var/home/core/zuul-output/0000755000175000017500000000000015113323172014523 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015113351030015461 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006705201515113351021017673 0ustar rootrootDec 01 14:33:53 crc systemd[1]: Starting Kubernetes Kubelet... Dec 01 14:33:53 crc restorecon[4737]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:53 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 14:33:54 crc restorecon[4737]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 14:33:54 crc restorecon[4737]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 01 14:33:54 crc kubenswrapper[4810]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 14:33:54 crc kubenswrapper[4810]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 01 14:33:54 crc kubenswrapper[4810]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 14:33:54 crc kubenswrapper[4810]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 14:33:54 crc kubenswrapper[4810]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 01 14:33:54 crc kubenswrapper[4810]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.342237 4810 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344687 4810 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344705 4810 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344709 4810 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344715 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344719 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344723 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344727 4810 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344732 4810 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344736 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344741 4810 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344746 4810 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344752 4810 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344756 4810 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344762 4810 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344766 4810 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344770 4810 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344774 4810 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344779 4810 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344784 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344788 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344793 4810 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344798 4810 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344802 4810 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344806 4810 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344810 4810 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344814 4810 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344818 4810 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344822 4810 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344825 4810 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344829 4810 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344833 4810 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344837 4810 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344840 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344844 4810 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344847 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344851 4810 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344855 4810 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344859 4810 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344862 4810 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344866 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344869 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344873 4810 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344877 4810 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344881 4810 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344884 4810 feature_gate.go:330] unrecognized feature gate: Example Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344889 4810 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344893 4810 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344897 4810 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344901 4810 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344905 4810 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344909 4810 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344913 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344917 4810 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344920 4810 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344924 4810 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344927 4810 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344931 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344934 4810 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344938 4810 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344941 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344944 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344948 4810 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344954 4810 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344958 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344961 4810 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344965 4810 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344968 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344972 4810 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344976 4810 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344979 4810 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.344983 4810 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345285 4810 flags.go:64] FLAG: --address="0.0.0.0" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345297 4810 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345304 4810 flags.go:64] FLAG: --anonymous-auth="true" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345310 4810 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345317 4810 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345321 4810 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345327 4810 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345332 4810 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345336 4810 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345340 4810 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345344 4810 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345349 4810 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345353 4810 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345357 4810 flags.go:64] FLAG: --cgroup-root="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345361 4810 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345365 4810 flags.go:64] FLAG: --client-ca-file="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345369 4810 flags.go:64] FLAG: --cloud-config="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345373 4810 flags.go:64] FLAG: --cloud-provider="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345377 4810 flags.go:64] FLAG: --cluster-dns="[]" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345381 4810 flags.go:64] FLAG: --cluster-domain="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345385 4810 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345389 4810 flags.go:64] FLAG: --config-dir="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345393 4810 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345397 4810 flags.go:64] FLAG: --container-log-max-files="5" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345403 4810 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345407 4810 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345411 4810 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345415 4810 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345420 4810 flags.go:64] FLAG: --contention-profiling="false" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345424 4810 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345428 4810 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345432 4810 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345436 4810 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345441 4810 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345445 4810 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345450 4810 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345456 4810 flags.go:64] FLAG: --enable-load-reader="false" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345460 4810 flags.go:64] FLAG: --enable-server="true" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345479 4810 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345485 4810 flags.go:64] FLAG: --event-burst="100" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345489 4810 flags.go:64] FLAG: --event-qps="50" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345493 4810 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345498 4810 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345502 4810 flags.go:64] FLAG: --eviction-hard="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345508 4810 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345512 4810 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345516 4810 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345520 4810 flags.go:64] FLAG: --eviction-soft="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345524 4810 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345528 4810 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345532 4810 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345537 4810 flags.go:64] FLAG: --experimental-mounter-path="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345541 4810 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345545 4810 flags.go:64] FLAG: --fail-swap-on="true" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345549 4810 flags.go:64] FLAG: --feature-gates="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345554 4810 flags.go:64] FLAG: --file-check-frequency="20s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345558 4810 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345562 4810 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345567 4810 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345570 4810 flags.go:64] FLAG: --healthz-port="10248" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345575 4810 flags.go:64] FLAG: --help="false" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345579 4810 flags.go:64] FLAG: --hostname-override="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345583 4810 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345587 4810 flags.go:64] FLAG: --http-check-frequency="20s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345591 4810 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345595 4810 flags.go:64] FLAG: --image-credential-provider-config="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345598 4810 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345604 4810 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345610 4810 flags.go:64] FLAG: --image-service-endpoint="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345614 4810 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345618 4810 flags.go:64] FLAG: --kube-api-burst="100" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345622 4810 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345626 4810 flags.go:64] FLAG: --kube-api-qps="50" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345630 4810 flags.go:64] FLAG: --kube-reserved="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345634 4810 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345638 4810 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345642 4810 flags.go:64] FLAG: --kubelet-cgroups="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345646 4810 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345650 4810 flags.go:64] FLAG: --lock-file="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345654 4810 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345658 4810 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345661 4810 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345667 4810 flags.go:64] FLAG: --log-json-split-stream="false" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345671 4810 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345676 4810 flags.go:64] FLAG: --log-text-split-stream="false" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345680 4810 flags.go:64] FLAG: --logging-format="text" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345683 4810 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345688 4810 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345692 4810 flags.go:64] FLAG: --manifest-url="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345696 4810 flags.go:64] FLAG: --manifest-url-header="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345701 4810 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345705 4810 flags.go:64] FLAG: --max-open-files="1000000" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345710 4810 flags.go:64] FLAG: --max-pods="110" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345714 4810 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345719 4810 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345723 4810 flags.go:64] FLAG: --memory-manager-policy="None" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345728 4810 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345732 4810 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345736 4810 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345741 4810 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345751 4810 flags.go:64] FLAG: --node-status-max-images="50" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345755 4810 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345759 4810 flags.go:64] FLAG: --oom-score-adj="-999" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345764 4810 flags.go:64] FLAG: --pod-cidr="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345768 4810 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345775 4810 flags.go:64] FLAG: --pod-manifest-path="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345779 4810 flags.go:64] FLAG: --pod-max-pids="-1" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345783 4810 flags.go:64] FLAG: --pods-per-core="0" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345787 4810 flags.go:64] FLAG: --port="10250" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345792 4810 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345796 4810 flags.go:64] FLAG: --provider-id="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345800 4810 flags.go:64] FLAG: --qos-reserved="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345804 4810 flags.go:64] FLAG: --read-only-port="10255" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345807 4810 flags.go:64] FLAG: --register-node="true" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345812 4810 flags.go:64] FLAG: --register-schedulable="true" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345815 4810 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345822 4810 flags.go:64] FLAG: --registry-burst="10" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345826 4810 flags.go:64] FLAG: --registry-qps="5" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345830 4810 flags.go:64] FLAG: --reserved-cpus="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345834 4810 flags.go:64] FLAG: --reserved-memory="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345839 4810 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345843 4810 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345847 4810 flags.go:64] FLAG: --rotate-certificates="false" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345851 4810 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345855 4810 flags.go:64] FLAG: --runonce="false" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345859 4810 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345862 4810 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345867 4810 flags.go:64] FLAG: --seccomp-default="false" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345872 4810 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345877 4810 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345882 4810 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345888 4810 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345892 4810 flags.go:64] FLAG: --storage-driver-password="root" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345898 4810 flags.go:64] FLAG: --storage-driver-secure="false" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345902 4810 flags.go:64] FLAG: --storage-driver-table="stats" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345906 4810 flags.go:64] FLAG: --storage-driver-user="root" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345910 4810 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345914 4810 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345919 4810 flags.go:64] FLAG: --system-cgroups="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345923 4810 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345929 4810 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345933 4810 flags.go:64] FLAG: --tls-cert-file="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345937 4810 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345942 4810 flags.go:64] FLAG: --tls-min-version="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345946 4810 flags.go:64] FLAG: --tls-private-key-file="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345950 4810 flags.go:64] FLAG: --topology-manager-policy="none" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345955 4810 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345958 4810 flags.go:64] FLAG: --topology-manager-scope="container" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345962 4810 flags.go:64] FLAG: --v="2" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345968 4810 flags.go:64] FLAG: --version="false" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345973 4810 flags.go:64] FLAG: --vmodule="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345978 4810 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.345983 4810 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346079 4810 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346084 4810 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346088 4810 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346092 4810 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346095 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346099 4810 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346104 4810 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346109 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346113 4810 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346117 4810 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346121 4810 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346125 4810 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346130 4810 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346134 4810 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346137 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346141 4810 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346144 4810 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346148 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346151 4810 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346155 4810 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346158 4810 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346162 4810 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346165 4810 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346169 4810 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346173 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346177 4810 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346180 4810 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346185 4810 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346188 4810 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346192 4810 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346195 4810 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346198 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346202 4810 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346206 4810 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346211 4810 feature_gate.go:330] unrecognized feature gate: Example Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346215 4810 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346220 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346223 4810 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346227 4810 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346230 4810 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346234 4810 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346237 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346241 4810 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346245 4810 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346250 4810 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346253 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346257 4810 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346261 4810 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346264 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346268 4810 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346271 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346275 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346278 4810 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346282 4810 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346285 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346289 4810 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346294 4810 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346298 4810 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346301 4810 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346305 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346309 4810 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346313 4810 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346316 4810 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346319 4810 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346323 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346326 4810 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346330 4810 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346333 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346336 4810 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346340 4810 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.346343 4810 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.346349 4810 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.355664 4810 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.355707 4810 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355788 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355795 4810 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355801 4810 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355805 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355809 4810 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355813 4810 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355816 4810 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355820 4810 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355823 4810 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355827 4810 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355831 4810 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355837 4810 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355841 4810 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355845 4810 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355848 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355853 4810 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355859 4810 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355863 4810 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355867 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355871 4810 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355876 4810 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355882 4810 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355885 4810 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355889 4810 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355893 4810 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355897 4810 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355901 4810 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355905 4810 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355908 4810 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355912 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355915 4810 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355922 4810 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355925 4810 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355930 4810 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355935 4810 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355939 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355943 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355946 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355950 4810 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355954 4810 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355958 4810 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355961 4810 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355965 4810 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355969 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355973 4810 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355977 4810 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355980 4810 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355984 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355987 4810 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355991 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.355995 4810 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356000 4810 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356004 4810 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356008 4810 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356012 4810 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356016 4810 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356020 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356024 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356027 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356031 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356035 4810 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356039 4810 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356042 4810 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356047 4810 feature_gate.go:330] unrecognized feature gate: Example Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356051 4810 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356055 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356059 4810 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356063 4810 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356066 4810 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356071 4810 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356077 4810 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.356084 4810 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356217 4810 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356224 4810 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356229 4810 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356233 4810 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356237 4810 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356241 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356245 4810 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356248 4810 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356252 4810 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356256 4810 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356261 4810 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356267 4810 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356272 4810 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356277 4810 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356281 4810 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356285 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356289 4810 feature_gate.go:330] unrecognized feature gate: Example Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356293 4810 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356297 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356301 4810 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356306 4810 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356310 4810 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356316 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356321 4810 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356326 4810 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356332 4810 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356337 4810 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356342 4810 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356348 4810 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356354 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356359 4810 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356365 4810 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356369 4810 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356373 4810 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356378 4810 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356382 4810 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356385 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356389 4810 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356393 4810 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356397 4810 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356400 4810 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356404 4810 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356407 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356411 4810 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356415 4810 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356418 4810 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356423 4810 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356427 4810 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356432 4810 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356436 4810 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356440 4810 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356444 4810 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356448 4810 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356452 4810 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356457 4810 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356461 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356465 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356484 4810 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356488 4810 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356492 4810 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356496 4810 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356499 4810 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356503 4810 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356507 4810 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356511 4810 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356514 4810 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356518 4810 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356522 4810 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356526 4810 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356529 4810 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.356534 4810 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.356540 4810 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.356735 4810 server.go:940] "Client rotation is on, will bootstrap in background" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.359327 4810 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.359429 4810 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.359983 4810 server.go:997] "Starting client certificate rotation" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.359997 4810 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.360170 4810 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-28 03:56:01.876962877 +0000 UTC Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.360238 4810 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 637h22m7.516728363s for next certificate rotation Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.371128 4810 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.373636 4810 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.382240 4810 log.go:25] "Validated CRI v1 runtime API" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.402233 4810 log.go:25] "Validated CRI v1 image API" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.404086 4810 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.406933 4810 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-01-14-29-17-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.406969 4810 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.422200 4810 manager.go:217] Machine: {Timestamp:2025-12-01 14:33:54.420635613 +0000 UTC m=+0.184145236 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:4b632de5-2300-4f1e-a112-9403c29ed772 BootID:4654003b-86dd-4016-8fc1-f1d206961e79 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:0c:d2:38 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:0c:d2:38 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:96:60:87 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:1c:21:6c Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:b7:a4:dd Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:2a:d3:ae Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:2b:0d:ca Speed:-1 Mtu:1496} {Name:eth10 MacAddress:82:c6:26:2e:ec:1f Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:b2:57:78:45:15:13 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.422450 4810 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.422631 4810 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.423400 4810 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.423711 4810 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.423774 4810 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.424026 4810 topology_manager.go:138] "Creating topology manager with none policy" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.424040 4810 container_manager_linux.go:303] "Creating device plugin manager" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.424377 4810 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.424522 4810 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.424701 4810 state_mem.go:36] "Initialized new in-memory state store" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.424796 4810 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.425413 4810 kubelet.go:418] "Attempting to sync node with API server" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.425430 4810 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.425610 4810 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.425630 4810 kubelet.go:324] "Adding apiserver pod source" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.425647 4810 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.427629 4810 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.428785 4810 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.429688 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.429670 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Dec 01 14:33:54 crc kubenswrapper[4810]: E1201 14:33:54.429862 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Dec 01 14:33:54 crc kubenswrapper[4810]: E1201 14:33:54.429785 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.431059 4810 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.431697 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.431721 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.431731 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.431740 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.431755 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.431764 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.431773 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.431787 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.431798 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.431808 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.431822 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.431832 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.432540 4810 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.433121 4810 server.go:1280] "Started kubelet" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.433293 4810 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.433523 4810 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.433522 4810 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.434007 4810 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 01 14:33:54 crc systemd[1]: Started Kubernetes Kubelet. Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.436625 4810 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.436653 4810 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 01 14:33:54 crc kubenswrapper[4810]: E1201 14:33:54.436880 4810 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.437014 4810 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.437153 4810 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 01 14:33:54 crc kubenswrapper[4810]: E1201 14:33:54.437330 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="200ms" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.437087 4810 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 13:14:36.795184557 +0000 UTC Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.437382 4810 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 550h40m42.357808307s for next certificate rotation Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.437395 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Dec 01 14:33:54 crc kubenswrapper[4810]: E1201 14:33:54.437440 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Dec 01 14:33:54 crc kubenswrapper[4810]: E1201 14:33:54.437288 4810 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.182:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187d1e04ad2b12f5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 14:33:54.433086197 +0000 UTC m=+0.196595800,LastTimestamp:2025-12-01 14:33:54.433086197 +0000 UTC m=+0.196595800,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.438042 4810 factory.go:55] Registering systemd factory Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.438131 4810 factory.go:221] Registration of the systemd container factory successfully Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.438173 4810 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.438444 4810 factory.go:153] Registering CRI-O factory Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.438487 4810 factory.go:221] Registration of the crio container factory successfully Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.438570 4810 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.438600 4810 factory.go:103] Registering Raw factory Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.438621 4810 manager.go:1196] Started watching for new ooms in manager Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.440174 4810 manager.go:319] Starting recovery of all containers Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.440906 4810 server.go:460] "Adding debug handlers to kubelet server" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450558 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450619 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450636 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450651 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450665 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450679 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450696 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450711 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450726 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450739 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450753 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450778 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450792 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450809 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450823 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450836 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450872 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450885 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450899 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450911 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450948 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450961 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450977 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.450995 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451014 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451053 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451073 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451089 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451105 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451124 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451143 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451186 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451201 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451217 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451236 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451255 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451273 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451287 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451300 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451319 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451339 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451356 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451375 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451393 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451412 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451430 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451447 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451465 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451505 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451523 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451541 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451558 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451580 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451599 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451617 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451638 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451658 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451675 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451693 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451711 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451727 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451746 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451765 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451783 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451801 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451821 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451840 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451857 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451876 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451894 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451912 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451930 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451948 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451966 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.451986 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452002 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452019 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452037 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452057 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452077 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452097 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452116 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452134 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452149 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452162 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452175 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452188 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452239 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452253 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452266 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452279 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452292 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452306 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452319 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452332 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452346 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452360 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452374 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452386 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452400 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452412 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452427 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452441 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452453 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452504 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452519 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452534 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452548 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452562 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452578 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452593 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452609 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452645 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452663 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452678 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452691 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452704 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452718 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452734 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452748 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452762 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452777 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452790 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452803 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452816 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452831 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452844 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452858 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452872 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452884 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452897 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452910 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452923 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452939 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452952 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452964 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452977 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.452990 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.453004 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.454970 4810 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455007 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455023 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455035 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455049 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455062 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455075 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455087 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455100 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455112 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455123 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455134 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455146 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455184 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455196 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455209 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455221 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455232 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455244 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455257 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455270 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455283 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455295 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455308 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455320 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455334 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455347 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455361 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455374 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455386 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455400 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455413 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455426 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455439 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455451 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455464 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455498 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455512 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455525 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455537 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455553 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455567 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455580 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455593 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455605 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455618 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455632 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455646 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455659 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455673 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455687 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455702 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455716 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455729 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455743 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455756 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455770 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455787 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455802 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455819 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455837 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455855 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455875 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455890 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455903 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455916 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455928 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455940 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455954 4810 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455967 4810 reconstruct.go:97] "Volume reconstruction finished" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.455976 4810 reconciler.go:26] "Reconciler: start to sync state" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.461295 4810 manager.go:324] Recovery completed Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.470217 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.472190 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.472333 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.472496 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.473083 4810 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.473175 4810 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.473269 4810 state_mem.go:36] "Initialized new in-memory state store" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.481559 4810 policy_none.go:49] "None policy: Start" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.482493 4810 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.482522 4810 state_mem.go:35] "Initializing new in-memory state store" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.488128 4810 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.489802 4810 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.489835 4810 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.489858 4810 kubelet.go:2335] "Starting kubelet main sync loop" Dec 01 14:33:54 crc kubenswrapper[4810]: E1201 14:33:54.489897 4810 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.490484 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Dec 01 14:33:54 crc kubenswrapper[4810]: E1201 14:33:54.490528 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.530572 4810 manager.go:334] "Starting Device Plugin manager" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.530697 4810 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.530709 4810 server.go:79] "Starting device plugin registration server" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.531067 4810 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.531083 4810 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.531410 4810 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.531525 4810 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.531535 4810 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 01 14:33:54 crc kubenswrapper[4810]: E1201 14:33:54.544798 4810 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.590534 4810 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.590729 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.592662 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.592721 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.592742 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.592972 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.593286 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.593350 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.593846 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.593879 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.593890 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.593995 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.594159 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.594240 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.594167 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.594304 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.594317 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.594714 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.594742 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.594755 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.594854 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.595086 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.595167 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.595248 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.595277 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.595286 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.595644 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.595670 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.595679 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.595791 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.595914 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.595940 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.596060 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.596087 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.596100 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.596308 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.596331 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.596339 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.596436 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.596453 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.596767 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.596806 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.596823 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.596945 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.596966 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.596978 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.631495 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.632532 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.632567 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.632578 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.632601 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 14:33:54 crc kubenswrapper[4810]: E1201 14:33:54.632996 4810 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.182:6443: connect: connection refused" node="crc" Dec 01 14:33:54 crc kubenswrapper[4810]: E1201 14:33:54.638984 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="400ms" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.657525 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.657760 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.657965 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.658123 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.658285 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.658434 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.658568 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.658722 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.658871 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.659013 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.659170 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.659318 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.659442 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.659605 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.659736 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761372 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761444 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761538 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761588 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761606 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761674 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761660 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761621 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761761 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761700 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761791 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761682 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761821 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761854 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761868 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761888 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761900 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761919 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761925 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761951 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761826 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761963 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761964 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.761982 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.762006 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.762027 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.762040 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.762070 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.762094 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.762171 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.834099 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.835358 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.835414 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.835427 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.835451 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 14:33:54 crc kubenswrapper[4810]: E1201 14:33:54.835943 4810 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.182:6443: connect: connection refused" node="crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.922295 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.947603 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.952782 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-d51eb7c6129a97a13c3b611dc2f85317a71cc15ea95e9ca75fb24cd6420acf15 WatchSource:0}: Error finding container d51eb7c6129a97a13c3b611dc2f85317a71cc15ea95e9ca75fb24cd6420acf15: Status 404 returned error can't find the container with id d51eb7c6129a97a13c3b611dc2f85317a71cc15ea95e9ca75fb24cd6420acf15 Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.961179 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.970303 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.975199 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-66b5ebceaa3d5ff36ad487a80b3219edb502fc5942ffbb73d91aae0f7a029d2d WatchSource:0}: Error finding container 66b5ebceaa3d5ff36ad487a80b3219edb502fc5942ffbb73d91aae0f7a029d2d: Status 404 returned error can't find the container with id 66b5ebceaa3d5ff36ad487a80b3219edb502fc5942ffbb73d91aae0f7a029d2d Dec 01 14:33:54 crc kubenswrapper[4810]: W1201 14:33:54.984391 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-033fc0e6de80ab7c3902fc0c3d0b60629bbf483a600890fbbd21202796a334d3 WatchSource:0}: Error finding container 033fc0e6de80ab7c3902fc0c3d0b60629bbf483a600890fbbd21202796a334d3: Status 404 returned error can't find the container with id 033fc0e6de80ab7c3902fc0c3d0b60629bbf483a600890fbbd21202796a334d3 Dec 01 14:33:54 crc kubenswrapper[4810]: I1201 14:33:54.990126 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 14:33:55 crc kubenswrapper[4810]: W1201 14:33:55.005488 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-81cb169400d25457841b4b51eeb7b72163bd31543cb0f63960393a5ee047d536 WatchSource:0}: Error finding container 81cb169400d25457841b4b51eeb7b72163bd31543cb0f63960393a5ee047d536: Status 404 returned error can't find the container with id 81cb169400d25457841b4b51eeb7b72163bd31543cb0f63960393a5ee047d536 Dec 01 14:33:55 crc kubenswrapper[4810]: E1201 14:33:55.039801 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="800ms" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.236395 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.237992 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.238031 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.238042 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.238065 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 14:33:55 crc kubenswrapper[4810]: E1201 14:33:55.238499 4810 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.182:6443: connect: connection refused" node="crc" Dec 01 14:33:55 crc kubenswrapper[4810]: W1201 14:33:55.336638 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Dec 01 14:33:55 crc kubenswrapper[4810]: E1201 14:33:55.336765 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Dec 01 14:33:55 crc kubenswrapper[4810]: W1201 14:33:55.343321 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Dec 01 14:33:55 crc kubenswrapper[4810]: E1201 14:33:55.343404 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Dec 01 14:33:55 crc kubenswrapper[4810]: W1201 14:33:55.369374 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Dec 01 14:33:55 crc kubenswrapper[4810]: E1201 14:33:55.369463 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.435639 4810 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.495182 4810 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="add82f0661fa60e28a4d8a1891f472f0c836ff78b9e67205f39d588ba7fd43ed" exitCode=0 Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.495284 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"add82f0661fa60e28a4d8a1891f472f0c836ff78b9e67205f39d588ba7fd43ed"} Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.495413 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"81cb169400d25457841b4b51eeb7b72163bd31543cb0f63960393a5ee047d536"} Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.495553 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.496813 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.496874 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.496886 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.497201 4810 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf" exitCode=0 Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.497315 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf"} Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.497357 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"033fc0e6de80ab7c3902fc0c3d0b60629bbf483a600890fbbd21202796a334d3"} Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.497525 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.498592 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e"} Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.498629 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.498632 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"66b5ebceaa3d5ff36ad487a80b3219edb502fc5942ffbb73d91aae0f7a029d2d"} Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.498661 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.498701 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.500973 4810 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6" exitCode=0 Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.501007 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6"} Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.501039 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b2a36b85474b7520d4de25cc2ed2959617748357fc221d847e832fd3d0344577"} Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.501124 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.502258 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.502295 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.502310 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.502856 4810 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179" exitCode=0 Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.502888 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179"} Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.502905 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d51eb7c6129a97a13c3b611dc2f85317a71cc15ea95e9ca75fb24cd6420acf15"} Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.502988 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.506457 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.506519 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.506536 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.507118 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.508661 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.508731 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:55 crc kubenswrapper[4810]: I1201 14:33:55.508745 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:55 crc kubenswrapper[4810]: W1201 14:33:55.689918 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Dec 01 14:33:55 crc kubenswrapper[4810]: E1201 14:33:55.689987 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Dec 01 14:33:55 crc kubenswrapper[4810]: E1201 14:33:55.840990 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="1.6s" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.039013 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.040770 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.040824 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.040837 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.040869 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.508065 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"306efebabacfe9294c92f371147b6fbf1ce71cd3d468eb9fcf0ca40534da5aaf"} Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.508247 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.511442 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.511546 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.511577 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.515845 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b24a67bb57bd282f48b4185451f120aaebab8373d0e368a70c97596b53a95213"} Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.515903 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2f7016800cb62ec155217c88e1d0968f86b6f807ab7a1e410579713f15741a4e"} Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.515915 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"46594b36eb97f9787facc7ee07e0d3e2aecd45816488438a8af9e619ff37aeb5"} Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.516031 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.517451 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.517725 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.517771 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.520143 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.520138 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769"} Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.520385 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460"} Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.520409 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c"} Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.521753 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.521833 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.521853 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.526423 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a"} Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.526489 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88"} Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.526508 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066"} Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.526524 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3"} Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.526539 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef"} Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.526664 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.527677 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.527716 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.527731 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.528568 4810 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec" exitCode=0 Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.528616 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec"} Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.528756 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.529611 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.529652 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:56 crc kubenswrapper[4810]: I1201 14:33:56.529668 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:57 crc kubenswrapper[4810]: I1201 14:33:57.536431 4810 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c" exitCode=0 Dec 01 14:33:57 crc kubenswrapper[4810]: I1201 14:33:57.536532 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c"} Dec 01 14:33:57 crc kubenswrapper[4810]: I1201 14:33:57.536583 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 14:33:57 crc kubenswrapper[4810]: I1201 14:33:57.536622 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:57 crc kubenswrapper[4810]: I1201 14:33:57.536703 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:57 crc kubenswrapper[4810]: I1201 14:33:57.536734 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:57 crc kubenswrapper[4810]: I1201 14:33:57.538203 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:57 crc kubenswrapper[4810]: I1201 14:33:57.538286 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:57 crc kubenswrapper[4810]: I1201 14:33:57.538322 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:57 crc kubenswrapper[4810]: I1201 14:33:57.538630 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:57 crc kubenswrapper[4810]: I1201 14:33:57.538790 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:57 crc kubenswrapper[4810]: I1201 14:33:57.538856 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:57 crc kubenswrapper[4810]: I1201 14:33:57.539332 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:57 crc kubenswrapper[4810]: I1201 14:33:57.539373 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:57 crc kubenswrapper[4810]: I1201 14:33:57.539387 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:58 crc kubenswrapper[4810]: I1201 14:33:58.536115 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:33:58 crc kubenswrapper[4810]: I1201 14:33:58.544977 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401"} Dec 01 14:33:58 crc kubenswrapper[4810]: I1201 14:33:58.545018 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684"} Dec 01 14:33:58 crc kubenswrapper[4810]: I1201 14:33:58.545031 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:58 crc kubenswrapper[4810]: I1201 14:33:58.545790 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:58 crc kubenswrapper[4810]: I1201 14:33:58.545817 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:58 crc kubenswrapper[4810]: I1201 14:33:58.545825 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:59 crc kubenswrapper[4810]: I1201 14:33:59.551702 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413"} Dec 01 14:33:59 crc kubenswrapper[4810]: I1201 14:33:59.551767 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51"} Dec 01 14:33:59 crc kubenswrapper[4810]: I1201 14:33:59.551789 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd"} Dec 01 14:33:59 crc kubenswrapper[4810]: I1201 14:33:59.551936 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:59 crc kubenswrapper[4810]: I1201 14:33:59.553163 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:59 crc kubenswrapper[4810]: I1201 14:33:59.553205 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:59 crc kubenswrapper[4810]: I1201 14:33:59.553221 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:33:59 crc kubenswrapper[4810]: I1201 14:33:59.769220 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:33:59 crc kubenswrapper[4810]: I1201 14:33:59.769462 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:33:59 crc kubenswrapper[4810]: I1201 14:33:59.770784 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:33:59 crc kubenswrapper[4810]: I1201 14:33:59.770831 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:33:59 crc kubenswrapper[4810]: I1201 14:33:59.770841 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.377340 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.377544 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.378904 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.378985 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.379003 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.555031 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.556813 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.556868 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.556885 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.602785 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.603035 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.604710 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.604781 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.604798 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.871362 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.936351 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.936752 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.939043 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.939147 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:00 crc kubenswrapper[4810]: I1201 14:34:00.939173 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:01 crc kubenswrapper[4810]: I1201 14:34:01.167760 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 01 14:34:01 crc kubenswrapper[4810]: I1201 14:34:01.557061 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:34:01 crc kubenswrapper[4810]: I1201 14:34:01.557062 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:34:01 crc kubenswrapper[4810]: I1201 14:34:01.558149 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:01 crc kubenswrapper[4810]: I1201 14:34:01.558194 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:01 crc kubenswrapper[4810]: I1201 14:34:01.558210 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:01 crc kubenswrapper[4810]: I1201 14:34:01.558287 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:01 crc kubenswrapper[4810]: I1201 14:34:01.558308 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:01 crc kubenswrapper[4810]: I1201 14:34:01.558317 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:01 crc kubenswrapper[4810]: I1201 14:34:01.571347 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:34:01 crc kubenswrapper[4810]: I1201 14:34:01.571786 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:34:01 crc kubenswrapper[4810]: I1201 14:34:01.573135 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:01 crc kubenswrapper[4810]: I1201 14:34:01.573203 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:01 crc kubenswrapper[4810]: I1201 14:34:01.573219 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:03 crc kubenswrapper[4810]: I1201 14:34:03.528984 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:34:03 crc kubenswrapper[4810]: I1201 14:34:03.529157 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:34:03 crc kubenswrapper[4810]: I1201 14:34:03.530437 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:03 crc kubenswrapper[4810]: I1201 14:34:03.530534 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:03 crc kubenswrapper[4810]: I1201 14:34:03.530552 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:03 crc kubenswrapper[4810]: I1201 14:34:03.534254 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:34:03 crc kubenswrapper[4810]: I1201 14:34:03.562365 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:34:03 crc kubenswrapper[4810]: I1201 14:34:03.563379 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:03 crc kubenswrapper[4810]: I1201 14:34:03.563420 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:03 crc kubenswrapper[4810]: I1201 14:34:03.563433 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:04 crc kubenswrapper[4810]: E1201 14:34:04.545578 4810 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 01 14:34:04 crc kubenswrapper[4810]: I1201 14:34:04.572341 4810 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 01 14:34:04 crc kubenswrapper[4810]: I1201 14:34:04.572400 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 01 14:34:06 crc kubenswrapper[4810]: E1201 14:34:06.041806 4810 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 01 14:34:06 crc kubenswrapper[4810]: I1201 14:34:06.435044 4810 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 01 14:34:06 crc kubenswrapper[4810]: I1201 14:34:06.484837 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 01 14:34:06 crc kubenswrapper[4810]: I1201 14:34:06.485141 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:34:06 crc kubenswrapper[4810]: I1201 14:34:06.486810 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:06 crc kubenswrapper[4810]: I1201 14:34:06.486857 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:06 crc kubenswrapper[4810]: I1201 14:34:06.486869 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:07 crc kubenswrapper[4810]: W1201 14:34:07.251422 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 01 14:34:07 crc kubenswrapper[4810]: I1201 14:34:07.251669 4810 trace.go:236] Trace[1028951567]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 14:33:57.250) (total time: 10001ms): Dec 01 14:34:07 crc kubenswrapper[4810]: Trace[1028951567]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (14:34:07.251) Dec 01 14:34:07 crc kubenswrapper[4810]: Trace[1028951567]: [10.001588861s] [10.001588861s] END Dec 01 14:34:07 crc kubenswrapper[4810]: E1201 14:34:07.251728 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 01 14:34:07 crc kubenswrapper[4810]: W1201 14:34:07.353848 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 01 14:34:07 crc kubenswrapper[4810]: I1201 14:34:07.353932 4810 trace.go:236] Trace[1358584749]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 14:33:57.352) (total time: 10001ms): Dec 01 14:34:07 crc kubenswrapper[4810]: Trace[1358584749]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (14:34:07.353) Dec 01 14:34:07 crc kubenswrapper[4810]: Trace[1358584749]: [10.001515088s] [10.001515088s] END Dec 01 14:34:07 crc kubenswrapper[4810]: E1201 14:34:07.353956 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 01 14:34:07 crc kubenswrapper[4810]: E1201 14:34:07.442703 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 01 14:34:07 crc kubenswrapper[4810]: W1201 14:34:07.595267 4810 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 01 14:34:07 crc kubenswrapper[4810]: I1201 14:34:07.595392 4810 trace.go:236] Trace[44294609]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 14:33:57.594) (total time: 10001ms): Dec 01 14:34:07 crc kubenswrapper[4810]: Trace[44294609]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (14:34:07.595) Dec 01 14:34:07 crc kubenswrapper[4810]: Trace[44294609]: [10.001140339s] [10.001140339s] END Dec 01 14:34:07 crc kubenswrapper[4810]: E1201 14:34:07.595414 4810 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 01 14:34:07 crc kubenswrapper[4810]: I1201 14:34:07.642099 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:34:07 crc kubenswrapper[4810]: I1201 14:34:07.643676 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:07 crc kubenswrapper[4810]: I1201 14:34:07.643791 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:07 crc kubenswrapper[4810]: I1201 14:34:07.643882 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:07 crc kubenswrapper[4810]: I1201 14:34:07.643983 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 14:34:08 crc kubenswrapper[4810]: I1201 14:34:08.184860 4810 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 01 14:34:08 crc kubenswrapper[4810]: I1201 14:34:08.184925 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 01 14:34:08 crc kubenswrapper[4810]: I1201 14:34:08.194242 4810 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 01 14:34:08 crc kubenswrapper[4810]: I1201 14:34:08.194309 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 01 14:34:08 crc kubenswrapper[4810]: I1201 14:34:08.543091 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:34:08 crc kubenswrapper[4810]: I1201 14:34:08.543251 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:34:08 crc kubenswrapper[4810]: I1201 14:34:08.544616 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:08 crc kubenswrapper[4810]: I1201 14:34:08.544638 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:08 crc kubenswrapper[4810]: I1201 14:34:08.544647 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:10 crc kubenswrapper[4810]: I1201 14:34:10.610540 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:34:10 crc kubenswrapper[4810]: I1201 14:34:10.610678 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:34:10 crc kubenswrapper[4810]: I1201 14:34:10.611892 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:10 crc kubenswrapper[4810]: I1201 14:34:10.612058 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:10 crc kubenswrapper[4810]: I1201 14:34:10.612238 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:10 crc kubenswrapper[4810]: I1201 14:34:10.618249 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:34:11 crc kubenswrapper[4810]: I1201 14:34:11.581267 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 14:34:11 crc kubenswrapper[4810]: I1201 14:34:11.581324 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:34:11 crc kubenswrapper[4810]: I1201 14:34:11.582297 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:11 crc kubenswrapper[4810]: I1201 14:34:11.582347 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:11 crc kubenswrapper[4810]: I1201 14:34:11.582358 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.221460 4810 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.435139 4810 apiserver.go:52] "Watching apiserver" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.437984 4810 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.438224 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.438605 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.438690 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.438701 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 14:34:12 crc kubenswrapper[4810]: E1201 14:34:12.438809 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:12 crc kubenswrapper[4810]: E1201 14:34:12.438884 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.439230 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.439271 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.439678 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 14:34:12 crc kubenswrapper[4810]: E1201 14:34:12.440088 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.440586 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.440684 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.440946 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.441945 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.442216 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.443330 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.443443 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.443455 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.443637 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.465764 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.477582 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.491882 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.528138 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.539409 4810 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.546726 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.556658 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.566711 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.575215 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:12 crc kubenswrapper[4810]: I1201 14:34:12.620001 4810 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.190932 4810 trace.go:236] Trace[1379694306]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 14:33:58.547) (total time: 14643ms): Dec 01 14:34:13 crc kubenswrapper[4810]: Trace[1379694306]: ---"Objects listed" error: 14643ms (14:34:13.190) Dec 01 14:34:13 crc kubenswrapper[4810]: Trace[1379694306]: [14.643849416s] [14.643849416s] END Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.190972 4810 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.191372 4810 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.191913 4810 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.219177 4810 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51924->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.219246 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51924->192.168.126.11:17697: read: connection reset by peer" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.219267 4810 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51938->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.219352 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51938->192.168.126.11:17697: read: connection reset by peer" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.219668 4810 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.219726 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.269761 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.281992 4810 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.283329 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.286726 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.289800 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.291983 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292089 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292117 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292142 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292167 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292189 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292212 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292237 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.292262 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:34:13.79223458 +0000 UTC m=+19.555744203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292316 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292352 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292378 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292400 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292425 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292447 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292458 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292543 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292572 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292585 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292520 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292640 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292664 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292681 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292699 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292714 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292756 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292774 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292782 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292792 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292791 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292804 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292844 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292874 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292898 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292925 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292938 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292948 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292971 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.292991 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293010 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293019 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293062 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293091 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293104 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293106 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293137 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293154 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293183 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293200 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293199 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293219 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293265 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293263 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293281 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293297 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293301 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293312 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293326 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293379 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293395 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293409 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293423 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293423 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293449 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293464 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293464 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293502 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293508 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293518 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293533 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293549 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293564 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293585 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293599 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293612 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293627 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293641 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293655 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293653 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293671 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293689 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293705 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293720 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293736 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293778 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293792 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293809 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293825 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293840 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293854 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293870 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293884 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293899 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293917 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293933 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293948 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293994 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294010 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294026 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294040 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294056 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294071 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294088 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294104 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294119 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294136 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294152 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294166 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294182 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294197 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294214 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294229 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294244 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294258 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294272 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294287 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294302 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294317 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294394 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294409 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294427 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294443 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294463 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294494 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294512 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294530 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294547 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294562 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294577 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294593 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294607 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294623 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294637 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294653 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294669 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294684 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294700 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294717 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294734 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294750 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294765 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294780 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294798 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294813 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294828 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294843 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294860 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294876 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294892 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294909 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294925 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294941 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294955 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294970 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294989 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295003 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295020 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295035 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295052 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295067 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295082 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295098 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295114 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295131 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295147 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295163 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295180 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295196 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295213 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295228 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295246 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295261 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295277 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295297 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295313 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295329 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295344 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295361 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295380 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295396 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295413 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295429 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295445 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295461 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295490 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295505 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295522 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295560 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295584 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295609 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295636 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295661 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295683 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295706 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295735 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295758 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295815 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295843 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295868 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295891 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295914 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295939 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295964 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295988 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296008 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296032 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296058 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296082 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296108 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296132 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296156 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296181 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296203 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296228 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296279 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296307 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296328 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296353 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296379 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296405 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296443 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296493 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296516 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296544 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296572 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296604 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296628 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296651 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296723 4810 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296740 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296756 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296769 4810 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296781 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296793 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296804 4810 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296818 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296831 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296842 4810 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296854 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296866 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296878 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296891 4810 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296904 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296919 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296934 4810 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296948 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293714 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293874 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293910 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293989 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.293992 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294175 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294253 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294262 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294400 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294419 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294458 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294491 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294647 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294677 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294683 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294707 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294772 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294836 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294888 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.294893 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295026 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295076 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295114 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295163 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295202 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295341 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295374 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.295593 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296181 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296215 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296386 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296530 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296551 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296757 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296790 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296919 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.296935 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.297088 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.310871 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:13.810846714 +0000 UTC m=+19.574356317 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.310894 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.310933 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.297148 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.297309 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.297340 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.297397 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.297418 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.298066 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.311199 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.311276 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.311461 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.311643 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.297126 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.298242 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.298276 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.298693 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.298767 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.299196 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.299325 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.299331 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.299464 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.299554 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.300083 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.299977 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.300139 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.300412 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.300431 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.300436 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.300541 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.300735 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.300773 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.300790 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.300792 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.300806 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.301177 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.301703 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.301977 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.302278 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.302560 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.302676 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.302883 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.302998 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.303307 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.303336 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.303578 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.303766 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.304019 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.304670 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.304988 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.305654 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.305601 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.306025 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.306442 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.307003 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.307174 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.308118 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.308364 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.308452 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.308646 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.309076 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.309236 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.309672 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.309537 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.312352 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.309809 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.310275 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.310364 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.310770 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.312067 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.312865 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.313480 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.313487 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.313672 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.313944 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.314227 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.314292 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.314406 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.314604 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.314709 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.314913 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.315312 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.315410 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.315526 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.315720 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.315764 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.315799 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.316183 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.316319 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.316331 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.316685 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.316724 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.316811 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.316835 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.316863 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.316942 4810 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.317240 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.317298 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.317383 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.317615 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.317778 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.317830 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.317832 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.317939 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.317987 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.318002 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.318674 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.318698 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.319024 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.298064 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.320616 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.325687 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:13.825601425 +0000 UTC m=+19.589111198 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.332854 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.332860 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.332807 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.333006 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.333189 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.333243 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.333457 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.333695 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.333742 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.333774 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.333873 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:13.833841893 +0000 UTC m=+19.597351536 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.334753 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.335036 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.335064 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.335082 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.335151 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:13.835123751 +0000 UTC m=+19.598633564 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.337129 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.337386 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.339730 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.340362 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.342926 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.343748 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.344018 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.344777 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.345555 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.345951 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.346174 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.346429 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.347348 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.347406 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.348085 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.348257 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.349732 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.348846 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.351295 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.351285 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.351623 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.352214 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.352381 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.353494 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.353627 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.354082 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.365899 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.367689 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.369743 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.381909 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.382109 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.384786 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.386128 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.395287 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.400985 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.405451 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.406961 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411250 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411380 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411445 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411539 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411614 4810 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411632 4810 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411650 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411665 4810 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411679 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411694 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411706 4810 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411718 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411731 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411743 4810 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411755 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411768 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411767 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411781 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411845 4810 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411863 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411882 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411897 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411912 4810 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411926 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411942 4810 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411958 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411974 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.411990 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412005 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412020 4810 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412033 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412046 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412059 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412071 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412085 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412103 4810 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412120 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412137 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412153 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412168 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412186 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412203 4810 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412248 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412267 4810 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412284 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412301 4810 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412320 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412337 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412354 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412369 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412384 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412400 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412413 4810 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412425 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412436 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412450 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412466 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412511 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412529 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412545 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412565 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412582 4810 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412856 4810 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412879 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412895 4810 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412914 4810 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412932 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412946 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412961 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412978 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.412994 4810 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413009 4810 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413051 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413070 4810 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413085 4810 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413100 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413117 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413132 4810 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413147 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413166 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413196 4810 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413212 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413229 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413245 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413263 4810 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413280 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413297 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413313 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413331 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413346 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413362 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413376 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413392 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413408 4810 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413428 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413445 4810 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413495 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413512 4810 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413527 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413543 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413557 4810 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413573 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413588 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413612 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413628 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413643 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413658 4810 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413674 4810 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413702 4810 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413722 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413742 4810 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413759 4810 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413775 4810 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413790 4810 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413807 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413823 4810 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413841 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413860 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413880 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413896 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413911 4810 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413926 4810 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413941 4810 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413956 4810 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413974 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.413990 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.414007 4810 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.414022 4810 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.414037 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.414052 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.414067 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.414083 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.414276 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.414307 4810 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.414324 4810 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.414339 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.414357 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.414373 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.414388 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.414404 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.414422 4810 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.414437 4810 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.414453 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415669 4810 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415689 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415705 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415723 4810 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415739 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415754 4810 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415769 4810 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415783 4810 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415799 4810 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415815 4810 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415829 4810 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415844 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415858 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415873 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415888 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415904 4810 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415921 4810 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415936 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415951 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415967 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415983 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.415999 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.416016 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.416030 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.416044 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.416060 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.416075 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.416093 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.416112 4810 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.416129 4810 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.416146 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.416162 4810 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.416178 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.416192 4810 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.416207 4810 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.416222 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.431837 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.438650 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.444043 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.456798 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.465103 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.472122 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.488329 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.516635 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.587075 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.588788 4810 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a" exitCode=255 Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.588859 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a"} Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.590536 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff"} Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.590588 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a"} Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.590600 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d752fc77e91417ce0b960b90a259d5dc8ea26926489de642c9688132e1b18a4f"} Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.598899 4810 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.599516 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.599699 4810 scope.go:117] "RemoveContainer" containerID="871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.601185 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.611242 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.625180 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.637194 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.646104 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.665078 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.674272 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.680417 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.697167 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 14:34:13 crc kubenswrapper[4810]: W1201 14:34:13.701642 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-c3f9650ab84bd626f23c17028a132180795d64077ac66ff0dee716f191212a40 WatchSource:0}: Error finding container c3f9650ab84bd626f23c17028a132180795d64077ac66ff0dee716f191212a40: Status 404 returned error can't find the container with id c3f9650ab84bd626f23c17028a132180795d64077ac66ff0dee716f191212a40 Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.726455 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:13Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.742173 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:13Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.756702 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:13Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.768534 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:13Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.783622 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:13Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.799631 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:13Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.817766 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:13Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.819866 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.819951 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.820024 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.820074 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:34:14.820044309 +0000 UTC m=+20.583553912 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.820116 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:14.82010581 +0000 UTC m=+20.583615413 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.834888 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:13Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.920409 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.920455 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:13 crc kubenswrapper[4810]: I1201 14:34:13.920520 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.920604 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.920639 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.920650 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.920682 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.920695 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:14.920681496 +0000 UTC m=+20.684191099 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.920702 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.920716 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.920765 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:14.920750247 +0000 UTC m=+20.684259900 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.920619 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 14:34:13 crc kubenswrapper[4810]: E1201 14:34:13.920800 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:14.920792408 +0000 UTC m=+20.684302111 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.490260 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.490368 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:14 crc kubenswrapper[4810]: E1201 14:34:14.490377 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.490385 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:14 crc kubenswrapper[4810]: E1201 14:34:14.490565 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:14 crc kubenswrapper[4810]: E1201 14:34:14.490659 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.494035 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.494684 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.495291 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.495862 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.496396 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.497649 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.498181 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.499052 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.499625 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.500454 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.500980 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.502010 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.502446 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.502932 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.503829 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.504321 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.505311 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.505975 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.506659 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.507784 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.508307 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.509452 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.510053 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.511232 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.511786 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.512466 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.513894 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.514450 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.515729 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.516212 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.517122 4810 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.517223 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.518873 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.519875 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.520316 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.521871 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.522312 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.522641 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.523528 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.524152 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.525203 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.525859 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.526944 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.527582 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.528614 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.529137 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.530067 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.530752 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.531995 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.532504 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.533312 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.533841 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.535013 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.535769 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.536307 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.546540 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.581338 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.593590 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.594029 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"60ab2ac0eae4a77e37819321b03067567241e4e69e2cd139fde0432566a3ee8a"} Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.595183 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995"} Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.595209 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c3f9650ab84bd626f23c17028a132180795d64077ac66ff0dee716f191212a40"} Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.596742 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.598882 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135"} Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.598912 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.605201 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.617145 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.629113 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.643990 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.661408 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.691676 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.707508 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.718849 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.729431 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.740026 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.753710 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.767041 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.827331 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.827420 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:14 crc kubenswrapper[4810]: E1201 14:34:14.827495 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 14:34:14 crc kubenswrapper[4810]: E1201 14:34:14.827510 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:34:16.827460059 +0000 UTC m=+22.590969682 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:34:14 crc kubenswrapper[4810]: E1201 14:34:14.827553 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:16.827540081 +0000 UTC m=+22.591049694 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.928673 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.928718 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:14 crc kubenswrapper[4810]: I1201 14:34:14.928737 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:14 crc kubenswrapper[4810]: E1201 14:34:14.928832 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 14:34:14 crc kubenswrapper[4810]: E1201 14:34:14.928845 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 14:34:14 crc kubenswrapper[4810]: E1201 14:34:14.928854 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:14 crc kubenswrapper[4810]: E1201 14:34:14.928922 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:16.928910154 +0000 UTC m=+22.692419757 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:14 crc kubenswrapper[4810]: E1201 14:34:14.929198 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 14:34:14 crc kubenswrapper[4810]: E1201 14:34:14.929227 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:16.92922043 +0000 UTC m=+22.692730033 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 14:34:14 crc kubenswrapper[4810]: E1201 14:34:14.929265 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 14:34:14 crc kubenswrapper[4810]: E1201 14:34:14.929274 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 14:34:14 crc kubenswrapper[4810]: E1201 14:34:14.929283 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:14 crc kubenswrapper[4810]: E1201 14:34:14.929302 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:16.929295592 +0000 UTC m=+22.692805195 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.392941 4810 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.394491 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.394523 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.394534 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.394589 4810 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.401187 4810 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.401516 4810 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.402546 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.402576 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.402620 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.402636 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.402647 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:16Z","lastTransitionTime":"2025-12-01T14:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.419165 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.422451 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.422503 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.422520 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.422540 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.422843 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:16Z","lastTransitionTime":"2025-12-01T14:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.433426 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.436376 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.436410 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.436418 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.436459 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.436485 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:16Z","lastTransitionTime":"2025-12-01T14:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.451148 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.455434 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.455485 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.455494 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.455521 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.455531 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:16Z","lastTransitionTime":"2025-12-01T14:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.466417 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.471007 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.471042 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.471050 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.471064 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.471073 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:16Z","lastTransitionTime":"2025-12-01T14:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.481944 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.482056 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.483621 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.483656 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.483667 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.483855 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.483869 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:16Z","lastTransitionTime":"2025-12-01T14:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.490440 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.490484 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.490440 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.490577 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.490674 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.490744 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.504063 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.517812 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.530417 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.531403 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.556369 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.568108 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.579758 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.585906 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.585943 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.585956 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.585972 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.585982 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:16Z","lastTransitionTime":"2025-12-01T14:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.589972 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.602499 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.603729 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf"} Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.612515 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.622802 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.634916 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.646820 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.658060 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.669758 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.680341 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.688311 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.688352 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.688375 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.688392 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.688403 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:16Z","lastTransitionTime":"2025-12-01T14:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.700802 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.711583 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.725393 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.743635 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:16Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.790398 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.790435 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.790444 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.790459 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.790483 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:16Z","lastTransitionTime":"2025-12-01T14:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.844831 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.844962 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.845032 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:34:20.844999729 +0000 UTC m=+26.608509332 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.845073 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.845156 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:20.845133632 +0000 UTC m=+26.608643285 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.892721 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.892757 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.892766 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.892781 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.892791 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:16Z","lastTransitionTime":"2025-12-01T14:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.945959 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.946029 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.946074 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.946217 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.946286 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:20.946265849 +0000 UTC m=+26.709775482 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.946621 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.946647 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.946659 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.946697 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:20.946685209 +0000 UTC m=+26.710194822 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.946766 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.946825 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.946840 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:16 crc kubenswrapper[4810]: E1201 14:34:16.946906 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:20.946885893 +0000 UTC m=+26.710395496 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.994708 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.994746 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.994755 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.994769 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:16 crc kubenswrapper[4810]: I1201 14:34:16.994779 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:16Z","lastTransitionTime":"2025-12-01T14:34:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.096974 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.097033 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.097049 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.097070 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.097084 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:17Z","lastTransitionTime":"2025-12-01T14:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.199530 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.199583 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.199595 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.199612 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.199626 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:17Z","lastTransitionTime":"2025-12-01T14:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.301952 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.302008 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.302023 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.302044 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.302057 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:17Z","lastTransitionTime":"2025-12-01T14:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.404652 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.404693 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.404704 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.404720 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.404731 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:17Z","lastTransitionTime":"2025-12-01T14:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.506601 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.506646 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.506660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.506679 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.506693 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:17Z","lastTransitionTime":"2025-12-01T14:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.608692 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.608732 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.608745 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.608762 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.608778 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:17Z","lastTransitionTime":"2025-12-01T14:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.711612 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.711661 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.711673 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.711691 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.711702 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:17Z","lastTransitionTime":"2025-12-01T14:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.813560 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.813608 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.813617 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.813632 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.813642 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:17Z","lastTransitionTime":"2025-12-01T14:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.915882 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.915930 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.915947 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.915965 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:17 crc kubenswrapper[4810]: I1201 14:34:17.915976 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:17Z","lastTransitionTime":"2025-12-01T14:34:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.018195 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.018233 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.018243 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.018257 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.018267 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:18Z","lastTransitionTime":"2025-12-01T14:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.120931 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.120965 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.120976 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.120988 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.120997 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:18Z","lastTransitionTime":"2025-12-01T14:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.222677 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.222712 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.222722 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.222735 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.222747 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:18Z","lastTransitionTime":"2025-12-01T14:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.324963 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.325009 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.325022 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.325039 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.325051 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:18Z","lastTransitionTime":"2025-12-01T14:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.427164 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.427198 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.427208 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.427222 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.427234 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:18Z","lastTransitionTime":"2025-12-01T14:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.474604 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-v7dlj"] Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.474912 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-v7dlj" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.477309 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.477489 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.478351 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.489783 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.490018 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.490034 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.490122 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:18 crc kubenswrapper[4810]: E1201 14:34:18.490222 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:18 crc kubenswrapper[4810]: E1201 14:34:18.490310 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:18 crc kubenswrapper[4810]: E1201 14:34:18.490493 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.509594 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.522748 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.531394 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.531440 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.531455 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.531492 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.531508 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:18Z","lastTransitionTime":"2025-12-01T14:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.549895 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.560366 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/293d2ee5-5bc6-40ca-a176-6ea87131aa9d-hosts-file\") pod \"node-resolver-v7dlj\" (UID: \"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\") " pod="openshift-dns/node-resolver-v7dlj" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.560422 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4xws\" (UniqueName: \"kubernetes.io/projected/293d2ee5-5bc6-40ca-a176-6ea87131aa9d-kube-api-access-v4xws\") pod \"node-resolver-v7dlj\" (UID: \"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\") " pod="openshift-dns/node-resolver-v7dlj" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.567629 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.613465 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.634865 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.634909 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.634921 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.634939 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.634949 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:18Z","lastTransitionTime":"2025-12-01T14:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.651667 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.661152 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/293d2ee5-5bc6-40ca-a176-6ea87131aa9d-hosts-file\") pod \"node-resolver-v7dlj\" (UID: \"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\") " pod="openshift-dns/node-resolver-v7dlj" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.661199 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4xws\" (UniqueName: \"kubernetes.io/projected/293d2ee5-5bc6-40ca-a176-6ea87131aa9d-kube-api-access-v4xws\") pod \"node-resolver-v7dlj\" (UID: \"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\") " pod="openshift-dns/node-resolver-v7dlj" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.661348 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/293d2ee5-5bc6-40ca-a176-6ea87131aa9d-hosts-file\") pod \"node-resolver-v7dlj\" (UID: \"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\") " pod="openshift-dns/node-resolver-v7dlj" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.688729 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4xws\" (UniqueName: \"kubernetes.io/projected/293d2ee5-5bc6-40ca-a176-6ea87131aa9d-kube-api-access-v4xws\") pod \"node-resolver-v7dlj\" (UID: \"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\") " pod="openshift-dns/node-resolver-v7dlj" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.696892 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.731993 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.737672 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.737726 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.737739 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.737757 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.737769 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:18Z","lastTransitionTime":"2025-12-01T14:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.754099 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.787512 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-v7dlj" Dec 01 14:34:18 crc kubenswrapper[4810]: W1201 14:34:18.806526 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod293d2ee5_5bc6_40ca_a176_6ea87131aa9d.slice/crio-8a0959a00d6cdfaeb0baaf89445b0819b1756e32fb22a251c88f9a51eac932e9 WatchSource:0}: Error finding container 8a0959a00d6cdfaeb0baaf89445b0819b1756e32fb22a251c88f9a51eac932e9: Status 404 returned error can't find the container with id 8a0959a00d6cdfaeb0baaf89445b0819b1756e32fb22a251c88f9a51eac932e9 Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.840616 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.840663 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.840696 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.840713 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.840724 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:18Z","lastTransitionTime":"2025-12-01T14:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.896012 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-rbt4q"] Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.896531 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-s4cc7"] Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.896730 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.899069 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-s4cc7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.899093 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-kbfbn"] Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.899957 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.900872 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.901830 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.902026 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.902133 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.902156 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.902326 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.902408 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.902559 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.902673 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.902867 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.903009 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.903308 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.913316 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.931280 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.943829 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.944133 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.944244 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.944335 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.944429 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:18Z","lastTransitionTime":"2025-12-01T14:34:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.945362 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.957013 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963153 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/df7ecc53-6478-49cd-bfb6-6ee80e850a19-cni-binary-copy\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963188 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-host-var-lib-kubelet\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963223 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/48f9858e-b927-4c2f-ab39-ded37b102dcd-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963242 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/48f9858e-b927-4c2f-ab39-ded37b102dcd-os-release\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963255 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/48f9858e-b927-4c2f-ab39-ded37b102dcd-tuning-conf-dir\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963277 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/48f9858e-b927-4c2f-ab39-ded37b102dcd-system-cni-dir\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963292 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/48f9858e-b927-4c2f-ab39-ded37b102dcd-cnibin\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963307 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-multus-cni-dir\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963325 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-host-run-multus-certs\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963353 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-os-release\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963367 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tg8g\" (UniqueName: \"kubernetes.io/projected/48f9858e-b927-4c2f-ab39-ded37b102dcd-kube-api-access-4tg8g\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963381 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/48f9858e-b927-4c2f-ab39-ded37b102dcd-cni-binary-copy\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963415 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-host-var-lib-cni-multus\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963432 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-multus-conf-dir\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963447 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6c275317-c741-4d61-a399-d196f37f1745-proxy-tls\") pod \"machine-config-daemon-rbt4q\" (UID: \"6c275317-c741-4d61-a399-d196f37f1745\") " pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963480 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwkk7\" (UniqueName: \"kubernetes.io/projected/6c275317-c741-4d61-a399-d196f37f1745-kube-api-access-mwkk7\") pod \"machine-config-daemon-rbt4q\" (UID: \"6c275317-c741-4d61-a399-d196f37f1745\") " pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963495 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-hostroot\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963510 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-system-cni-dir\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963533 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6c275317-c741-4d61-a399-d196f37f1745-mcd-auth-proxy-config\") pod \"machine-config-daemon-rbt4q\" (UID: \"6c275317-c741-4d61-a399-d196f37f1745\") " pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963547 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-cnibin\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963560 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/df7ecc53-6478-49cd-bfb6-6ee80e850a19-multus-daemon-config\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963579 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnqq5\" (UniqueName: \"kubernetes.io/projected/df7ecc53-6478-49cd-bfb6-6ee80e850a19-kube-api-access-jnqq5\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963594 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6c275317-c741-4d61-a399-d196f37f1745-rootfs\") pod \"machine-config-daemon-rbt4q\" (UID: \"6c275317-c741-4d61-a399-d196f37f1745\") " pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963616 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-multus-socket-dir-parent\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963633 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-host-run-netns\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963647 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-etc-kubernetes\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963662 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-host-run-k8s-cni-cncf-io\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.963682 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-host-var-lib-cni-bin\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.979401 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:18 crc kubenswrapper[4810]: I1201 14:34:18.991099 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.008215 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.019162 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.034222 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.047223 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.047586 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.047733 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.047862 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.047995 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:19Z","lastTransitionTime":"2025-12-01T14:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.048924 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.061762 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068084 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6c275317-c741-4d61-a399-d196f37f1745-proxy-tls\") pod \"machine-config-daemon-rbt4q\" (UID: \"6c275317-c741-4d61-a399-d196f37f1745\") " pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068127 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwkk7\" (UniqueName: \"kubernetes.io/projected/6c275317-c741-4d61-a399-d196f37f1745-kube-api-access-mwkk7\") pod \"machine-config-daemon-rbt4q\" (UID: \"6c275317-c741-4d61-a399-d196f37f1745\") " pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068166 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-hostroot\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068199 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-system-cni-dir\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068217 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6c275317-c741-4d61-a399-d196f37f1745-mcd-auth-proxy-config\") pod \"machine-config-daemon-rbt4q\" (UID: \"6c275317-c741-4d61-a399-d196f37f1745\") " pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068253 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-cnibin\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068271 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/df7ecc53-6478-49cd-bfb6-6ee80e850a19-multus-daemon-config\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068289 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnqq5\" (UniqueName: \"kubernetes.io/projected/df7ecc53-6478-49cd-bfb6-6ee80e850a19-kube-api-access-jnqq5\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068322 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6c275317-c741-4d61-a399-d196f37f1745-rootfs\") pod \"machine-config-daemon-rbt4q\" (UID: \"6c275317-c741-4d61-a399-d196f37f1745\") " pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068340 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-multus-socket-dir-parent\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068357 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-host-run-netns\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068374 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-etc-kubernetes\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068411 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-host-run-k8s-cni-cncf-io\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068428 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-host-var-lib-cni-bin\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068445 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/df7ecc53-6478-49cd-bfb6-6ee80e850a19-cni-binary-copy\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068497 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-host-var-lib-kubelet\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068524 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/48f9858e-b927-4c2f-ab39-ded37b102dcd-os-release\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068539 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/48f9858e-b927-4c2f-ab39-ded37b102dcd-tuning-conf-dir\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068572 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/48f9858e-b927-4c2f-ab39-ded37b102dcd-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068601 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/48f9858e-b927-4c2f-ab39-ded37b102dcd-system-cni-dir\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068616 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/48f9858e-b927-4c2f-ab39-ded37b102dcd-cnibin\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068649 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-multus-cni-dir\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068666 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-host-run-multus-certs\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068692 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-os-release\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068729 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tg8g\" (UniqueName: \"kubernetes.io/projected/48f9858e-b927-4c2f-ab39-ded37b102dcd-kube-api-access-4tg8g\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068748 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/48f9858e-b927-4c2f-ab39-ded37b102dcd-cni-binary-copy\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068769 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-host-var-lib-cni-multus\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068806 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-multus-conf-dir\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068846 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-host-run-k8s-cni-cncf-io\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068892 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-multus-conf-dir\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.068938 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-host-var-lib-cni-bin\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.069184 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-hostroot\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.069407 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-system-cni-dir\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.069456 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/6c275317-c741-4d61-a399-d196f37f1745-rootfs\") pod \"machine-config-daemon-rbt4q\" (UID: \"6c275317-c741-4d61-a399-d196f37f1745\") " pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.069543 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-cnibin\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.069991 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/df7ecc53-6478-49cd-bfb6-6ee80e850a19-cni-binary-copy\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.070045 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6c275317-c741-4d61-a399-d196f37f1745-mcd-auth-proxy-config\") pod \"machine-config-daemon-rbt4q\" (UID: \"6c275317-c741-4d61-a399-d196f37f1745\") " pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.070064 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-host-var-lib-kubelet\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.070107 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-multus-socket-dir-parent\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.070136 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-host-run-netns\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.070153 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/48f9858e-b927-4c2f-ab39-ded37b102dcd-os-release\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.070161 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-etc-kubernetes\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.070554 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/48f9858e-b927-4c2f-ab39-ded37b102dcd-tuning-conf-dir\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.070169 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/df7ecc53-6478-49cd-bfb6-6ee80e850a19-multus-daemon-config\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.070672 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-host-var-lib-cni-multus\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.070752 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-multus-cni-dir\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.070787 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/48f9858e-b927-4c2f-ab39-ded37b102dcd-cni-binary-copy\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.070828 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/48f9858e-b927-4c2f-ab39-ded37b102dcd-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.070852 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-host-run-multus-certs\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.070833 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/df7ecc53-6478-49cd-bfb6-6ee80e850a19-os-release\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.070903 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/48f9858e-b927-4c2f-ab39-ded37b102dcd-system-cni-dir\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.070918 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/48f9858e-b927-4c2f-ab39-ded37b102dcd-cnibin\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.071751 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6c275317-c741-4d61-a399-d196f37f1745-proxy-tls\") pod \"machine-config-daemon-rbt4q\" (UID: \"6c275317-c741-4d61-a399-d196f37f1745\") " pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.078896 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.089035 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnqq5\" (UniqueName: \"kubernetes.io/projected/df7ecc53-6478-49cd-bfb6-6ee80e850a19-kube-api-access-jnqq5\") pod \"multus-s4cc7\" (UID: \"df7ecc53-6478-49cd-bfb6-6ee80e850a19\") " pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.093065 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tg8g\" (UniqueName: \"kubernetes.io/projected/48f9858e-b927-4c2f-ab39-ded37b102dcd-kube-api-access-4tg8g\") pod \"multus-additional-cni-plugins-kbfbn\" (UID: \"48f9858e-b927-4c2f-ab39-ded37b102dcd\") " pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.094924 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.096732 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwkk7\" (UniqueName: \"kubernetes.io/projected/6c275317-c741-4d61-a399-d196f37f1745-kube-api-access-mwkk7\") pod \"machine-config-daemon-rbt4q\" (UID: \"6c275317-c741-4d61-a399-d196f37f1745\") " pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.108737 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.121165 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.131258 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.143534 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.149788 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.149816 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.149825 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.149839 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.149848 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:19Z","lastTransitionTime":"2025-12-01T14:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.153834 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.165457 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.180300 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.194343 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.213908 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.213958 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.223283 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-s4cc7" Dec 01 14:34:19 crc kubenswrapper[4810]: W1201 14:34:19.224756 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c275317_c741_4d61_a399_d196f37f1745.slice/crio-4610eec6204516603e1f38e0890cf23fce75ec40433011f162b08c706c6e007b WatchSource:0}: Error finding container 4610eec6204516603e1f38e0890cf23fce75ec40433011f162b08c706c6e007b: Status 404 returned error can't find the container with id 4610eec6204516603e1f38e0890cf23fce75ec40433011f162b08c706c6e007b Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.230203 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.252486 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.252533 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.252543 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.252559 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.252569 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:19Z","lastTransitionTime":"2025-12-01T14:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.261925 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.273069 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p7pvm"] Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.273892 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.275845 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.278429 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.278451 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.278602 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.278713 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.278855 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.278863 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.279035 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.322950 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.337439 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.362353 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.362384 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.362393 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.362407 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.362417 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:19Z","lastTransitionTime":"2025-12-01T14:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.371985 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-var-lib-openvswitch\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.372106 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.372199 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-run-ovn-kubernetes\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.372114 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.372391 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7xmf\" (UniqueName: \"kubernetes.io/projected/efba5556-97da-447f-9dda-40c2ea6d3e3d-kube-api-access-q7xmf\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.372516 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-node-log\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.372610 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-run-netns\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.372677 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-cni-netd\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.372816 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/efba5556-97da-447f-9dda-40c2ea6d3e3d-ovnkube-script-lib\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.372887 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-systemd-units\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.372957 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-slash\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.373040 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-cni-bin\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.373108 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-run-systemd\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.373174 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-etc-openvswitch\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.373250 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-run-openvswitch\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.373319 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-log-socket\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.373392 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-run-ovn\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.373462 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/efba5556-97da-447f-9dda-40c2ea6d3e3d-ovnkube-config\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.373565 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/efba5556-97da-447f-9dda-40c2ea6d3e3d-ovn-node-metrics-cert\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.373646 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/efba5556-97da-447f-9dda-40c2ea6d3e3d-env-overrides\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.373732 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-kubelet\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.388947 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.400952 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.426124 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.443181 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.455820 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.465348 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.465396 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.465405 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.465422 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.465431 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:19Z","lastTransitionTime":"2025-12-01T14:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.468816 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474335 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-var-lib-openvswitch\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474378 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474406 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-run-ovn-kubernetes\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474421 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7xmf\" (UniqueName: \"kubernetes.io/projected/efba5556-97da-447f-9dda-40c2ea6d3e3d-kube-api-access-q7xmf\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474436 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-node-log\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474458 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-cni-netd\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474488 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/efba5556-97da-447f-9dda-40c2ea6d3e3d-ovnkube-script-lib\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474503 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-run-netns\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474517 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-systemd-units\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474530 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-slash\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474552 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-cni-bin\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474567 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-run-systemd\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474583 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-etc-openvswitch\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474599 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-run-openvswitch\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474613 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-log-socket\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474628 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/efba5556-97da-447f-9dda-40c2ea6d3e3d-ovn-node-metrics-cert\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474643 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-run-ovn\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474656 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/efba5556-97da-447f-9dda-40c2ea6d3e3d-ovnkube-config\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474669 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/efba5556-97da-447f-9dda-40c2ea6d3e3d-env-overrides\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474685 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-kubelet\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474741 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-kubelet\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474773 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-var-lib-openvswitch\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474795 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.474814 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-run-ovn-kubernetes\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.475044 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-node-log\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.475075 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-cni-netd\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.475722 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/efba5556-97da-447f-9dda-40c2ea6d3e3d-ovnkube-script-lib\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.475765 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-run-netns\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.475832 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-systemd-units\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.475856 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-slash\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.475883 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-cni-bin\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.475918 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-run-systemd\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.475949 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-etc-openvswitch\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.475980 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-run-openvswitch\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.476009 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-log-socket\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.476505 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-run-ovn\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.476849 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/efba5556-97da-447f-9dda-40c2ea6d3e3d-env-overrides\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.477892 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/efba5556-97da-447f-9dda-40c2ea6d3e3d-ovnkube-config\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.483131 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/efba5556-97da-447f-9dda-40c2ea6d3e3d-ovn-node-metrics-cert\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.489006 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.493945 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7xmf\" (UniqueName: \"kubernetes.io/projected/efba5556-97da-447f-9dda-40c2ea6d3e3d-kube-api-access-q7xmf\") pod \"ovnkube-node-p7pvm\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.507143 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.519511 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.531012 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.545402 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.567660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.567700 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.567710 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.567724 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.567734 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:19Z","lastTransitionTime":"2025-12-01T14:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.598037 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.613031 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-v7dlj" event={"ID":"293d2ee5-5bc6-40ca-a176-6ea87131aa9d","Type":"ContainerStarted","Data":"1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.613071 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-v7dlj" event={"ID":"293d2ee5-5bc6-40ca-a176-6ea87131aa9d","Type":"ContainerStarted","Data":"8a0959a00d6cdfaeb0baaf89445b0819b1756e32fb22a251c88f9a51eac932e9"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.615546 4810 generic.go:334] "Generic (PLEG): container finished" podID="48f9858e-b927-4c2f-ab39-ded37b102dcd" containerID="9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8" exitCode=0 Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.615624 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" event={"ID":"48f9858e-b927-4c2f-ab39-ded37b102dcd","Type":"ContainerDied","Data":"9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.615673 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" event={"ID":"48f9858e-b927-4c2f-ab39-ded37b102dcd","Type":"ContainerStarted","Data":"d2eb84e9c4fc2fe7c98eb2b9a65c77d5a78f853dac83164b99675f88f44bc297"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.622218 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-s4cc7" event={"ID":"df7ecc53-6478-49cd-bfb6-6ee80e850a19","Type":"ContainerStarted","Data":"70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.622302 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-s4cc7" event={"ID":"df7ecc53-6478-49cd-bfb6-6ee80e850a19","Type":"ContainerStarted","Data":"002593f2b2a846d034e4186546e01d17626dcaadf425ace8ed5fb6012bc805c0"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.625055 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.625651 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.625703 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.625721 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"4610eec6204516603e1f38e0890cf23fce75ec40433011f162b08c706c6e007b"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.640595 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.664626 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: W1201 14:34:19.668076 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podefba5556_97da_447f_9dda_40c2ea6d3e3d.slice/crio-5410ffd03955b8b5558821baec0e9cef19e443be2c486539b91a5a1a35cec10d WatchSource:0}: Error finding container 5410ffd03955b8b5558821baec0e9cef19e443be2c486539b91a5a1a35cec10d: Status 404 returned error can't find the container with id 5410ffd03955b8b5558821baec0e9cef19e443be2c486539b91a5a1a35cec10d Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.672851 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.672893 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.672903 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.672919 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.672929 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:19Z","lastTransitionTime":"2025-12-01T14:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.681401 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.694349 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.709950 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.724940 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.738029 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.758939 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.779321 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.779359 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.779368 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.779384 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.779395 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:19Z","lastTransitionTime":"2025-12-01T14:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.780747 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.799287 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.817981 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.830532 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.843700 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.858645 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.870121 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.880268 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.882155 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.882215 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.882232 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.882258 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.882277 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:19Z","lastTransitionTime":"2025-12-01T14:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.892521 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.910613 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.922463 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.935131 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.947112 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.965891 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.980247 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.987138 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.987179 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.987188 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.987202 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.987213 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:19Z","lastTransitionTime":"2025-12-01T14:34:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:19 crc kubenswrapper[4810]: I1201 14:34:19.995323 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.005449 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.025633 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.070390 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.089674 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.089705 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.089713 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.089729 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.089738 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:20Z","lastTransitionTime":"2025-12-01T14:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.193111 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.193458 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.193495 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.193516 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.193529 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:20Z","lastTransitionTime":"2025-12-01T14:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.296387 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.296432 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.296446 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.296462 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.297182 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:20Z","lastTransitionTime":"2025-12-01T14:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.404019 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.404054 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.404062 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.404076 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.404085 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:20Z","lastTransitionTime":"2025-12-01T14:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.490775 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.490826 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:20 crc kubenswrapper[4810]: E1201 14:34:20.490894 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.490775 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:20 crc kubenswrapper[4810]: E1201 14:34:20.490983 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:20 crc kubenswrapper[4810]: E1201 14:34:20.491147 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.505672 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.505706 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.505713 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.505729 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.505738 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:20Z","lastTransitionTime":"2025-12-01T14:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.608573 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.608612 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.608621 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.608638 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.608649 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:20Z","lastTransitionTime":"2025-12-01T14:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.630350 4810 generic.go:334] "Generic (PLEG): container finished" podID="48f9858e-b927-4c2f-ab39-ded37b102dcd" containerID="eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26" exitCode=0 Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.630446 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" event={"ID":"48f9858e-b927-4c2f-ab39-ded37b102dcd","Type":"ContainerDied","Data":"eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26"} Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.632559 4810 generic.go:334] "Generic (PLEG): container finished" podID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerID="13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96" exitCode=0 Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.632587 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerDied","Data":"13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96"} Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.632602 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerStarted","Data":"5410ffd03955b8b5558821baec0e9cef19e443be2c486539b91a5a1a35cec10d"} Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.643236 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.661652 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.674333 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.690576 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.713656 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.713692 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.713703 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.713721 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.713732 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:20Z","lastTransitionTime":"2025-12-01T14:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.714276 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.728391 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.749055 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.766046 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.784740 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.785860 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-h8m66"] Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.787032 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-h8m66" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.789445 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.789733 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.789875 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.789997 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.799893 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.816942 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.819015 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.819049 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.819059 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.819077 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.819087 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:20Z","lastTransitionTime":"2025-12-01T14:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.829190 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.843067 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.856257 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.889360 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.890268 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.890400 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/87c01a68-ffa6-46eb-97b2-65fc457d8a79-host\") pod \"node-ca-h8m66\" (UID: \"87c01a68-ffa6-46eb-97b2-65fc457d8a79\") " pod="openshift-image-registry/node-ca-h8m66" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.890440 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:20 crc kubenswrapper[4810]: E1201 14:34:20.890552 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 14:34:20 crc kubenswrapper[4810]: E1201 14:34:20.890549 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:34:28.890491636 +0000 UTC m=+34.654001249 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.890634 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/87c01a68-ffa6-46eb-97b2-65fc457d8a79-serviceca\") pod \"node-ca-h8m66\" (UID: \"87c01a68-ffa6-46eb-97b2-65fc457d8a79\") " pod="openshift-image-registry/node-ca-h8m66" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.890724 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcs69\" (UniqueName: \"kubernetes.io/projected/87c01a68-ffa6-46eb-97b2-65fc457d8a79-kube-api-access-fcs69\") pod \"node-ca-h8m66\" (UID: \"87c01a68-ffa6-46eb-97b2-65fc457d8a79\") " pod="openshift-image-registry/node-ca-h8m66" Dec 01 14:34:20 crc kubenswrapper[4810]: E1201 14:34:20.890776 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:28.890722161 +0000 UTC m=+34.654231754 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.918719 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.921447 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.921498 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.921508 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.921522 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.921533 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:20Z","lastTransitionTime":"2025-12-01T14:34:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.938482 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.949769 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.961360 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.971659 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.991491 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/87c01a68-ffa6-46eb-97b2-65fc457d8a79-host\") pod \"node-ca-h8m66\" (UID: \"87c01a68-ffa6-46eb-97b2-65fc457d8a79\") " pod="openshift-image-registry/node-ca-h8m66" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.991570 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.991595 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/87c01a68-ffa6-46eb-97b2-65fc457d8a79-serviceca\") pod \"node-ca-h8m66\" (UID: \"87c01a68-ffa6-46eb-97b2-65fc457d8a79\") " pod="openshift-image-registry/node-ca-h8m66" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.991619 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.991641 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcs69\" (UniqueName: \"kubernetes.io/projected/87c01a68-ffa6-46eb-97b2-65fc457d8a79-kube-api-access-fcs69\") pod \"node-ca-h8m66\" (UID: \"87c01a68-ffa6-46eb-97b2-65fc457d8a79\") " pod="openshift-image-registry/node-ca-h8m66" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.991666 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.991657 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/87c01a68-ffa6-46eb-97b2-65fc457d8a79-host\") pod \"node-ca-h8m66\" (UID: \"87c01a68-ffa6-46eb-97b2-65fc457d8a79\") " pod="openshift-image-registry/node-ca-h8m66" Dec 01 14:34:20 crc kubenswrapper[4810]: E1201 14:34:20.992046 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 14:34:20 crc kubenswrapper[4810]: E1201 14:34:20.992118 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 14:34:20 crc kubenswrapper[4810]: E1201 14:34:20.992153 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 14:34:20 crc kubenswrapper[4810]: E1201 14:34:20.992179 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 14:34:20 crc kubenswrapper[4810]: E1201 14:34:20.992189 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:28.992150495 +0000 UTC m=+34.755660108 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 14:34:20 crc kubenswrapper[4810]: E1201 14:34:20.992194 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:20 crc kubenswrapper[4810]: E1201 14:34:20.992256 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:28.992246357 +0000 UTC m=+34.755755970 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:20 crc kubenswrapper[4810]: E1201 14:34:20.992159 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 14:34:20 crc kubenswrapper[4810]: E1201 14:34:20.992331 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:20 crc kubenswrapper[4810]: E1201 14:34:20.992399 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:28.99237834 +0000 UTC m=+34.755888143 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.993221 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/87c01a68-ffa6-46eb-97b2-65fc457d8a79-serviceca\") pod \"node-ca-h8m66\" (UID: \"87c01a68-ffa6-46eb-97b2-65fc457d8a79\") " pod="openshift-image-registry/node-ca-h8m66" Dec 01 14:34:20 crc kubenswrapper[4810]: I1201 14:34:20.997431 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.023190 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcs69\" (UniqueName: \"kubernetes.io/projected/87c01a68-ffa6-46eb-97b2-65fc457d8a79-kube-api-access-fcs69\") pod \"node-ca-h8m66\" (UID: \"87c01a68-ffa6-46eb-97b2-65fc457d8a79\") " pod="openshift-image-registry/node-ca-h8m66" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.029643 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.029690 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.029704 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.029720 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.029740 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:21Z","lastTransitionTime":"2025-12-01T14:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.054206 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.087963 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.125816 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.128921 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-h8m66" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.131403 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.131438 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.131447 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.131479 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.131489 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:21Z","lastTransitionTime":"2025-12-01T14:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:21 crc kubenswrapper[4810]: W1201 14:34:21.142281 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87c01a68_ffa6_46eb_97b2_65fc457d8a79.slice/crio-b9ab5b35d3b73eb9e4a8ff18ae59ab91e89f13dd31cc84c960a44fc73ea974b5 WatchSource:0}: Error finding container b9ab5b35d3b73eb9e4a8ff18ae59ab91e89f13dd31cc84c960a44fc73ea974b5: Status 404 returned error can't find the container with id b9ab5b35d3b73eb9e4a8ff18ae59ab91e89f13dd31cc84c960a44fc73ea974b5 Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.165073 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.208085 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.234060 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.234097 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.234109 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.234123 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.234132 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:21Z","lastTransitionTime":"2025-12-01T14:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.247528 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.297245 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.327773 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.337284 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.337318 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.337328 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.337343 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.337353 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:21Z","lastTransitionTime":"2025-12-01T14:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.440089 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.440131 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.440144 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.440162 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.440173 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:21Z","lastTransitionTime":"2025-12-01T14:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.542193 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.542276 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.542289 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.542306 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.542344 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:21Z","lastTransitionTime":"2025-12-01T14:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.637094 4810 generic.go:334] "Generic (PLEG): container finished" podID="48f9858e-b927-4c2f-ab39-ded37b102dcd" containerID="f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748" exitCode=0 Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.637179 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" event={"ID":"48f9858e-b927-4c2f-ab39-ded37b102dcd","Type":"ContainerDied","Data":"f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748"} Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.643499 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerStarted","Data":"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8"} Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.643554 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerStarted","Data":"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb"} Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.643566 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerStarted","Data":"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce"} Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.643574 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerStarted","Data":"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f"} Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.643583 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerStarted","Data":"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026"} Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.643591 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerStarted","Data":"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0"} Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.644695 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.644722 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.644731 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.644747 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.644761 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:21Z","lastTransitionTime":"2025-12-01T14:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.646189 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-h8m66" event={"ID":"87c01a68-ffa6-46eb-97b2-65fc457d8a79","Type":"ContainerStarted","Data":"a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3"} Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.646242 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-h8m66" event={"ID":"87c01a68-ffa6-46eb-97b2-65fc457d8a79","Type":"ContainerStarted","Data":"b9ab5b35d3b73eb9e4a8ff18ae59ab91e89f13dd31cc84c960a44fc73ea974b5"} Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.652929 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.664645 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.685022 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.702294 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.713857 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.725443 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.739282 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.746439 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.746496 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.746507 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.746523 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.746532 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:21Z","lastTransitionTime":"2025-12-01T14:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.754138 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.768417 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.778873 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.796361 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.813722 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.846962 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.849851 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.849884 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.849895 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.849911 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.849920 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:21Z","lastTransitionTime":"2025-12-01T14:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.883885 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.924585 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.952066 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.952108 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.952118 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.952133 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.952145 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:21Z","lastTransitionTime":"2025-12-01T14:34:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:21 crc kubenswrapper[4810]: I1201 14:34:21.964596 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.010729 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.054947 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.054981 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.054989 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.055004 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.055013 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:22Z","lastTransitionTime":"2025-12-01T14:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.056640 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.086225 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.128314 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.157866 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.158193 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.158342 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.158538 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.158712 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:22Z","lastTransitionTime":"2025-12-01T14:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.169917 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.203872 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.245743 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.261486 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.261724 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.261791 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.261856 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.261926 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:22Z","lastTransitionTime":"2025-12-01T14:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.285262 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.333737 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.363626 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.363667 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.363676 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.363691 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.363701 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:22Z","lastTransitionTime":"2025-12-01T14:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.366734 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.405702 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.449768 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.466454 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.466527 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.466544 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.466567 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.466583 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:22Z","lastTransitionTime":"2025-12-01T14:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.485995 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.492707 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:22 crc kubenswrapper[4810]: E1201 14:34:22.492917 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.493418 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:22 crc kubenswrapper[4810]: E1201 14:34:22.493656 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.493723 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:22 crc kubenswrapper[4810]: E1201 14:34:22.493868 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.534510 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.568642 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.568678 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.568686 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.568702 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.568714 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:22Z","lastTransitionTime":"2025-12-01T14:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.655327 4810 generic.go:334] "Generic (PLEG): container finished" podID="48f9858e-b927-4c2f-ab39-ded37b102dcd" containerID="df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75" exitCode=0 Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.655401 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" event={"ID":"48f9858e-b927-4c2f-ab39-ded37b102dcd","Type":"ContainerDied","Data":"df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75"} Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.671035 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.671096 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.671113 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.671139 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.671158 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:22Z","lastTransitionTime":"2025-12-01T14:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.680693 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.705192 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.736289 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.751810 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.764591 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.773203 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.773247 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.773255 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.773270 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.773279 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:22Z","lastTransitionTime":"2025-12-01T14:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.777980 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.804672 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.853451 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.875517 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.875554 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.875564 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.875578 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.875586 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:22Z","lastTransitionTime":"2025-12-01T14:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.886066 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.926395 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.964810 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:22Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.977960 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.977987 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.977996 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.978008 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:22 crc kubenswrapper[4810]: I1201 14:34:22.978035 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:22Z","lastTransitionTime":"2025-12-01T14:34:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.005671 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.047583 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.079556 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.079587 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.079597 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.079610 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.079618 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:23Z","lastTransitionTime":"2025-12-01T14:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.085975 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.130737 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.182863 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.182907 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.182920 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.182938 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.182953 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:23Z","lastTransitionTime":"2025-12-01T14:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.285548 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.285587 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.285604 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.285621 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.285632 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:23Z","lastTransitionTime":"2025-12-01T14:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.388986 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.389033 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.389051 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.389072 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.389087 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:23Z","lastTransitionTime":"2025-12-01T14:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.490970 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.491146 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.491274 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.491400 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.491537 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:23Z","lastTransitionTime":"2025-12-01T14:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.594434 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.594865 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.594881 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.594901 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.594913 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:23Z","lastTransitionTime":"2025-12-01T14:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.662842 4810 generic.go:334] "Generic (PLEG): container finished" podID="48f9858e-b927-4c2f-ab39-ded37b102dcd" containerID="4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec" exitCode=0 Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.663021 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" event={"ID":"48f9858e-b927-4c2f-ab39-ded37b102dcd","Type":"ContainerDied","Data":"4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec"} Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.671715 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerStarted","Data":"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d"} Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.682363 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.697715 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.697772 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.697785 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.697801 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.697814 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:23Z","lastTransitionTime":"2025-12-01T14:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.701369 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.716846 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.729927 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.740419 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.753266 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.761597 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.781328 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.792725 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.801915 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.801965 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.801980 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.801999 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.802011 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:23Z","lastTransitionTime":"2025-12-01T14:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.806082 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.818892 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.833071 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.845029 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.858768 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.881849 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:23Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.904796 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.904838 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.904848 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.904865 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:23 crc kubenswrapper[4810]: I1201 14:34:23.904875 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:23Z","lastTransitionTime":"2025-12-01T14:34:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.007674 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.007715 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.007726 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.007741 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.007751 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:24Z","lastTransitionTime":"2025-12-01T14:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.110184 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.110213 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.110221 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.110234 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.110242 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:24Z","lastTransitionTime":"2025-12-01T14:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.212178 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.212241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.212251 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.212264 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.212273 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:24Z","lastTransitionTime":"2025-12-01T14:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.315292 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.315342 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.315352 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.315369 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.315384 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:24Z","lastTransitionTime":"2025-12-01T14:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.421206 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.421499 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.421638 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.421743 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.421827 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:24Z","lastTransitionTime":"2025-12-01T14:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.490419 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.490460 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.490527 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:24 crc kubenswrapper[4810]: E1201 14:34:24.490635 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:24 crc kubenswrapper[4810]: E1201 14:34:24.490739 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:24 crc kubenswrapper[4810]: E1201 14:34:24.490857 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.505585 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.519579 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.530119 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.539391 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.539650 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.539724 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.539782 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.539836 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:24Z","lastTransitionTime":"2025-12-01T14:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.551690 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.570840 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.584754 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.601908 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.617212 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.631608 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.642069 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.642101 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.642110 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.642123 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.642133 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:24Z","lastTransitionTime":"2025-12-01T14:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.644943 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.663942 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.676897 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.679624 4810 generic.go:334] "Generic (PLEG): container finished" podID="48f9858e-b927-4c2f-ab39-ded37b102dcd" containerID="5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de" exitCode=0 Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.679659 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" event={"ID":"48f9858e-b927-4c2f-ab39-ded37b102dcd","Type":"ContainerDied","Data":"5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de"} Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.689215 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.699183 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.710148 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.723179 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.735133 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.744180 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.744446 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.744543 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.744606 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.744668 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:24Z","lastTransitionTime":"2025-12-01T14:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.746989 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.764268 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.777388 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.791771 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.803798 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.816786 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.837244 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.847272 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.847338 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.847348 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.847362 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.847371 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:24Z","lastTransitionTime":"2025-12-01T14:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.852235 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.864361 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.877667 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.887848 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.899846 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.930132 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:24Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.950365 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.950646 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.950781 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.950875 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:24 crc kubenswrapper[4810]: I1201 14:34:24.950964 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:24Z","lastTransitionTime":"2025-12-01T14:34:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.053341 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.053407 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.053424 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.053447 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.053462 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:25Z","lastTransitionTime":"2025-12-01T14:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.156218 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.156631 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.156645 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.156661 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.156669 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:25Z","lastTransitionTime":"2025-12-01T14:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.259374 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.259409 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.259418 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.259431 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.259442 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:25Z","lastTransitionTime":"2025-12-01T14:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.361707 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.361735 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.361743 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.361758 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.361767 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:25Z","lastTransitionTime":"2025-12-01T14:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.464514 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.464551 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.464561 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.464575 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.464586 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:25Z","lastTransitionTime":"2025-12-01T14:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.566803 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.566837 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.566846 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.566861 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.566872 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:25Z","lastTransitionTime":"2025-12-01T14:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.672107 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.672141 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.672151 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.672166 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.672177 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:25Z","lastTransitionTime":"2025-12-01T14:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.687278 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" event={"ID":"48f9858e-b927-4c2f-ab39-ded37b102dcd","Type":"ContainerStarted","Data":"101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c"} Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.702797 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:25Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.716194 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:25Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.733264 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:25Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.748412 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:25Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.763032 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:25Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.775901 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.775938 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.775951 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.775969 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.775981 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:25Z","lastTransitionTime":"2025-12-01T14:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.777670 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:25Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.789613 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:25Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.823741 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:25Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.840418 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:25Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.856743 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:25Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.871602 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:25Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.880291 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.880355 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.880368 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.880390 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.880404 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:25Z","lastTransitionTime":"2025-12-01T14:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.888235 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:25Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.905109 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:25Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.918259 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:25Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.936849 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:25Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.983518 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.983569 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.983582 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.983601 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:25 crc kubenswrapper[4810]: I1201 14:34:25.983614 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:25Z","lastTransitionTime":"2025-12-01T14:34:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.085913 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.085950 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.085962 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.085980 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.085993 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:26Z","lastTransitionTime":"2025-12-01T14:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.188104 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.188138 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.188146 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.188164 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.188174 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:26Z","lastTransitionTime":"2025-12-01T14:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.290822 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.290861 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.290874 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.290891 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.290903 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:26Z","lastTransitionTime":"2025-12-01T14:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.393673 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.393703 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.393712 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.393726 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.393735 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:26Z","lastTransitionTime":"2025-12-01T14:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.492861 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:26 crc kubenswrapper[4810]: E1201 14:34:26.493004 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.492853 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.492861 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:26 crc kubenswrapper[4810]: E1201 14:34:26.493081 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:26 crc kubenswrapper[4810]: E1201 14:34:26.493329 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.496729 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.496794 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.496821 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.496851 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.496873 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:26Z","lastTransitionTime":"2025-12-01T14:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.600272 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.600309 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.600316 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.600329 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.600338 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:26Z","lastTransitionTime":"2025-12-01T14:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.695620 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerStarted","Data":"f24941df3884bd7df557177fb9087ceaec0c30c2c3b2f1d86fd33d8bfe8c21c8"} Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.696291 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.702773 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.702835 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.702852 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.702874 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.702894 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:26Z","lastTransitionTime":"2025-12-01T14:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.714454 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.724992 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.735389 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.753983 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.772614 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.787933 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.806152 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.806192 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.806202 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.806219 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.806232 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:26Z","lastTransitionTime":"2025-12-01T14:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.814101 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.832186 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.847038 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.860612 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.866601 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.866648 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.866658 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.866676 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.866687 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:26Z","lastTransitionTime":"2025-12-01T14:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.870818 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: E1201 14:34:26.876829 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.880069 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.880111 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.880124 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.880146 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.880159 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:26Z","lastTransitionTime":"2025-12-01T14:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.885955 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: E1201 14:34:26.891305 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.898360 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.898390 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.898399 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.898416 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.898429 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:26Z","lastTransitionTime":"2025-12-01T14:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.905864 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: E1201 14:34:26.912988 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.916904 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.916937 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.916946 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.916961 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.916972 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:26Z","lastTransitionTime":"2025-12-01T14:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.922978 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: E1201 14:34:26.929293 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.934270 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.934314 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.934327 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.934349 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.934364 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:26Z","lastTransitionTime":"2025-12-01T14:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.938576 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: E1201 14:34:26.945557 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: E1201 14:34:26.945679 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.947362 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.947402 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.947415 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.947432 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.947444 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:26Z","lastTransitionTime":"2025-12-01T14:34:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.960314 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f24941df3884bd7df557177fb9087ceaec0c30c2c3b2f1d86fd33d8bfe8c21c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.975704 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:26 crc kubenswrapper[4810]: I1201 14:34:26.989893 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:26Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.010614 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f24941df3884bd7df557177fb9087ceaec0c30c2c3b2f1d86fd33d8bfe8c21c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.029690 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.046611 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.050872 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.050910 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.050923 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.050947 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.050962 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:27Z","lastTransitionTime":"2025-12-01T14:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.065292 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.081727 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.101329 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.120967 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.137596 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.151918 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.153710 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.153755 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.153768 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.153788 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.153803 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:27Z","lastTransitionTime":"2025-12-01T14:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.171270 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.187757 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.217719 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.242286 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.256522 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.256596 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.256628 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.256672 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.256707 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:27Z","lastTransitionTime":"2025-12-01T14:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.360380 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.360500 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.360539 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.360578 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.360603 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:27Z","lastTransitionTime":"2025-12-01T14:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.465687 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.466272 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.466293 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.466326 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.466347 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:27Z","lastTransitionTime":"2025-12-01T14:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.568380 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.568413 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.568421 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.568436 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.568444 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:27Z","lastTransitionTime":"2025-12-01T14:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.670942 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.670981 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.670992 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.671008 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.671020 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:27Z","lastTransitionTime":"2025-12-01T14:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.698698 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.699098 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.717174 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.729417 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.739915 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.760027 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f24941df3884bd7df557177fb9087ceaec0c30c2c3b2f1d86fd33d8bfe8c21c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.773634 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.773704 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.773728 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.773760 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.773782 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:27Z","lastTransitionTime":"2025-12-01T14:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.774654 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.787886 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.798316 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.806767 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.833307 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.845024 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.858634 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.869365 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.877946 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.878017 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.878037 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.878065 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.878122 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:27Z","lastTransitionTime":"2025-12-01T14:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.881087 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.902604 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.919633 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.939355 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:27Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.980530 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.980578 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.980589 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.980606 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:27 crc kubenswrapper[4810]: I1201 14:34:27.980616 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:27Z","lastTransitionTime":"2025-12-01T14:34:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.082494 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.082529 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.082537 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.082552 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.082560 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:28Z","lastTransitionTime":"2025-12-01T14:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.184761 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.184799 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.184809 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.184823 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.184833 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:28Z","lastTransitionTime":"2025-12-01T14:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.287358 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.287670 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.288673 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.288761 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.288819 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:28Z","lastTransitionTime":"2025-12-01T14:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.391999 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.392054 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.392067 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.392089 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.392115 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:28Z","lastTransitionTime":"2025-12-01T14:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.490919 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.491250 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.491369 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:28 crc kubenswrapper[4810]: E1201 14:34:28.491323 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:28 crc kubenswrapper[4810]: E1201 14:34:28.491802 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:28 crc kubenswrapper[4810]: E1201 14:34:28.492069 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.496367 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.496410 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.496420 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.496436 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.496446 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:28Z","lastTransitionTime":"2025-12-01T14:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.600184 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.600259 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.600280 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.600306 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.600324 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:28Z","lastTransitionTime":"2025-12-01T14:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.702080 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.702129 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.702140 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.702157 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.702168 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:28Z","lastTransitionTime":"2025-12-01T14:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.703211 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovnkube-controller/0.log" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.705762 4810 generic.go:334] "Generic (PLEG): container finished" podID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerID="f24941df3884bd7df557177fb9087ceaec0c30c2c3b2f1d86fd33d8bfe8c21c8" exitCode=1 Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.705835 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerDied","Data":"f24941df3884bd7df557177fb9087ceaec0c30c2c3b2f1d86fd33d8bfe8c21c8"} Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.706920 4810 scope.go:117] "RemoveContainer" containerID="f24941df3884bd7df557177fb9087ceaec0c30c2c3b2f1d86fd33d8bfe8c21c8" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.723037 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:28Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.738196 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:28Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.757443 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:28Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.773352 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:28Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.805131 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.805173 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.805184 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.805200 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.805211 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:28Z","lastTransitionTime":"2025-12-01T14:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.807045 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:28Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.825695 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:28Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.842193 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:28Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.860376 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:28Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.874551 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:28Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.892188 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:28Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.906057 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:28Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.907906 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.907950 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.907960 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.907977 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.907988 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:28Z","lastTransitionTime":"2025-12-01T14:34:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.920747 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:28Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.943350 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:28Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.964622 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:28Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.983413 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.983679 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:28 crc kubenswrapper[4810]: E1201 14:34:28.983760 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:34:44.983713529 +0000 UTC m=+50.747223172 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:34:28 crc kubenswrapper[4810]: E1201 14:34:28.983861 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 14:34:28 crc kubenswrapper[4810]: E1201 14:34:28.983994 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:44.983956844 +0000 UTC m=+50.747466477 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 14:34:28 crc kubenswrapper[4810]: I1201 14:34:28.990265 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f24941df3884bd7df557177fb9087ceaec0c30c2c3b2f1d86fd33d8bfe8c21c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f24941df3884bd7df557177fb9087ceaec0c30c2c3b2f1d86fd33d8bfe8c21c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:28Z\\\",\\\"message\\\":\\\"1 14:34:28.055865 6125 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 14:34:28.055920 6125 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 14:34:28.055983 6125 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 14:34:28.056117 6125 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 14:34:28.056373 6125 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1201 14:34:28.056682 6125 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 14:34:28.056707 6125 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 14:34:28.056736 6125 factory.go:656] Stopping watch factory\\\\nI1201 14:34:28.056741 6125 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 14:34:28.056768 6125 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 14:34:28.056785 6125 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:28Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.011249 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.011318 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.011331 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.011357 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.011371 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:29Z","lastTransitionTime":"2025-12-01T14:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.085309 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.085411 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.085465 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:29 crc kubenswrapper[4810]: E1201 14:34:29.085689 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 14:34:29 crc kubenswrapper[4810]: E1201 14:34:29.085838 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 14:34:29 crc kubenswrapper[4810]: E1201 14:34:29.085861 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 14:34:29 crc kubenswrapper[4810]: E1201 14:34:29.085951 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 14:34:29 crc kubenswrapper[4810]: E1201 14:34:29.085977 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:29 crc kubenswrapper[4810]: E1201 14:34:29.085878 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:45.085840048 +0000 UTC m=+50.849349691 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 14:34:29 crc kubenswrapper[4810]: E1201 14:34:29.085881 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 14:34:29 crc kubenswrapper[4810]: E1201 14:34:29.086126 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:29 crc kubenswrapper[4810]: E1201 14:34:29.086144 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:45.086098083 +0000 UTC m=+50.849607726 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:29 crc kubenswrapper[4810]: E1201 14:34:29.086266 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:45.086173525 +0000 UTC m=+50.849683388 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.113951 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.114004 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.114017 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.114034 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.114046 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:29Z","lastTransitionTime":"2025-12-01T14:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.216411 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.216466 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.216509 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.216532 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.216548 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:29Z","lastTransitionTime":"2025-12-01T14:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.318615 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.318649 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.318658 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.318671 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.318680 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:29Z","lastTransitionTime":"2025-12-01T14:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.422169 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.422229 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.422245 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.422274 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.422291 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:29Z","lastTransitionTime":"2025-12-01T14:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.525007 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.525047 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.525056 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.525069 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.525077 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:29Z","lastTransitionTime":"2025-12-01T14:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.627361 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.627401 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.627410 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.627447 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.627489 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:29Z","lastTransitionTime":"2025-12-01T14:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.712602 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovnkube-controller/0.log" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.716292 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerStarted","Data":"55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec"} Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.716533 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.730024 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.730083 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.730094 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.730112 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.730125 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:29Z","lastTransitionTime":"2025-12-01T14:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.741538 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.755130 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.777662 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.817378 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f24941df3884bd7df557177fb9087ceaec0c30c2c3b2f1d86fd33d8bfe8c21c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:28Z\\\",\\\"message\\\":\\\"1 14:34:28.055865 6125 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 14:34:28.055920 6125 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 14:34:28.055983 6125 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 14:34:28.056117 6125 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 14:34:28.056373 6125 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1201 14:34:28.056682 6125 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 14:34:28.056707 6125 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 14:34:28.056736 6125 factory.go:656] Stopping watch factory\\\\nI1201 14:34:28.056741 6125 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 14:34:28.056768 6125 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 14:34:28.056785 6125 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.832892 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.832946 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.832962 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.832982 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.832994 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:29Z","lastTransitionTime":"2025-12-01T14:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.835026 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.850317 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.864340 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.880044 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.910351 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.930906 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.935880 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.935911 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.935919 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.935935 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.935947 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:29Z","lastTransitionTime":"2025-12-01T14:34:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.947276 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.962794 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:29 crc kubenswrapper[4810]: I1201 14:34:29.984919 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.010954 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.027127 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.039200 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.039271 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.039291 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.039343 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.039364 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:30Z","lastTransitionTime":"2025-12-01T14:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.142521 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.142559 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.142570 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.142586 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.142597 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:30Z","lastTransitionTime":"2025-12-01T14:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.246049 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.246096 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.246107 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.246125 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.246137 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:30Z","lastTransitionTime":"2025-12-01T14:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.349152 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.349214 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.349225 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.349241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.349252 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:30Z","lastTransitionTime":"2025-12-01T14:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.382315 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.394811 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.405501 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.414391 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.428314 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.439450 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.451730 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.451770 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.451780 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.451798 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.451810 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:30Z","lastTransitionTime":"2025-12-01T14:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.491043 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.491069 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.491122 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:30 crc kubenswrapper[4810]: E1201 14:34:30.491540 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:30 crc kubenswrapper[4810]: E1201 14:34:30.491639 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:30 crc kubenswrapper[4810]: E1201 14:34:30.491642 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.493640 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.515572 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.533978 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.545530 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.554194 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.554229 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.554238 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.554251 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.554261 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:30Z","lastTransitionTime":"2025-12-01T14:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.563199 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f24941df3884bd7df557177fb9087ceaec0c30c2c3b2f1d86fd33d8bfe8c21c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:28Z\\\",\\\"message\\\":\\\"1 14:34:28.055865 6125 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 14:34:28.055920 6125 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 14:34:28.055983 6125 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 14:34:28.056117 6125 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 14:34:28.056373 6125 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1201 14:34:28.056682 6125 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 14:34:28.056707 6125 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 14:34:28.056736 6125 factory.go:656] Stopping watch factory\\\\nI1201 14:34:28.056741 6125 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 14:34:28.056768 6125 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 14:34:28.056785 6125 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.575011 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.586622 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.596581 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.608649 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.620379 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.656031 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.656064 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.656073 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.656086 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.656095 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:30Z","lastTransitionTime":"2025-12-01T14:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.720865 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovnkube-controller/1.log" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.721994 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovnkube-controller/0.log" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.727679 4810 generic.go:334] "Generic (PLEG): container finished" podID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerID="55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec" exitCode=1 Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.727715 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerDied","Data":"55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec"} Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.727749 4810 scope.go:117] "RemoveContainer" containerID="f24941df3884bd7df557177fb9087ceaec0c30c2c3b2f1d86fd33d8bfe8c21c8" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.728548 4810 scope.go:117] "RemoveContainer" containerID="55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec" Dec 01 14:34:30 crc kubenswrapper[4810]: E1201 14:34:30.728691 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.742444 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.754936 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.758378 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.758422 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.758437 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.758453 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.758464 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:30Z","lastTransitionTime":"2025-12-01T14:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.767759 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.778174 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.797769 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.810537 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.827965 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.846015 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.858576 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.861125 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.861197 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.861210 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.861229 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.861266 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:30Z","lastTransitionTime":"2025-12-01T14:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.874454 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.887229 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.904733 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.918132 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.936458 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.953926 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f24941df3884bd7df557177fb9087ceaec0c30c2c3b2f1d86fd33d8bfe8c21c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:28Z\\\",\\\"message\\\":\\\"1 14:34:28.055865 6125 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 14:34:28.055920 6125 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 14:34:28.055983 6125 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 14:34:28.056117 6125 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 14:34:28.056373 6125 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1201 14:34:28.056682 6125 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 14:34:28.056707 6125 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 14:34:28.056736 6125 factory.go:656] Stopping watch factory\\\\nI1201 14:34:28.056741 6125 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 14:34:28.056768 6125 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 14:34:28.056785 6125 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:29Z\\\",\\\"message\\\":\\\"ort:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1201 14:34:29.634576 6247 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nI1201 14:34:29.634605 6247 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nF1201 14:34:29.634613 6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:30Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.963666 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.963709 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.963721 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.963946 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:30 crc kubenswrapper[4810]: I1201 14:34:30.963975 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:30Z","lastTransitionTime":"2025-12-01T14:34:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.066151 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.066196 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.066209 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.066224 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.066234 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:31Z","lastTransitionTime":"2025-12-01T14:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.170552 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.170652 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.170670 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.170704 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.170725 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:31Z","lastTransitionTime":"2025-12-01T14:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.253660 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg"] Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.254329 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.257016 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.257253 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.273667 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.273731 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.273749 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.273777 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.273799 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:31Z","lastTransitionTime":"2025-12-01T14:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.278202 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.292590 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.306328 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/42ef3ab2-0451-4901-96d9-2d96faa37567-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mlglg\" (UID: \"42ef3ab2-0451-4901-96d9-2d96faa37567\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.306386 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/42ef3ab2-0451-4901-96d9-2d96faa37567-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mlglg\" (UID: \"42ef3ab2-0451-4901-96d9-2d96faa37567\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.306419 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpfv5\" (UniqueName: \"kubernetes.io/projected/42ef3ab2-0451-4901-96d9-2d96faa37567-kube-api-access-bpfv5\") pod \"ovnkube-control-plane-749d76644c-mlglg\" (UID: \"42ef3ab2-0451-4901-96d9-2d96faa37567\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.306453 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/42ef3ab2-0451-4901-96d9-2d96faa37567-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mlglg\" (UID: \"42ef3ab2-0451-4901-96d9-2d96faa37567\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.312124 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.334627 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f24941df3884bd7df557177fb9087ceaec0c30c2c3b2f1d86fd33d8bfe8c21c8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:28Z\\\",\\\"message\\\":\\\"1 14:34:28.055865 6125 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1201 14:34:28.055920 6125 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 14:34:28.055983 6125 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 14:34:28.056117 6125 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1201 14:34:28.056373 6125 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1201 14:34:28.056682 6125 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 14:34:28.056707 6125 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 14:34:28.056736 6125 factory.go:656] Stopping watch factory\\\\nI1201 14:34:28.056741 6125 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 14:34:28.056768 6125 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 14:34:28.056785 6125 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:29Z\\\",\\\"message\\\":\\\"ort:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1201 14:34:29.634576 6247 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nI1201 14:34:29.634605 6247 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nF1201 14:34:29.634613 6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.349859 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.363575 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.376778 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.376838 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.376853 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.376874 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.376886 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:31Z","lastTransitionTime":"2025-12-01T14:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.378196 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.392850 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.404324 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.406856 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpfv5\" (UniqueName: \"kubernetes.io/projected/42ef3ab2-0451-4901-96d9-2d96faa37567-kube-api-access-bpfv5\") pod \"ovnkube-control-plane-749d76644c-mlglg\" (UID: \"42ef3ab2-0451-4901-96d9-2d96faa37567\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.406894 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/42ef3ab2-0451-4901-96d9-2d96faa37567-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mlglg\" (UID: \"42ef3ab2-0451-4901-96d9-2d96faa37567\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.406910 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/42ef3ab2-0451-4901-96d9-2d96faa37567-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mlglg\" (UID: \"42ef3ab2-0451-4901-96d9-2d96faa37567\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.406943 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/42ef3ab2-0451-4901-96d9-2d96faa37567-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mlglg\" (UID: \"42ef3ab2-0451-4901-96d9-2d96faa37567\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.407639 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/42ef3ab2-0451-4901-96d9-2d96faa37567-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-mlglg\" (UID: \"42ef3ab2-0451-4901-96d9-2d96faa37567\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.407700 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/42ef3ab2-0451-4901-96d9-2d96faa37567-env-overrides\") pod \"ovnkube-control-plane-749d76644c-mlglg\" (UID: \"42ef3ab2-0451-4901-96d9-2d96faa37567\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.417012 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/42ef3ab2-0451-4901-96d9-2d96faa37567-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-mlglg\" (UID: \"42ef3ab2-0451-4901-96d9-2d96faa37567\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.421227 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.423688 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpfv5\" (UniqueName: \"kubernetes.io/projected/42ef3ab2-0451-4901-96d9-2d96faa37567-kube-api-access-bpfv5\") pod \"ovnkube-control-plane-749d76644c-mlglg\" (UID: \"42ef3ab2-0451-4901-96d9-2d96faa37567\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.434841 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.455208 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ef3ab2-0451-4901-96d9-2d96faa37567\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mlglg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.476330 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.479794 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.479867 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.479898 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.479932 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.479952 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:31Z","lastTransitionTime":"2025-12-01T14:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.490158 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.503423 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.513263 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.567315 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.582859 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.582907 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.582921 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.582940 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.582953 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:31Z","lastTransitionTime":"2025-12-01T14:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.687961 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.688022 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.688062 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.688106 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.688128 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:31Z","lastTransitionTime":"2025-12-01T14:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.735662 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovnkube-controller/1.log" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.739890 4810 scope.go:117] "RemoveContainer" containerID="55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec" Dec 01 14:34:31 crc kubenswrapper[4810]: E1201 14:34:31.740016 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.741594 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" event={"ID":"42ef3ab2-0451-4901-96d9-2d96faa37567","Type":"ContainerStarted","Data":"c0ff734541250c2a3987c49035f25ca9dc95e0ab32c752266479cab77d523e62"} Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.753515 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ef3ab2-0451-4901-96d9-2d96faa37567\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mlglg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.781969 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.790980 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.791035 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.791051 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.791077 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.791093 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:31Z","lastTransitionTime":"2025-12-01T14:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.797211 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.813306 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.829903 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.843074 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.859922 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.871006 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.886236 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.900011 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.900264 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.900308 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.900323 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.900345 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.900358 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:31Z","lastTransitionTime":"2025-12-01T14:34:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.915853 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.940977 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:29Z\\\",\\\"message\\\":\\\"ort:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1201 14:34:29.634576 6247 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nI1201 14:34:29.634605 6247 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nF1201 14:34:29.634613 6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.958735 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.970243 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.983072 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:31 crc kubenswrapper[4810]: I1201 14:34:31.996374 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:31Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.003553 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.003591 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.003599 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.003617 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.003628 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:32Z","lastTransitionTime":"2025-12-01T14:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.106423 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.106461 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.106487 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.106506 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.106516 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:32Z","lastTransitionTime":"2025-12-01T14:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.209164 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.209194 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.209202 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.209214 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.209222 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:32Z","lastTransitionTime":"2025-12-01T14:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.312426 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.312526 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.312539 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.312556 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.312568 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:32Z","lastTransitionTime":"2025-12-01T14:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.416011 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.416067 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.416083 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.416113 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.416135 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:32Z","lastTransitionTime":"2025-12-01T14:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.490591 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:32 crc kubenswrapper[4810]: E1201 14:34:32.490763 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.490593 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.490805 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:32 crc kubenswrapper[4810]: E1201 14:34:32.490881 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:32 crc kubenswrapper[4810]: E1201 14:34:32.491235 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.519281 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.519342 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.519359 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.519384 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.519402 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:32Z","lastTransitionTime":"2025-12-01T14:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.622901 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.622978 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.623002 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.623030 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.623051 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:32Z","lastTransitionTime":"2025-12-01T14:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.726649 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.726729 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.726755 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.726792 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.726818 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:32Z","lastTransitionTime":"2025-12-01T14:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.755291 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" event={"ID":"42ef3ab2-0451-4901-96d9-2d96faa37567","Type":"ContainerStarted","Data":"f6ada971d2ea2d5b690fbc2086938a74afe6d886f3535687706f46b2425e3116"} Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.755385 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" event={"ID":"42ef3ab2-0451-4901-96d9-2d96faa37567","Type":"ContainerStarted","Data":"da8909387cd335182759cbebab25a30186b86ad060eccd57e825de4d2dcda211"} Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.763027 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-jlwwg"] Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.763889 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:32 crc kubenswrapper[4810]: E1201 14:34:32.763993 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.785366 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:32Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.802383 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:32Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.818830 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:32Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.829838 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.829906 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.829989 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.830071 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.830098 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:32Z","lastTransitionTime":"2025-12-01T14:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.835231 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:32Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.856904 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:32Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.872942 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:32Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.887160 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:32Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.904551 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:32Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.918071 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:32Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.918900 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djdpf\" (UniqueName: \"kubernetes.io/projected/d379bcb2-a35d-470f-894c-8629d5023df8-kube-api-access-djdpf\") pod \"network-metrics-daemon-jlwwg\" (UID: \"d379bcb2-a35d-470f-894c-8629d5023df8\") " pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.918947 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs\") pod \"network-metrics-daemon-jlwwg\" (UID: \"d379bcb2-a35d-470f-894c-8629d5023df8\") " pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.932341 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.932378 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.932391 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.932408 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.932420 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:32Z","lastTransitionTime":"2025-12-01T14:34:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.934992 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:32Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.948619 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:32Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.961244 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ef3ab2-0451-4901-96d9-2d96faa37567\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8909387cd335182759cbebab25a30186b86ad060eccd57e825de4d2dcda211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ada971d2ea2d5b690fbc2086938a74afe6d886f3535687706f46b2425e3116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mlglg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:32Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.977551 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:32Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:32 crc kubenswrapper[4810]: I1201 14:34:32.990339 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:32Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.002736 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.019759 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djdpf\" (UniqueName: \"kubernetes.io/projected/d379bcb2-a35d-470f-894c-8629d5023df8-kube-api-access-djdpf\") pod \"network-metrics-daemon-jlwwg\" (UID: \"d379bcb2-a35d-470f-894c-8629d5023df8\") " pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.019800 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs\") pod \"network-metrics-daemon-jlwwg\" (UID: \"d379bcb2-a35d-470f-894c-8629d5023df8\") " pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:33 crc kubenswrapper[4810]: E1201 14:34:33.019901 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 14:34:33 crc kubenswrapper[4810]: E1201 14:34:33.019944 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs podName:d379bcb2-a35d-470f-894c-8629d5023df8 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:33.519932444 +0000 UTC m=+39.283442047 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs") pod "network-metrics-daemon-jlwwg" (UID: "d379bcb2-a35d-470f-894c-8629d5023df8") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.020797 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:29Z\\\",\\\"message\\\":\\\"ort:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1201 14:34:29.634576 6247 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nI1201 14:34:29.634605 6247 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nF1201 14:34:29.634613 6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.035146 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.035278 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.035307 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.035318 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.035336 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.035347 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:33Z","lastTransitionTime":"2025-12-01T14:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.045889 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djdpf\" (UniqueName: \"kubernetes.io/projected/d379bcb2-a35d-470f-894c-8629d5023df8-kube-api-access-djdpf\") pod \"network-metrics-daemon-jlwwg\" (UID: \"d379bcb2-a35d-470f-894c-8629d5023df8\") " pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.050876 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.065129 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.077710 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.091069 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.101281 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.116527 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ef3ab2-0451-4901-96d9-2d96faa37567\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8909387cd335182759cbebab25a30186b86ad060eccd57e825de4d2dcda211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ada971d2ea2d5b690fbc2086938a74afe6d886f3535687706f46b2425e3116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mlglg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.138054 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.138092 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.138104 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.138122 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.138135 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:33Z","lastTransitionTime":"2025-12-01T14:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.141830 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.160018 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.176386 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.189868 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.213513 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:29Z\\\",\\\"message\\\":\\\"ort:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1201 14:34:29.634576 6247 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nI1201 14:34:29.634605 6247 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nF1201 14:34:29.634613 6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.226591 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jlwwg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d379bcb2-a35d-470f-894c-8629d5023df8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jlwwg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.240539 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.240568 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.240580 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.240597 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.240610 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:33Z","lastTransitionTime":"2025-12-01T14:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.243298 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.261593 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.274414 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.289606 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:33Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.343283 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.343365 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.343383 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.343412 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.343431 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:33Z","lastTransitionTime":"2025-12-01T14:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.446896 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.446973 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.446992 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.447021 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.447043 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:33Z","lastTransitionTime":"2025-12-01T14:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.524714 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs\") pod \"network-metrics-daemon-jlwwg\" (UID: \"d379bcb2-a35d-470f-894c-8629d5023df8\") " pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:33 crc kubenswrapper[4810]: E1201 14:34:33.524889 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 14:34:33 crc kubenswrapper[4810]: E1201 14:34:33.525200 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs podName:d379bcb2-a35d-470f-894c-8629d5023df8 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:34.525165791 +0000 UTC m=+40.288675434 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs") pod "network-metrics-daemon-jlwwg" (UID: "d379bcb2-a35d-470f-894c-8629d5023df8") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.550797 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.550866 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.550890 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.550923 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.550950 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:33Z","lastTransitionTime":"2025-12-01T14:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.654922 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.655728 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.656095 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.656352 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.656607 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:33Z","lastTransitionTime":"2025-12-01T14:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.760018 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.760182 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.760208 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.760240 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.760265 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:33Z","lastTransitionTime":"2025-12-01T14:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.866851 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.867567 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.867745 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.867896 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.868043 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:33Z","lastTransitionTime":"2025-12-01T14:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.973506 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.973551 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.973563 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.973584 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:33 crc kubenswrapper[4810]: I1201 14:34:33.973596 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:33Z","lastTransitionTime":"2025-12-01T14:34:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.076935 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.076991 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.077004 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.077025 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.077040 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:34Z","lastTransitionTime":"2025-12-01T14:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.180785 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.180876 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.180895 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.180928 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.180953 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:34Z","lastTransitionTime":"2025-12-01T14:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.283602 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.283656 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.283672 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.283699 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.283716 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:34Z","lastTransitionTime":"2025-12-01T14:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.386638 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.386692 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.386706 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.386728 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.386743 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:34Z","lastTransitionTime":"2025-12-01T14:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.489386 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.489418 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.489428 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.489443 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.489454 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:34Z","lastTransitionTime":"2025-12-01T14:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.490412 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.490436 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.490461 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:34 crc kubenswrapper[4810]: E1201 14:34:34.491102 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.490521 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:34 crc kubenswrapper[4810]: E1201 14:34:34.491318 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:34 crc kubenswrapper[4810]: E1201 14:34:34.491445 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:34 crc kubenswrapper[4810]: E1201 14:34:34.491734 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.508515 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:34Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.520031 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:34Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.533848 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:34Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.537045 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs\") pod \"network-metrics-daemon-jlwwg\" (UID: \"d379bcb2-a35d-470f-894c-8629d5023df8\") " pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:34 crc kubenswrapper[4810]: E1201 14:34:34.537269 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 14:34:34 crc kubenswrapper[4810]: E1201 14:34:34.537372 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs podName:d379bcb2-a35d-470f-894c-8629d5023df8 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:36.537350036 +0000 UTC m=+42.300859649 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs") pod "network-metrics-daemon-jlwwg" (UID: "d379bcb2-a35d-470f-894c-8629d5023df8") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.556857 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:34Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.573264 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:34Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.592701 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ef3ab2-0451-4901-96d9-2d96faa37567\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8909387cd335182759cbebab25a30186b86ad060eccd57e825de4d2dcda211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ada971d2ea2d5b690fbc2086938a74afe6d886f3535687706f46b2425e3116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mlglg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:34Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.594420 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.594525 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.594544 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.594574 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.594594 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:34Z","lastTransitionTime":"2025-12-01T14:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.631588 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:34Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.653341 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:34Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.681352 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:34Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.697910 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.697977 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.697995 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.698024 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.698206 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:34Z","lastTransitionTime":"2025-12-01T14:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.705853 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:34Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.731097 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:29Z\\\",\\\"message\\\":\\\"ort:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1201 14:34:29.634576 6247 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nI1201 14:34:29.634605 6247 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nF1201 14:34:29.634613 6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:34Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.747323 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jlwwg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d379bcb2-a35d-470f-894c-8629d5023df8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jlwwg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:34Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.764356 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:34Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.782310 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:34Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.802160 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.802244 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.802273 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.802304 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.802328 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:34Z","lastTransitionTime":"2025-12-01T14:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.802157 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:34Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.826354 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:34Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.847372 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:34Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.905805 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.905884 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.905903 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.905936 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:34 crc kubenswrapper[4810]: I1201 14:34:34.905954 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:34Z","lastTransitionTime":"2025-12-01T14:34:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.009535 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.009607 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.009627 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.009655 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.009676 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:35Z","lastTransitionTime":"2025-12-01T14:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.112995 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.113077 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.113096 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.113124 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.113144 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:35Z","lastTransitionTime":"2025-12-01T14:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.216626 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.216684 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.216698 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.216719 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.216735 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:35Z","lastTransitionTime":"2025-12-01T14:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.320125 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.320185 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.320241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.320462 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.320529 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:35Z","lastTransitionTime":"2025-12-01T14:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.437880 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.437912 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.437923 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.437939 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.437951 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:35Z","lastTransitionTime":"2025-12-01T14:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.542620 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.542693 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.542716 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.542744 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.542775 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:35Z","lastTransitionTime":"2025-12-01T14:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.646252 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.646331 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.646353 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.646383 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.646401 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:35Z","lastTransitionTime":"2025-12-01T14:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.749253 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.749353 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.749379 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.749403 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.749422 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:35Z","lastTransitionTime":"2025-12-01T14:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.853121 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.853212 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.853244 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.853287 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.853314 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:35Z","lastTransitionTime":"2025-12-01T14:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.957735 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.957780 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.957792 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.957808 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:35 crc kubenswrapper[4810]: I1201 14:34:35.957820 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:35Z","lastTransitionTime":"2025-12-01T14:34:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.060411 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.060543 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.060566 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.060591 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.060610 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:36Z","lastTransitionTime":"2025-12-01T14:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.163964 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.164392 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.164598 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.164774 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.164966 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:36Z","lastTransitionTime":"2025-12-01T14:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.268600 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.268667 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.268685 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.268710 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.269334 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:36Z","lastTransitionTime":"2025-12-01T14:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.373894 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.373960 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.373982 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.374010 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.374047 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:36Z","lastTransitionTime":"2025-12-01T14:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.476367 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.476442 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.476467 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.476536 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.476563 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:36Z","lastTransitionTime":"2025-12-01T14:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.490112 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.490137 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.490215 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:36 crc kubenswrapper[4810]: E1201 14:34:36.490397 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.490540 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:36 crc kubenswrapper[4810]: E1201 14:34:36.490840 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:36 crc kubenswrapper[4810]: E1201 14:34:36.491099 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:36 crc kubenswrapper[4810]: E1201 14:34:36.492617 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.562888 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs\") pod \"network-metrics-daemon-jlwwg\" (UID: \"d379bcb2-a35d-470f-894c-8629d5023df8\") " pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:36 crc kubenswrapper[4810]: E1201 14:34:36.563196 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 14:34:36 crc kubenswrapper[4810]: E1201 14:34:36.563369 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs podName:d379bcb2-a35d-470f-894c-8629d5023df8 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:40.5633318 +0000 UTC m=+46.326841433 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs") pod "network-metrics-daemon-jlwwg" (UID: "d379bcb2-a35d-470f-894c-8629d5023df8") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.579065 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.579166 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.579194 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.579229 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.579254 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:36Z","lastTransitionTime":"2025-12-01T14:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.682238 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.682302 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.682320 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.682346 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.682367 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:36Z","lastTransitionTime":"2025-12-01T14:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.786091 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.786141 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.786165 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.786191 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.786210 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:36Z","lastTransitionTime":"2025-12-01T14:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.889939 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.889986 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.889998 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.890019 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.890031 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:36Z","lastTransitionTime":"2025-12-01T14:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.993306 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.993350 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.993363 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.993385 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:36 crc kubenswrapper[4810]: I1201 14:34:36.993397 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:36Z","lastTransitionTime":"2025-12-01T14:34:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.096510 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.096541 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.096552 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.096568 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.096581 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:37Z","lastTransitionTime":"2025-12-01T14:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.170320 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.170381 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.170398 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.170422 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.170456 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:37Z","lastTransitionTime":"2025-12-01T14:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:37 crc kubenswrapper[4810]: E1201 14:34:37.188042 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:37Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.192338 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.192372 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.192382 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.192400 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.192412 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:37Z","lastTransitionTime":"2025-12-01T14:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:37 crc kubenswrapper[4810]: E1201 14:34:37.211202 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:37Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.215803 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.215955 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.215985 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.216021 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.216043 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:37Z","lastTransitionTime":"2025-12-01T14:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:37 crc kubenswrapper[4810]: E1201 14:34:37.233614 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:37Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.238029 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.238078 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.238094 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.238112 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.238129 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:37Z","lastTransitionTime":"2025-12-01T14:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:37 crc kubenswrapper[4810]: E1201 14:34:37.257452 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:37Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.263222 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.263316 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.263340 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.263370 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.263390 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:37Z","lastTransitionTime":"2025-12-01T14:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:37 crc kubenswrapper[4810]: E1201 14:34:37.287751 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:37Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:37 crc kubenswrapper[4810]: E1201 14:34:37.287976 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.290899 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.290973 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.291030 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.291060 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.291078 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:37Z","lastTransitionTime":"2025-12-01T14:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.393900 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.394018 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.394046 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.394110 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.394137 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:37Z","lastTransitionTime":"2025-12-01T14:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.497928 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.498004 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.498013 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.498029 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.498042 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:37Z","lastTransitionTime":"2025-12-01T14:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.602360 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.602446 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.602465 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.602535 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.602556 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:37Z","lastTransitionTime":"2025-12-01T14:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.610562 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.612097 4810 scope.go:117] "RemoveContainer" containerID="55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec" Dec 01 14:34:37 crc kubenswrapper[4810]: E1201 14:34:37.612395 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.706933 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.707066 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.707086 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.707117 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.707138 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:37Z","lastTransitionTime":"2025-12-01T14:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.810268 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.810333 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.810350 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.810376 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.810393 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:37Z","lastTransitionTime":"2025-12-01T14:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.915180 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.915287 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.915318 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.915360 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:37 crc kubenswrapper[4810]: I1201 14:34:37.915386 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:37Z","lastTransitionTime":"2025-12-01T14:34:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.017540 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.017621 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.017636 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.017664 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.017683 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:38Z","lastTransitionTime":"2025-12-01T14:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.120302 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.120349 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.120363 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.120379 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.120389 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:38Z","lastTransitionTime":"2025-12-01T14:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.223009 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.223083 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.223107 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.223135 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.223153 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:38Z","lastTransitionTime":"2025-12-01T14:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.326111 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.326190 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.326214 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.326245 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.326271 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:38Z","lastTransitionTime":"2025-12-01T14:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.428736 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.428787 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.428800 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.428820 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.428833 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:38Z","lastTransitionTime":"2025-12-01T14:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.490521 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.490633 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:38 crc kubenswrapper[4810]: E1201 14:34:38.490692 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.490772 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.490960 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:38 crc kubenswrapper[4810]: E1201 14:34:38.490958 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:38 crc kubenswrapper[4810]: E1201 14:34:38.491115 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:34:38 crc kubenswrapper[4810]: E1201 14:34:38.491201 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.541149 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.541215 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.541238 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.541267 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.541290 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:38Z","lastTransitionTime":"2025-12-01T14:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.644342 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.644378 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.644389 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.644407 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.644419 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:38Z","lastTransitionTime":"2025-12-01T14:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.747741 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.747776 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.747783 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.747798 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.747809 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:38Z","lastTransitionTime":"2025-12-01T14:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.850912 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.850979 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.850997 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.851024 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.851044 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:38Z","lastTransitionTime":"2025-12-01T14:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.954236 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.954296 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.954318 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.954343 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:38 crc kubenswrapper[4810]: I1201 14:34:38.954361 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:38Z","lastTransitionTime":"2025-12-01T14:34:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.057607 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.057687 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.057704 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.057727 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.057744 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:39Z","lastTransitionTime":"2025-12-01T14:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.160698 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.160800 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.160819 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.160843 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.160860 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:39Z","lastTransitionTime":"2025-12-01T14:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.264255 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.264318 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.264335 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.264361 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.264380 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:39Z","lastTransitionTime":"2025-12-01T14:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.368067 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.368153 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.368176 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.368212 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.368231 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:39Z","lastTransitionTime":"2025-12-01T14:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.471335 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.471398 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.471421 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.471456 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.471508 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:39Z","lastTransitionTime":"2025-12-01T14:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.574335 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.574555 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.574591 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.574660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.574681 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:39Z","lastTransitionTime":"2025-12-01T14:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.677793 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.677851 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.677889 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.677915 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.677933 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:39Z","lastTransitionTime":"2025-12-01T14:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.780968 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.781046 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.781066 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.781101 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.781125 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:39Z","lastTransitionTime":"2025-12-01T14:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.884402 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.884466 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.884514 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.884538 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.884556 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:39Z","lastTransitionTime":"2025-12-01T14:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.987974 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.988046 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.988081 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.988112 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:39 crc kubenswrapper[4810]: I1201 14:34:39.988134 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:39Z","lastTransitionTime":"2025-12-01T14:34:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.091955 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.092020 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.092035 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.092058 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.092071 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:40Z","lastTransitionTime":"2025-12-01T14:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.195271 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.195416 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.195440 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.195464 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.195509 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:40Z","lastTransitionTime":"2025-12-01T14:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.297789 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.297844 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.297860 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.297879 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.297892 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:40Z","lastTransitionTime":"2025-12-01T14:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.400574 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.400619 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.400679 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.400709 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.400726 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:40Z","lastTransitionTime":"2025-12-01T14:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.491766 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.491816 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.491818 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:40 crc kubenswrapper[4810]: E1201 14:34:40.491914 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.491956 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:40 crc kubenswrapper[4810]: E1201 14:34:40.492031 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:40 crc kubenswrapper[4810]: E1201 14:34:40.492091 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:40 crc kubenswrapper[4810]: E1201 14:34:40.492605 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.503543 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.503584 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.503595 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.503613 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.503627 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:40Z","lastTransitionTime":"2025-12-01T14:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.606623 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.606776 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.606808 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.606848 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.606876 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:40Z","lastTransitionTime":"2025-12-01T14:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.613451 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs\") pod \"network-metrics-daemon-jlwwg\" (UID: \"d379bcb2-a35d-470f-894c-8629d5023df8\") " pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:40 crc kubenswrapper[4810]: E1201 14:34:40.613626 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 14:34:40 crc kubenswrapper[4810]: E1201 14:34:40.613691 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs podName:d379bcb2-a35d-470f-894c-8629d5023df8 nodeName:}" failed. No retries permitted until 2025-12-01 14:34:48.613672153 +0000 UTC m=+54.377181766 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs") pod "network-metrics-daemon-jlwwg" (UID: "d379bcb2-a35d-470f-894c-8629d5023df8") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.711167 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.711267 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.711292 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.711328 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.711354 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:40Z","lastTransitionTime":"2025-12-01T14:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.814397 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.814437 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.814454 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.814524 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.814548 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:40Z","lastTransitionTime":"2025-12-01T14:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.918249 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.918309 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.918324 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.918368 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.918387 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:40Z","lastTransitionTime":"2025-12-01T14:34:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.942238 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.955421 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.967785 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:40Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:40 crc kubenswrapper[4810]: I1201 14:34:40.988833 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:40Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.008055 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:41Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.020991 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.021037 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.021048 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.021068 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.021085 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:41Z","lastTransitionTime":"2025-12-01T14:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.041202 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:29Z\\\",\\\"message\\\":\\\"ort:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1201 14:34:29.634576 6247 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nI1201 14:34:29.634605 6247 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nF1201 14:34:29.634613 6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:41Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.056750 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jlwwg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d379bcb2-a35d-470f-894c-8629d5023df8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jlwwg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:41Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.079977 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:41Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.105438 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:41Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.125042 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.125127 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.125147 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.125182 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.125204 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:41Z","lastTransitionTime":"2025-12-01T14:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.125690 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:41Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.150603 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:41Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.170674 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:41Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.191893 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:41Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.209019 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:41Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.221526 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:41Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.228281 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.228338 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.228362 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.228395 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.228418 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:41Z","lastTransitionTime":"2025-12-01T14:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.240518 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:41Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.257296 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:41Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.271621 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ef3ab2-0451-4901-96d9-2d96faa37567\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8909387cd335182759cbebab25a30186b86ad060eccd57e825de4d2dcda211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ada971d2ea2d5b690fbc2086938a74afe6d886f3535687706f46b2425e3116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mlglg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:41Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.297741 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:41Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.331661 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.331708 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.331719 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.331736 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.331747 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:41Z","lastTransitionTime":"2025-12-01T14:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.434890 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.434957 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.434976 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.435005 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.435025 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:41Z","lastTransitionTime":"2025-12-01T14:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.537730 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.537778 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.537790 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.537822 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.537847 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:41Z","lastTransitionTime":"2025-12-01T14:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.641179 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.641248 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.641269 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.641298 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.641323 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:41Z","lastTransitionTime":"2025-12-01T14:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.746363 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.746410 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.746421 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.746438 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.746450 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:41Z","lastTransitionTime":"2025-12-01T14:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.850440 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.850536 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.850554 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.850580 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.850598 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:41Z","lastTransitionTime":"2025-12-01T14:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.953629 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.953691 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.953711 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.953738 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:41 crc kubenswrapper[4810]: I1201 14:34:41.953756 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:41Z","lastTransitionTime":"2025-12-01T14:34:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.056835 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.056917 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.056955 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.056993 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.057011 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:42Z","lastTransitionTime":"2025-12-01T14:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.160000 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.160070 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.160094 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.160150 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.160174 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:42Z","lastTransitionTime":"2025-12-01T14:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.263855 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.263901 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.263912 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.263929 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.263941 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:42Z","lastTransitionTime":"2025-12-01T14:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.366848 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.366896 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.366907 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.366926 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.366939 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:42Z","lastTransitionTime":"2025-12-01T14:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.468985 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.469025 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.469035 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.469050 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.469059 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:42Z","lastTransitionTime":"2025-12-01T14:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.490412 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.490462 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.490569 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:42 crc kubenswrapper[4810]: E1201 14:34:42.490739 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.490759 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:42 crc kubenswrapper[4810]: E1201 14:34:42.490857 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:34:42 crc kubenswrapper[4810]: E1201 14:34:42.491074 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:42 crc kubenswrapper[4810]: E1201 14:34:42.491126 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.572375 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.572509 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.572535 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.572562 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.572582 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:42Z","lastTransitionTime":"2025-12-01T14:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.675062 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.675139 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.675162 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.675202 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.675227 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:42Z","lastTransitionTime":"2025-12-01T14:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.778221 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.778289 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.778308 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.778333 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.778353 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:42Z","lastTransitionTime":"2025-12-01T14:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.880719 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.880768 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.880778 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.880793 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.880802 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:42Z","lastTransitionTime":"2025-12-01T14:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.983978 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.984042 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.984064 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.984090 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:42 crc kubenswrapper[4810]: I1201 14:34:42.984110 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:42Z","lastTransitionTime":"2025-12-01T14:34:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.087506 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.087563 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.087580 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.087605 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.087623 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:43Z","lastTransitionTime":"2025-12-01T14:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.191520 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.191571 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.191585 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.191604 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.191616 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:43Z","lastTransitionTime":"2025-12-01T14:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.294159 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.294204 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.294214 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.294228 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.294237 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:43Z","lastTransitionTime":"2025-12-01T14:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.397228 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.397273 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.397285 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.397301 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.397314 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:43Z","lastTransitionTime":"2025-12-01T14:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.501047 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.501111 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.501133 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.501157 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.501220 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:43Z","lastTransitionTime":"2025-12-01T14:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.605456 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.605612 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.605630 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.605656 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.605703 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:43Z","lastTransitionTime":"2025-12-01T14:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.708690 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.708741 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.708754 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.708774 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.708786 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:43Z","lastTransitionTime":"2025-12-01T14:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.811651 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.811712 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.811729 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.811757 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.811780 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:43Z","lastTransitionTime":"2025-12-01T14:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.914763 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.914857 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.914881 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.914919 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:43 crc kubenswrapper[4810]: I1201 14:34:43.914945 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:43Z","lastTransitionTime":"2025-12-01T14:34:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.018310 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.018367 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.018381 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.018406 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.018430 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:44Z","lastTransitionTime":"2025-12-01T14:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.122781 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.122898 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.122923 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.122962 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.122987 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:44Z","lastTransitionTime":"2025-12-01T14:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.226916 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.227020 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.227046 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.227085 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.227106 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:44Z","lastTransitionTime":"2025-12-01T14:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.331391 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.331437 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.331455 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.331527 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.331554 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:44Z","lastTransitionTime":"2025-12-01T14:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.435630 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.435703 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.435726 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.435755 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.435779 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:44Z","lastTransitionTime":"2025-12-01T14:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.490443 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:44 crc kubenswrapper[4810]: E1201 14:34:44.490611 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.491160 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:44 crc kubenswrapper[4810]: E1201 14:34:44.491244 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.491299 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:44 crc kubenswrapper[4810]: E1201 14:34:44.491360 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.491447 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:44 crc kubenswrapper[4810]: E1201 14:34:44.491550 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.511017 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bd862d7-0562-4f45-a529-96b05cc7b1c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46594b36eb97f9787facc7ee07e0d3e2aecd45816488438a8af9e619ff37aeb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f7016800cb62ec155217c88e1d0968f86b6f807ab7a1e410579713f15741a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b24a67bb57bd282f48b4185451f120aaebab8373d0e368a70c97596b53a95213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:44Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.536984 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:44Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.539764 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.539868 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.539928 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.539962 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.540035 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:44Z","lastTransitionTime":"2025-12-01T14:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.558684 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jlwwg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d379bcb2-a35d-470f-894c-8629d5023df8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jlwwg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:44Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.574342 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:44Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.589821 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:44Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.613134 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:29Z\\\",\\\"message\\\":\\\"ort:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1201 14:34:29.634576 6247 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nI1201 14:34:29.634605 6247 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nF1201 14:34:29.634613 6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:44Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.628917 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:44Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.642730 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.643040 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.643169 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.643150 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:44Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.643491 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.643623 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:44Z","lastTransitionTime":"2025-12-01T14:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.658801 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:44Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.673105 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:44Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.686300 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:44Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.703462 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:44Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.714307 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:44Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.731930 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ef3ab2-0451-4901-96d9-2d96faa37567\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8909387cd335182759cbebab25a30186b86ad060eccd57e825de4d2dcda211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ada971d2ea2d5b690fbc2086938a74afe6d886f3535687706f46b2425e3116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mlglg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:44Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.746952 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.747019 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.747040 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.747066 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.747085 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:44Z","lastTransitionTime":"2025-12-01T14:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.759322 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:44Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.784822 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:44Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.805803 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:44Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.821199 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:44Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.850213 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.850254 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.850269 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.850290 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.850304 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:44Z","lastTransitionTime":"2025-12-01T14:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.953419 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.953560 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.953596 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.953629 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:44 crc kubenswrapper[4810]: I1201 14:34:44.953653 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:44Z","lastTransitionTime":"2025-12-01T14:34:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.056715 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.056755 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.056766 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.056784 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.056796 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:45Z","lastTransitionTime":"2025-12-01T14:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.070725 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.070876 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:45 crc kubenswrapper[4810]: E1201 14:34:45.070983 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:35:17.070963693 +0000 UTC m=+82.834473306 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:34:45 crc kubenswrapper[4810]: E1201 14:34:45.071107 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 14:34:45 crc kubenswrapper[4810]: E1201 14:34:45.071230 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 14:35:17.071200799 +0000 UTC m=+82.834710422 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.159775 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.159815 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.159827 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.159845 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.159857 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:45Z","lastTransitionTime":"2025-12-01T14:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.172507 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.172592 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.172658 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:45 crc kubenswrapper[4810]: E1201 14:34:45.172684 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 14:34:45 crc kubenswrapper[4810]: E1201 14:34:45.172750 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 14:35:17.172729724 +0000 UTC m=+82.936239347 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 14:34:45 crc kubenswrapper[4810]: E1201 14:34:45.172800 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 14:34:45 crc kubenswrapper[4810]: E1201 14:34:45.172830 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 14:34:45 crc kubenswrapper[4810]: E1201 14:34:45.172846 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:45 crc kubenswrapper[4810]: E1201 14:34:45.172877 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 14:34:45 crc kubenswrapper[4810]: E1201 14:34:45.172895 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 14:35:17.172879538 +0000 UTC m=+82.936389151 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:45 crc kubenswrapper[4810]: E1201 14:34:45.172910 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 14:34:45 crc kubenswrapper[4810]: E1201 14:34:45.172935 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:45 crc kubenswrapper[4810]: E1201 14:34:45.173031 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 14:35:17.173003841 +0000 UTC m=+82.936513504 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.262458 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.262549 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.262574 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.262607 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.262630 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:45Z","lastTransitionTime":"2025-12-01T14:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.365370 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.365429 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.365447 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.365505 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.365541 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:45Z","lastTransitionTime":"2025-12-01T14:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.469315 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.469384 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.469407 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.469436 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.469460 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:45Z","lastTransitionTime":"2025-12-01T14:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.572962 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.573028 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.573048 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.573074 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.573092 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:45Z","lastTransitionTime":"2025-12-01T14:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.676556 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.676888 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.677234 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.677517 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.677669 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:45Z","lastTransitionTime":"2025-12-01T14:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.779952 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.780207 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.780278 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.780361 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.780424 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:45Z","lastTransitionTime":"2025-12-01T14:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.882973 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.883013 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.883026 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.883043 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.883056 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:45Z","lastTransitionTime":"2025-12-01T14:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.985489 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.985684 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.985793 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.985868 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:45 crc kubenswrapper[4810]: I1201 14:34:45.985954 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:45Z","lastTransitionTime":"2025-12-01T14:34:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.088040 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.088077 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.088087 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.088101 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.088111 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:46Z","lastTransitionTime":"2025-12-01T14:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.190083 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.190116 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.190128 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.190143 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.190154 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:46Z","lastTransitionTime":"2025-12-01T14:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.291956 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.292004 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.292020 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.292040 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.292053 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:46Z","lastTransitionTime":"2025-12-01T14:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.395219 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.395266 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.395279 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.395298 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.395309 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:46Z","lastTransitionTime":"2025-12-01T14:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.491152 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:46 crc kubenswrapper[4810]: E1201 14:34:46.491392 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.491745 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.491923 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.491881 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:46 crc kubenswrapper[4810]: E1201 14:34:46.492352 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:46 crc kubenswrapper[4810]: E1201 14:34:46.492533 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:46 crc kubenswrapper[4810]: E1201 14:34:46.492652 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.497776 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.498679 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.498737 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.498796 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.498824 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:46Z","lastTransitionTime":"2025-12-01T14:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.603022 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.603654 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.603704 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.603738 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.603760 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:46Z","lastTransitionTime":"2025-12-01T14:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.707047 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.707103 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.707123 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.707147 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.707165 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:46Z","lastTransitionTime":"2025-12-01T14:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.810100 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.810195 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.810214 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.810245 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.810268 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:46Z","lastTransitionTime":"2025-12-01T14:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.913649 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.913729 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.913758 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.913792 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:46 crc kubenswrapper[4810]: I1201 14:34:46.913850 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:46Z","lastTransitionTime":"2025-12-01T14:34:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.017652 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.018055 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.018261 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.018449 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.018675 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:47Z","lastTransitionTime":"2025-12-01T14:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.122590 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.122651 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.122670 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.122696 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.122715 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:47Z","lastTransitionTime":"2025-12-01T14:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.226437 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.226544 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.226571 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.226605 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.226629 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:47Z","lastTransitionTime":"2025-12-01T14:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.328905 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.328964 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.328976 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.328995 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.329008 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:47Z","lastTransitionTime":"2025-12-01T14:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.362044 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.362328 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.362391 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.362511 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.362602 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:47Z","lastTransitionTime":"2025-12-01T14:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:47 crc kubenswrapper[4810]: E1201 14:34:47.377114 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:47Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.380883 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.380918 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.380930 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.380944 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.380956 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:47Z","lastTransitionTime":"2025-12-01T14:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:47 crc kubenswrapper[4810]: E1201 14:34:47.396630 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:47Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.400600 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.400776 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.400842 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.400916 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.400983 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:47Z","lastTransitionTime":"2025-12-01T14:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:47 crc kubenswrapper[4810]: E1201 14:34:47.416109 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:47Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.420654 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.420723 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.420744 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.420770 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.420793 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:47Z","lastTransitionTime":"2025-12-01T14:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:47 crc kubenswrapper[4810]: E1201 14:34:47.434997 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:47Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.440210 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.440258 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.440272 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.440289 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.440301 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:47Z","lastTransitionTime":"2025-12-01T14:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:47 crc kubenswrapper[4810]: E1201 14:34:47.454355 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:47Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:47 crc kubenswrapper[4810]: E1201 14:34:47.454555 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.456405 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.456648 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.456777 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.456870 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.457078 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:47Z","lastTransitionTime":"2025-12-01T14:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.559613 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.560001 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.560146 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.560243 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.560349 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:47Z","lastTransitionTime":"2025-12-01T14:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.664550 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.664879 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.665007 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.665107 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.665207 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:47Z","lastTransitionTime":"2025-12-01T14:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.768649 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.769154 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.769540 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.769940 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.770256 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:47Z","lastTransitionTime":"2025-12-01T14:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.873449 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.873771 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.873887 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.873957 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.874051 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:47Z","lastTransitionTime":"2025-12-01T14:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.977130 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.977379 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.977436 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.977510 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:47 crc kubenswrapper[4810]: I1201 14:34:47.977578 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:47Z","lastTransitionTime":"2025-12-01T14:34:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.079731 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.080050 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.080182 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.080326 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.080451 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:48Z","lastTransitionTime":"2025-12-01T14:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.183213 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.183503 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.183590 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.183651 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.183712 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:48Z","lastTransitionTime":"2025-12-01T14:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.286511 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.286544 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.286554 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.286569 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.286583 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:48Z","lastTransitionTime":"2025-12-01T14:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.388580 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.388623 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.388632 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.388646 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.388657 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:48Z","lastTransitionTime":"2025-12-01T14:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.490111 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.490176 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:48 crc kubenswrapper[4810]: E1201 14:34:48.490242 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:48 crc kubenswrapper[4810]: E1201 14:34:48.490329 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.490412 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:48 crc kubenswrapper[4810]: E1201 14:34:48.490507 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.490550 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:48 crc kubenswrapper[4810]: E1201 14:34:48.490598 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.491546 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.491569 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.491577 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.491590 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.491599 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:48Z","lastTransitionTime":"2025-12-01T14:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.594904 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.594961 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.594979 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.595001 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.595017 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:48Z","lastTransitionTime":"2025-12-01T14:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.697104 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.697169 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.697185 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.697208 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.697229 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:48Z","lastTransitionTime":"2025-12-01T14:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.713394 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs\") pod \"network-metrics-daemon-jlwwg\" (UID: \"d379bcb2-a35d-470f-894c-8629d5023df8\") " pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:48 crc kubenswrapper[4810]: E1201 14:34:48.713612 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 14:34:48 crc kubenswrapper[4810]: E1201 14:34:48.713748 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs podName:d379bcb2-a35d-470f-894c-8629d5023df8 nodeName:}" failed. No retries permitted until 2025-12-01 14:35:04.713719435 +0000 UTC m=+70.477229108 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs") pod "network-metrics-daemon-jlwwg" (UID: "d379bcb2-a35d-470f-894c-8629d5023df8") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.799645 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.799686 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.799695 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.799722 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.799731 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:48Z","lastTransitionTime":"2025-12-01T14:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.902463 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.902534 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.902552 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.902575 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:48 crc kubenswrapper[4810]: I1201 14:34:48.902592 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:48Z","lastTransitionTime":"2025-12-01T14:34:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.006561 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.006630 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.006649 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.006676 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.006699 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:49Z","lastTransitionTime":"2025-12-01T14:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.110012 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.110069 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.110086 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.110114 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.110139 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:49Z","lastTransitionTime":"2025-12-01T14:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.212968 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.213002 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.213012 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.213025 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.213034 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:49Z","lastTransitionTime":"2025-12-01T14:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.315271 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.315361 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.315380 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.315405 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.315421 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:49Z","lastTransitionTime":"2025-12-01T14:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.417397 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.417453 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.417504 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.417522 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.417539 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:49Z","lastTransitionTime":"2025-12-01T14:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.491816 4810 scope.go:117] "RemoveContainer" containerID="55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.520365 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.520523 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.520535 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.520548 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.520558 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:49Z","lastTransitionTime":"2025-12-01T14:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.624502 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.624549 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.624561 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.624580 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.624593 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:49Z","lastTransitionTime":"2025-12-01T14:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.727754 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.728001 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.728016 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.728045 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.728061 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:49Z","lastTransitionTime":"2025-12-01T14:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.828811 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovnkube-controller/1.log" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.829980 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.830009 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.830019 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.830032 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.830041 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:49Z","lastTransitionTime":"2025-12-01T14:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.832138 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerStarted","Data":"a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6"} Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.832690 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.842569 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:49Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.852627 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ef3ab2-0451-4901-96d9-2d96faa37567\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8909387cd335182759cbebab25a30186b86ad060eccd57e825de4d2dcda211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ada971d2ea2d5b690fbc2086938a74afe6d886f3535687706f46b2425e3116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mlglg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:49Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.871408 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:49Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.890132 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:49Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.907649 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:49Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.918579 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:49Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.932224 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:49Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.932535 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.932558 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.932570 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.932587 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.932599 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:49Z","lastTransitionTime":"2025-12-01T14:34:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.948691 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:49Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.969157 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bd862d7-0562-4f45-a529-96b05cc7b1c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46594b36eb97f9787facc7ee07e0d3e2aecd45816488438a8af9e619ff37aeb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f7016800cb62ec155217c88e1d0968f86b6f807ab7a1e410579713f15741a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b24a67bb57bd282f48b4185451f120aaebab8373d0e368a70c97596b53a95213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:49Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.984999 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:49Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:49 crc kubenswrapper[4810]: I1201 14:34:49.995463 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:49Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.005359 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:50Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.023602 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:29Z\\\",\\\"message\\\":\\\"ort:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1201 14:34:29.634576 6247 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nI1201 14:34:29.634605 6247 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nF1201 14:34:29.634613 6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:50Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.032915 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jlwwg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d379bcb2-a35d-470f-894c-8629d5023df8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jlwwg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:50Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.034547 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.034584 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.034593 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.034608 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.034618 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:50Z","lastTransitionTime":"2025-12-01T14:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.045465 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:50Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.056800 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:50Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.070011 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:50Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.087259 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:50Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.136736 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.136798 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.136809 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.136829 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.136843 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:50Z","lastTransitionTime":"2025-12-01T14:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.239528 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.239571 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.239582 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.239597 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.239607 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:50Z","lastTransitionTime":"2025-12-01T14:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.342372 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.342422 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.342432 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.342455 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.342465 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:50Z","lastTransitionTime":"2025-12-01T14:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.445228 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.445326 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.445345 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.445368 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.445386 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:50Z","lastTransitionTime":"2025-12-01T14:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.490617 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.490684 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.490791 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:50 crc kubenswrapper[4810]: E1201 14:34:50.490961 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.491055 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:50 crc kubenswrapper[4810]: E1201 14:34:50.491190 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:50 crc kubenswrapper[4810]: E1201 14:34:50.491234 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:34:50 crc kubenswrapper[4810]: E1201 14:34:50.491338 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.549045 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.549111 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.549134 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.549194 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.549217 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:50Z","lastTransitionTime":"2025-12-01T14:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.658647 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.658717 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.658742 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.658771 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.658795 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:50Z","lastTransitionTime":"2025-12-01T14:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.761224 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.761273 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.761288 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.761309 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.761325 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:50Z","lastTransitionTime":"2025-12-01T14:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.863682 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.863740 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.863750 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.863764 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.863776 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:50Z","lastTransitionTime":"2025-12-01T14:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.966356 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.966410 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.966429 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.966455 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:50 crc kubenswrapper[4810]: I1201 14:34:50.966494 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:50Z","lastTransitionTime":"2025-12-01T14:34:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.069005 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.069059 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.069076 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.069094 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.069106 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:51Z","lastTransitionTime":"2025-12-01T14:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.172349 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.172403 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.172420 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.172444 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.172462 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:51Z","lastTransitionTime":"2025-12-01T14:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.274922 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.274956 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.274964 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.274977 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.274986 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:51Z","lastTransitionTime":"2025-12-01T14:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.377729 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.377770 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.377778 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.377793 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.377802 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:51Z","lastTransitionTime":"2025-12-01T14:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.480212 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.480278 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.480297 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.480323 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.480348 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:51Z","lastTransitionTime":"2025-12-01T14:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.582594 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.582628 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.582637 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.582652 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.582664 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:51Z","lastTransitionTime":"2025-12-01T14:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.686568 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.686627 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.686644 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.686670 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.686689 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:51Z","lastTransitionTime":"2025-12-01T14:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.789611 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.789691 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.789710 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.789739 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.789759 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:51Z","lastTransitionTime":"2025-12-01T14:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.841545 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovnkube-controller/2.log" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.842724 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovnkube-controller/1.log" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.847178 4810 generic.go:334] "Generic (PLEG): container finished" podID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerID="a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6" exitCode=1 Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.847245 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerDied","Data":"a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6"} Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.847322 4810 scope.go:117] "RemoveContainer" containerID="55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.849314 4810 scope.go:117] "RemoveContainer" containerID="a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6" Dec 01 14:34:51 crc kubenswrapper[4810]: E1201 14:34:51.849770 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.872852 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:51Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.893504 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:51Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.894036 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.894104 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.894129 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.894163 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.894192 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:51Z","lastTransitionTime":"2025-12-01T14:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.920116 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:51Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.947034 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:51Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.962558 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:51Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.978516 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:51Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.994350 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:51Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.997233 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.997302 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.997320 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.997342 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:51 crc kubenswrapper[4810]: I1201 14:34:51.997395 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:51Z","lastTransitionTime":"2025-12-01T14:34:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.014152 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:52Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.030214 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:52Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.046179 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ef3ab2-0451-4901-96d9-2d96faa37567\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8909387cd335182759cbebab25a30186b86ad060eccd57e825de4d2dcda211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ada971d2ea2d5b690fbc2086938a74afe6d886f3535687706f46b2425e3116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mlglg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:52Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.077421 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:52Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.094633 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:52Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.100239 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.100276 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.100288 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.100305 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.100316 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:52Z","lastTransitionTime":"2025-12-01T14:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.108553 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bd862d7-0562-4f45-a529-96b05cc7b1c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46594b36eb97f9787facc7ee07e0d3e2aecd45816488438a8af9e619ff37aeb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f7016800cb62ec155217c88e1d0968f86b6f807ab7a1e410579713f15741a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b24a67bb57bd282f48b4185451f120aaebab8373d0e368a70c97596b53a95213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:52Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.124786 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:52Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.137340 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:52Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.159506 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:29Z\\\",\\\"message\\\":\\\"ort:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1201 14:34:29.634576 6247 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nI1201 14:34:29.634605 6247 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nF1201 14:34:29.634613 6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:50Z\\\",\\\"message\\\":\\\"\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 14:34:50.479725 6482 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 14:34:50.479773 6482 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 14:34:50.479808 6482 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:52Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.175118 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jlwwg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d379bcb2-a35d-470f-894c-8629d5023df8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jlwwg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:52Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.194353 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:52Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.202571 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.202618 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.202629 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.202646 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.202656 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:52Z","lastTransitionTime":"2025-12-01T14:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.305366 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.305415 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.305427 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.305444 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.305457 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:52Z","lastTransitionTime":"2025-12-01T14:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.408137 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.408221 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.408252 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.408286 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.408312 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:52Z","lastTransitionTime":"2025-12-01T14:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.490381 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.490436 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.490381 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.490406 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:52 crc kubenswrapper[4810]: E1201 14:34:52.490556 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:52 crc kubenswrapper[4810]: E1201 14:34:52.490637 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:52 crc kubenswrapper[4810]: E1201 14:34:52.490671 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:52 crc kubenswrapper[4810]: E1201 14:34:52.490733 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.511205 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.511270 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.511282 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.511297 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.511306 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:52Z","lastTransitionTime":"2025-12-01T14:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.615199 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.615282 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.615306 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.615333 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.615364 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:52Z","lastTransitionTime":"2025-12-01T14:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.718679 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.718720 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.718730 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.718745 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.718757 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:52Z","lastTransitionTime":"2025-12-01T14:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.821345 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.821379 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.821390 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.821407 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.821418 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:52Z","lastTransitionTime":"2025-12-01T14:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.853727 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovnkube-controller/2.log" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.924532 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.924600 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.924624 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.924655 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:52 crc kubenswrapper[4810]: I1201 14:34:52.924678 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:52Z","lastTransitionTime":"2025-12-01T14:34:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.027542 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.027602 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.027618 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.027641 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.027658 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:53Z","lastTransitionTime":"2025-12-01T14:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.130999 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.131069 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.131096 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.131124 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.131147 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:53Z","lastTransitionTime":"2025-12-01T14:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.234078 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.234151 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.234174 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.234205 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.234224 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:53Z","lastTransitionTime":"2025-12-01T14:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.337342 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.337422 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.337448 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.337511 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.337542 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:53Z","lastTransitionTime":"2025-12-01T14:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.440432 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.440525 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.440550 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.440581 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.440605 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:53Z","lastTransitionTime":"2025-12-01T14:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.543510 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.543578 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.543605 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.543633 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.543657 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:53Z","lastTransitionTime":"2025-12-01T14:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.652652 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.652698 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.652713 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.652733 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.652748 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:53Z","lastTransitionTime":"2025-12-01T14:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.754810 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.754852 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.754861 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.754875 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.754888 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:53Z","lastTransitionTime":"2025-12-01T14:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.860087 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.860154 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.860163 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.860179 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.860187 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:53Z","lastTransitionTime":"2025-12-01T14:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.962503 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.962543 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.962555 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.962573 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:53 crc kubenswrapper[4810]: I1201 14:34:53.962585 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:53Z","lastTransitionTime":"2025-12-01T14:34:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.064515 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.064551 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.064562 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.064580 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.064591 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:54Z","lastTransitionTime":"2025-12-01T14:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.166352 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.166379 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.166387 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.166400 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.166408 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:54Z","lastTransitionTime":"2025-12-01T14:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.269662 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.269710 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.269722 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.269741 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.269757 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:54Z","lastTransitionTime":"2025-12-01T14:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.372991 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.373061 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.373088 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.373110 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.373126 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:54Z","lastTransitionTime":"2025-12-01T14:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.476222 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.476273 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.476286 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.476307 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.476327 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:54Z","lastTransitionTime":"2025-12-01T14:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.490888 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.490895 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.490940 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.491034 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:54 crc kubenswrapper[4810]: E1201 14:34:54.491150 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:54 crc kubenswrapper[4810]: E1201 14:34:54.491231 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:54 crc kubenswrapper[4810]: E1201 14:34:54.491341 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:34:54 crc kubenswrapper[4810]: E1201 14:34:54.492084 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.507179 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:54Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.519638 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bd862d7-0562-4f45-a529-96b05cc7b1c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46594b36eb97f9787facc7ee07e0d3e2aecd45816488438a8af9e619ff37aeb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f7016800cb62ec155217c88e1d0968f86b6f807ab7a1e410579713f15741a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b24a67bb57bd282f48b4185451f120aaebab8373d0e368a70c97596b53a95213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:54Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.536581 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:54Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.561293 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:54Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.579491 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.579528 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.579547 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.579566 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.579581 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:54Z","lastTransitionTime":"2025-12-01T14:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.591835 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:29Z\\\",\\\"message\\\":\\\"ort:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1201 14:34:29.634576 6247 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nI1201 14:34:29.634605 6247 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nF1201 14:34:29.634613 6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:50Z\\\",\\\"message\\\":\\\"\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 14:34:50.479725 6482 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 14:34:50.479773 6482 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 14:34:50.479808 6482 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:54Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.607440 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jlwwg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d379bcb2-a35d-470f-894c-8629d5023df8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jlwwg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:54Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.624893 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:54Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.642145 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:54Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.656184 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:54Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.674486 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:54Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.681677 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.681727 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.681739 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.681756 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.681768 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:54Z","lastTransitionTime":"2025-12-01T14:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.693362 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:54Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.709382 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:54Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.723364 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:54Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.737708 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:54Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.751769 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:54Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.765057 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:54Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.778077 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ef3ab2-0451-4901-96d9-2d96faa37567\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8909387cd335182759cbebab25a30186b86ad060eccd57e825de4d2dcda211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ada971d2ea2d5b690fbc2086938a74afe6d886f3535687706f46b2425e3116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mlglg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:54Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.783701 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.783731 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.783741 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.783757 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.783766 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:54Z","lastTransitionTime":"2025-12-01T14:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.803023 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:54Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.885835 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.885881 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.885889 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.885903 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.885913 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:54Z","lastTransitionTime":"2025-12-01T14:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.988389 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.988422 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.988430 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.988443 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:54 crc kubenswrapper[4810]: I1201 14:34:54.988452 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:54Z","lastTransitionTime":"2025-12-01T14:34:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.091532 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.091601 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.091625 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.091653 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.091675 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:55Z","lastTransitionTime":"2025-12-01T14:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.193846 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.194215 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.194311 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.194419 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.194548 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:55Z","lastTransitionTime":"2025-12-01T14:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.297261 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.297596 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.297672 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.297765 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.297854 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:55Z","lastTransitionTime":"2025-12-01T14:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.399946 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.399998 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.400019 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.400041 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.400057 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:55Z","lastTransitionTime":"2025-12-01T14:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.502453 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.503054 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.503144 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.503242 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.503397 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:55Z","lastTransitionTime":"2025-12-01T14:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.606266 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.606570 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.606665 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.606756 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.606851 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:55Z","lastTransitionTime":"2025-12-01T14:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.709012 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.709086 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.709096 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.709110 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.709120 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:55Z","lastTransitionTime":"2025-12-01T14:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.811609 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.811645 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.811656 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.811672 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.811685 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:55Z","lastTransitionTime":"2025-12-01T14:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.914994 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.915068 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.915092 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.915121 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:55 crc kubenswrapper[4810]: I1201 14:34:55.915142 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:55Z","lastTransitionTime":"2025-12-01T14:34:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.018449 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.018525 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.018541 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.018563 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.018584 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:56Z","lastTransitionTime":"2025-12-01T14:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.122637 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.123151 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.123168 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.123194 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.123215 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:56Z","lastTransitionTime":"2025-12-01T14:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.226178 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.226206 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.226218 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.226232 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.226243 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:56Z","lastTransitionTime":"2025-12-01T14:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.328833 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.328866 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.328878 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.328895 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.328909 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:56Z","lastTransitionTime":"2025-12-01T14:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.431688 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.431724 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.431734 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.431748 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.431759 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:56Z","lastTransitionTime":"2025-12-01T14:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.492535 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:56 crc kubenswrapper[4810]: E1201 14:34:56.492632 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.492777 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:56 crc kubenswrapper[4810]: E1201 14:34:56.492822 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.492909 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:56 crc kubenswrapper[4810]: E1201 14:34:56.492951 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.493064 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:56 crc kubenswrapper[4810]: E1201 14:34:56.493116 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.534362 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.534399 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.534412 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.534430 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.534443 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:56Z","lastTransitionTime":"2025-12-01T14:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.637294 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.637329 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.637341 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.637357 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.637368 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:56Z","lastTransitionTime":"2025-12-01T14:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.739758 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.739785 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.739795 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.739809 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.739818 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:56Z","lastTransitionTime":"2025-12-01T14:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.841896 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.841944 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.841962 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.841994 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.842012 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:56Z","lastTransitionTime":"2025-12-01T14:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.944174 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.944200 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.944207 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.944219 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:56 crc kubenswrapper[4810]: I1201 14:34:56.944227 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:56Z","lastTransitionTime":"2025-12-01T14:34:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.046124 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.046148 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.046156 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.046167 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.046175 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:57Z","lastTransitionTime":"2025-12-01T14:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.148573 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.148609 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.148620 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.148635 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.148648 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:57Z","lastTransitionTime":"2025-12-01T14:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.251611 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.251674 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.251693 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.251717 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.251735 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:57Z","lastTransitionTime":"2025-12-01T14:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.354619 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.354689 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.354711 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.354734 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.354751 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:57Z","lastTransitionTime":"2025-12-01T14:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.457709 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.457806 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.457827 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.457851 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.457868 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:57Z","lastTransitionTime":"2025-12-01T14:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.560426 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.560491 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.560503 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.560520 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.560534 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:57Z","lastTransitionTime":"2025-12-01T14:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.663807 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.663854 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.663865 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.663883 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.663894 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:57Z","lastTransitionTime":"2025-12-01T14:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.691619 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.691667 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.691679 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.691701 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.691718 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:57Z","lastTransitionTime":"2025-12-01T14:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:57 crc kubenswrapper[4810]: E1201 14:34:57.709759 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:57Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.715457 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.715603 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.715622 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.715653 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.715672 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:57Z","lastTransitionTime":"2025-12-01T14:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:57 crc kubenswrapper[4810]: E1201 14:34:57.738356 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:57Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.743591 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.743677 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.743693 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.743717 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.743759 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:57Z","lastTransitionTime":"2025-12-01T14:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:57 crc kubenswrapper[4810]: E1201 14:34:57.766236 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:57Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.772072 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.772112 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.772124 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.772143 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.772156 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:57Z","lastTransitionTime":"2025-12-01T14:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:57 crc kubenswrapper[4810]: E1201 14:34:57.795779 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:57Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.802120 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.802152 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.802163 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.802178 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.802189 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:57Z","lastTransitionTime":"2025-12-01T14:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:57 crc kubenswrapper[4810]: E1201 14:34:57.842525 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:57Z is after 2025-08-24T17:21:41Z" Dec 01 14:34:57 crc kubenswrapper[4810]: E1201 14:34:57.842692 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.844328 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.844370 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.844379 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.844393 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.844403 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:57Z","lastTransitionTime":"2025-12-01T14:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.950757 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.950819 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.950830 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.950848 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:57 crc kubenswrapper[4810]: I1201 14:34:57.950858 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:57Z","lastTransitionTime":"2025-12-01T14:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.054828 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.054895 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.054914 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.054935 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.054950 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:58Z","lastTransitionTime":"2025-12-01T14:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.157414 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.157451 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.157460 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.157486 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.157496 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:58Z","lastTransitionTime":"2025-12-01T14:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.259677 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.259715 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.259724 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.259739 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.259750 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:58Z","lastTransitionTime":"2025-12-01T14:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.362036 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.362074 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.362086 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.362104 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.362118 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:58Z","lastTransitionTime":"2025-12-01T14:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.464405 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.464441 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.464453 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.464485 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.464496 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:58Z","lastTransitionTime":"2025-12-01T14:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.490534 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.490750 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.490858 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.491264 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:34:58 crc kubenswrapper[4810]: E1201 14:34:58.491383 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:34:58 crc kubenswrapper[4810]: E1201 14:34:58.491563 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:34:58 crc kubenswrapper[4810]: E1201 14:34:58.491626 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:34:58 crc kubenswrapper[4810]: E1201 14:34:58.491663 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.567033 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.567067 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.567075 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.567089 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.567098 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:58Z","lastTransitionTime":"2025-12-01T14:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.668692 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.668718 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.668726 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.668738 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.668748 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:58Z","lastTransitionTime":"2025-12-01T14:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.771373 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.771419 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.771430 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.771453 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.771492 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:58Z","lastTransitionTime":"2025-12-01T14:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.873909 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.874194 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.874263 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.874334 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.874427 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:58Z","lastTransitionTime":"2025-12-01T14:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.976372 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.976651 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.976717 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.976808 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:58 crc kubenswrapper[4810]: I1201 14:34:58.976953 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:58Z","lastTransitionTime":"2025-12-01T14:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.079517 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.079780 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.079899 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.079972 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.080032 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:59Z","lastTransitionTime":"2025-12-01T14:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.181942 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.181976 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.181986 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.182002 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.182012 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:59Z","lastTransitionTime":"2025-12-01T14:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.284808 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.284855 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.284867 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.284884 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.284895 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:59Z","lastTransitionTime":"2025-12-01T14:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.388273 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.388319 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.388329 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.388345 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.388356 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:59Z","lastTransitionTime":"2025-12-01T14:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.490934 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.491183 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.491269 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.491436 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.491581 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:59Z","lastTransitionTime":"2025-12-01T14:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.594994 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.595833 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.595981 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.596125 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.596259 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:59Z","lastTransitionTime":"2025-12-01T14:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.699350 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.699389 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.699405 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.699427 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.699438 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:59Z","lastTransitionTime":"2025-12-01T14:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.802412 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.802463 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.802504 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.802520 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.802530 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:59Z","lastTransitionTime":"2025-12-01T14:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.905339 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.905388 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.905407 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.905427 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:34:59 crc kubenswrapper[4810]: I1201 14:34:59.905442 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:34:59Z","lastTransitionTime":"2025-12-01T14:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.007949 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.008014 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.008032 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.008049 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.008060 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:00Z","lastTransitionTime":"2025-12-01T14:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.110949 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.110995 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.111007 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.111025 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.111037 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:00Z","lastTransitionTime":"2025-12-01T14:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.213525 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.213579 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.213591 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.213606 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.213617 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:00Z","lastTransitionTime":"2025-12-01T14:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.316080 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.316126 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.316136 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.316152 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.316161 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:00Z","lastTransitionTime":"2025-12-01T14:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.418085 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.418129 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.418144 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.418166 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.418182 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:00Z","lastTransitionTime":"2025-12-01T14:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.490851 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.490941 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:00 crc kubenswrapper[4810]: E1201 14:35:00.491523 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.490976 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.490955 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:00 crc kubenswrapper[4810]: E1201 14:35:00.492223 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:00 crc kubenswrapper[4810]: E1201 14:35:00.491569 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:00 crc kubenswrapper[4810]: E1201 14:35:00.495868 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.520166 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.520224 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.520235 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.520251 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.520264 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:00Z","lastTransitionTime":"2025-12-01T14:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.622481 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.622528 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.622537 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.622552 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.622563 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:00Z","lastTransitionTime":"2025-12-01T14:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.725101 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.725158 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.725166 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.725181 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.725193 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:00Z","lastTransitionTime":"2025-12-01T14:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.827346 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.827616 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.827695 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.827771 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.827829 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:00Z","lastTransitionTime":"2025-12-01T14:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.931069 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.931344 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.931438 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.931533 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:00 crc kubenswrapper[4810]: I1201 14:35:00.931601 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:00Z","lastTransitionTime":"2025-12-01T14:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.033498 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.033536 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.033545 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.033561 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.033569 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:01Z","lastTransitionTime":"2025-12-01T14:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.135749 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.135780 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.135789 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.135803 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.135812 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:01Z","lastTransitionTime":"2025-12-01T14:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.238318 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.238358 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.238370 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.238384 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.238394 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:01Z","lastTransitionTime":"2025-12-01T14:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.340382 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.340421 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.340433 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.340450 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.340462 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:01Z","lastTransitionTime":"2025-12-01T14:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.442889 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.442928 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.442936 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.442950 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.442961 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:01Z","lastTransitionTime":"2025-12-01T14:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.545248 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.545298 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.545314 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.545338 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.545354 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:01Z","lastTransitionTime":"2025-12-01T14:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.647513 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.647547 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.647555 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.647567 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.647576 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:01Z","lastTransitionTime":"2025-12-01T14:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.749741 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.749780 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.749791 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.749807 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.749819 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:01Z","lastTransitionTime":"2025-12-01T14:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.851574 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.851638 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.851654 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.851680 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.851697 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:01Z","lastTransitionTime":"2025-12-01T14:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.953643 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.953684 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.953694 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.953709 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:01 crc kubenswrapper[4810]: I1201 14:35:01.953718 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:01Z","lastTransitionTime":"2025-12-01T14:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.055777 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.055819 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.055832 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.055848 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.055858 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:02Z","lastTransitionTime":"2025-12-01T14:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.158360 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.158401 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.158415 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.158431 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.158443 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:02Z","lastTransitionTime":"2025-12-01T14:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.261075 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.261119 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.261131 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.261148 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.261161 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:02Z","lastTransitionTime":"2025-12-01T14:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.363348 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.363375 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.363382 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.363394 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.363404 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:02Z","lastTransitionTime":"2025-12-01T14:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.466139 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.466388 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.466537 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.466645 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.466727 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:02Z","lastTransitionTime":"2025-12-01T14:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.490778 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.490845 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.490802 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:02 crc kubenswrapper[4810]: E1201 14:35:02.490928 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.490979 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:02 crc kubenswrapper[4810]: E1201 14:35:02.491100 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:02 crc kubenswrapper[4810]: E1201 14:35:02.491156 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:02 crc kubenswrapper[4810]: E1201 14:35:02.491224 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.569607 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.569638 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.569646 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.569660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.569669 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:02Z","lastTransitionTime":"2025-12-01T14:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.672003 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.672048 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.672064 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.672087 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.672102 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:02Z","lastTransitionTime":"2025-12-01T14:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.774709 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.774746 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.774756 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.774773 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.774783 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:02Z","lastTransitionTime":"2025-12-01T14:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.877718 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.877967 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.878052 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.878138 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.878221 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:02Z","lastTransitionTime":"2025-12-01T14:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.980616 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.980874 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.980939 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.980998 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:02 crc kubenswrapper[4810]: I1201 14:35:02.981051 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:02Z","lastTransitionTime":"2025-12-01T14:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.083285 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.083558 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.083653 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.083745 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.083806 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:03Z","lastTransitionTime":"2025-12-01T14:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.187076 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.187129 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.187143 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.187163 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.187179 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:03Z","lastTransitionTime":"2025-12-01T14:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.289283 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.289327 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.289336 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.289352 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.289363 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:03Z","lastTransitionTime":"2025-12-01T14:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.391334 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.391365 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.391373 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.391384 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.391392 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:03Z","lastTransitionTime":"2025-12-01T14:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.492981 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.493022 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.493034 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.493049 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.493061 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:03Z","lastTransitionTime":"2025-12-01T14:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.596605 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.596649 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.596660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.596677 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.596687 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:03Z","lastTransitionTime":"2025-12-01T14:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.698874 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.699355 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.699440 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.699593 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.699631 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:03Z","lastTransitionTime":"2025-12-01T14:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.802383 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.802412 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.802420 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.802433 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.802442 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:03Z","lastTransitionTime":"2025-12-01T14:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.904893 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.904928 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.904940 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.904956 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:03 crc kubenswrapper[4810]: I1201 14:35:03.904969 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:03Z","lastTransitionTime":"2025-12-01T14:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.007342 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.007373 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.007381 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.007394 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.007404 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:04Z","lastTransitionTime":"2025-12-01T14:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.110505 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.110557 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.110573 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.110600 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.110616 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:04Z","lastTransitionTime":"2025-12-01T14:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.213759 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.213986 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.214075 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.214141 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.214203 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:04Z","lastTransitionTime":"2025-12-01T14:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.316562 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.316602 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.316614 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.316629 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.316641 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:04Z","lastTransitionTime":"2025-12-01T14:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.419416 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.419498 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.419519 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.419545 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.419564 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:04Z","lastTransitionTime":"2025-12-01T14:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.490996 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.491088 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:04 crc kubenswrapper[4810]: E1201 14:35:04.491130 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.491261 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:04 crc kubenswrapper[4810]: E1201 14:35:04.491318 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.491393 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:04 crc kubenswrapper[4810]: E1201 14:35:04.491566 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:04 crc kubenswrapper[4810]: E1201 14:35:04.491705 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.507957 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.520502 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.521873 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.521914 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.521926 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.521943 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.521957 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:04Z","lastTransitionTime":"2025-12-01T14:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.531620 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.539889 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.548188 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.559317 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ef3ab2-0451-4901-96d9-2d96faa37567\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8909387cd335182759cbebab25a30186b86ad060eccd57e825de4d2dcda211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ada971d2ea2d5b690fbc2086938a74afe6d886f3535687706f46b2425e3116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mlglg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.576362 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.587891 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.599252 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.609829 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.618993 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.625254 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.625305 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.625316 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.625333 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.625349 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:04Z","lastTransitionTime":"2025-12-01T14:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.632248 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.642933 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bd862d7-0562-4f45-a529-96b05cc7b1c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46594b36eb97f9787facc7ee07e0d3e2aecd45816488438a8af9e619ff37aeb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f7016800cb62ec155217c88e1d0968f86b6f807ab7a1e410579713f15741a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b24a67bb57bd282f48b4185451f120aaebab8373d0e368a70c97596b53a95213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.661127 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.676189 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.690450 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.712181 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:29Z\\\",\\\"message\\\":\\\"ort:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1201 14:34:29.634576 6247 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nI1201 14:34:29.634605 6247 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nF1201 14:34:29.634613 6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:50Z\\\",\\\"message\\\":\\\"\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 14:34:50.479725 6482 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 14:34:50.479773 6482 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 14:34:50.479808 6482 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.723028 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jlwwg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d379bcb2-a35d-470f-894c-8629d5023df8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jlwwg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:04Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.727532 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.727567 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.727580 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.727597 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.727609 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:04Z","lastTransitionTime":"2025-12-01T14:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.787710 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs\") pod \"network-metrics-daemon-jlwwg\" (UID: \"d379bcb2-a35d-470f-894c-8629d5023df8\") " pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:04 crc kubenswrapper[4810]: E1201 14:35:04.787828 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 14:35:04 crc kubenswrapper[4810]: E1201 14:35:04.787916 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs podName:d379bcb2-a35d-470f-894c-8629d5023df8 nodeName:}" failed. No retries permitted until 2025-12-01 14:35:36.787882677 +0000 UTC m=+102.551392360 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs") pod "network-metrics-daemon-jlwwg" (UID: "d379bcb2-a35d-470f-894c-8629d5023df8") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.829925 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.829955 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.829963 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.829976 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.829984 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:04Z","lastTransitionTime":"2025-12-01T14:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.932282 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.932326 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.932339 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.932356 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:04 crc kubenswrapper[4810]: I1201 14:35:04.932368 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:04Z","lastTransitionTime":"2025-12-01T14:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.034424 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.034487 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.034500 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.034520 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.034533 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:05Z","lastTransitionTime":"2025-12-01T14:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.137059 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.137129 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.137153 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.137188 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.137210 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:05Z","lastTransitionTime":"2025-12-01T14:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.239335 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.239377 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.239388 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.239404 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.239416 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:05Z","lastTransitionTime":"2025-12-01T14:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.342001 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.342049 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.342062 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.342080 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.342094 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:05Z","lastTransitionTime":"2025-12-01T14:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.444647 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.444693 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.444702 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.444718 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.444731 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:05Z","lastTransitionTime":"2025-12-01T14:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.547520 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.547558 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.547567 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.547581 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.547590 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:05Z","lastTransitionTime":"2025-12-01T14:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.649977 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.650019 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.650029 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.650046 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.650059 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:05Z","lastTransitionTime":"2025-12-01T14:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.752829 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.752924 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.752941 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.752965 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.752985 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:05Z","lastTransitionTime":"2025-12-01T14:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.855556 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.855632 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.855650 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.855671 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.855688 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:05Z","lastTransitionTime":"2025-12-01T14:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.906982 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-s4cc7_df7ecc53-6478-49cd-bfb6-6ee80e850a19/kube-multus/0.log" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.907074 4810 generic.go:334] "Generic (PLEG): container finished" podID="df7ecc53-6478-49cd-bfb6-6ee80e850a19" containerID="70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd" exitCode=1 Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.907107 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-s4cc7" event={"ID":"df7ecc53-6478-49cd-bfb6-6ee80e850a19","Type":"ContainerDied","Data":"70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd"} Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.907491 4810 scope.go:117] "RemoveContainer" containerID="70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.921891 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bd862d7-0562-4f45-a529-96b05cc7b1c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46594b36eb97f9787facc7ee07e0d3e2aecd45816488438a8af9e619ff37aeb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f7016800cb62ec155217c88e1d0968f86b6f807ab7a1e410579713f15741a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b24a67bb57bd282f48b4185451f120aaebab8373d0e368a70c97596b53a95213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.943355 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.958149 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.958185 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.958193 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.958205 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.958216 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:05Z","lastTransitionTime":"2025-12-01T14:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.965509 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:05 crc kubenswrapper[4810]: I1201 14:35:05.984460 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:05Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.006196 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55ba0b9ed88a93ea3f48eb27e0235d54ce83994cb936748d0435560be39f1aec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:29Z\\\",\\\"message\\\":\\\"ort:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1201 14:34:29.634576 6247 services_controller.go:445] Built service openshift-dns/dns-default LB template configs for network=default: []services.lbConfig(nil)\\\\nI1201 14:34:29.634605 6247 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nF1201 14:34:29.634613 6247 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:34:29Z is after\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:50Z\\\",\\\"message\\\":\\\"\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 14:34:50.479725 6482 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 14:34:50.479773 6482 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 14:34:50.479808 6482 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.021294 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jlwwg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d379bcb2-a35d-470f-894c-8629d5023df8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jlwwg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.035800 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.049197 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.060170 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.060188 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.060196 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.060209 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.060217 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:06Z","lastTransitionTime":"2025-12-01T14:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.060496 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.070945 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.081075 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ef3ab2-0451-4901-96d9-2d96faa37567\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8909387cd335182759cbebab25a30186b86ad060eccd57e825de4d2dcda211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ada971d2ea2d5b690fbc2086938a74afe6d886f3535687706f46b2425e3116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mlglg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.098898 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.110721 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.121275 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.133439 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.142973 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.157843 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:35:05Z\\\",\\\"message\\\":\\\"2025-12-01T14:34:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e86a7ce5-d3be-4bf0-8ca0-792f1d2b97b0\\\\n2025-12-01T14:34:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e86a7ce5-d3be-4bf0-8ca0-792f1d2b97b0 to /host/opt/cni/bin/\\\\n2025-12-01T14:34:19Z [verbose] multus-daemon started\\\\n2025-12-01T14:34:19Z [verbose] Readiness Indicator file check\\\\n2025-12-01T14:35:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.161828 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.161853 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.161862 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.161876 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.161885 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:06Z","lastTransitionTime":"2025-12-01T14:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.169429 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.263877 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.263912 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.263925 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.263939 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.263947 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:06Z","lastTransitionTime":"2025-12-01T14:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.365667 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.365701 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.365711 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.365728 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.365740 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:06Z","lastTransitionTime":"2025-12-01T14:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.468917 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.468976 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.468996 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.469020 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.469036 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:06Z","lastTransitionTime":"2025-12-01T14:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.490678 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.490689 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:06 crc kubenswrapper[4810]: E1201 14:35:06.490765 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.490818 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:06 crc kubenswrapper[4810]: E1201 14:35:06.490987 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.491015 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:06 crc kubenswrapper[4810]: E1201 14:35:06.491439 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:06 crc kubenswrapper[4810]: E1201 14:35:06.491672 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.492028 4810 scope.go:117] "RemoveContainer" containerID="a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6" Dec 01 14:35:06 crc kubenswrapper[4810]: E1201 14:35:06.492324 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.507116 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bd862d7-0562-4f45-a529-96b05cc7b1c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46594b36eb97f9787facc7ee07e0d3e2aecd45816488438a8af9e619ff37aeb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f7016800cb62ec155217c88e1d0968f86b6f807ab7a1e410579713f15741a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b24a67bb57bd282f48b4185451f120aaebab8373d0e368a70c97596b53a95213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.528198 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.545246 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.559462 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.571757 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.571812 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.571830 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.571853 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.571871 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:06Z","lastTransitionTime":"2025-12-01T14:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.581271 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:50Z\\\",\\\"message\\\":\\\"\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 14:34:50.479725 6482 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 14:34:50.479773 6482 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 14:34:50.479808 6482 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.593567 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jlwwg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d379bcb2-a35d-470f-894c-8629d5023df8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jlwwg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.612617 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.649610 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.663122 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.674582 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.675378 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.675526 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.675665 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.675709 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.675880 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:06Z","lastTransitionTime":"2025-12-01T14:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.686178 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ef3ab2-0451-4901-96d9-2d96faa37567\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8909387cd335182759cbebab25a30186b86ad060eccd57e825de4d2dcda211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ada971d2ea2d5b690fbc2086938a74afe6d886f3535687706f46b2425e3116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mlglg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.706668 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.718111 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.728380 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.739439 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.752110 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.766849 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:35:05Z\\\",\\\"message\\\":\\\"2025-12-01T14:34:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e86a7ce5-d3be-4bf0-8ca0-792f1d2b97b0\\\\n2025-12-01T14:34:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e86a7ce5-d3be-4bf0-8ca0-792f1d2b97b0 to /host/opt/cni/bin/\\\\n2025-12-01T14:34:19Z [verbose] multus-daemon started\\\\n2025-12-01T14:34:19Z [verbose] Readiness Indicator file check\\\\n2025-12-01T14:35:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.777935 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.777968 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.777979 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.777996 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.778008 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:06Z","lastTransitionTime":"2025-12-01T14:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.781742 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.880710 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.880769 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.880785 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.880808 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.880825 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:06Z","lastTransitionTime":"2025-12-01T14:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.912441 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-s4cc7_df7ecc53-6478-49cd-bfb6-6ee80e850a19/kube-multus/0.log" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.912578 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-s4cc7" event={"ID":"df7ecc53-6478-49cd-bfb6-6ee80e850a19","Type":"ContainerStarted","Data":"bd96824c21af3e98dc4565197e8c1e036e7a99ce8855d92ff01b023cea04e7d5"} Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.929278 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.944787 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.962549 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.977398 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.984124 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.984216 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.984243 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.984278 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.984300 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:06Z","lastTransitionTime":"2025-12-01T14:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:06 crc kubenswrapper[4810]: I1201 14:35:06.990339 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:06Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.004934 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ef3ab2-0451-4901-96d9-2d96faa37567\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8909387cd335182759cbebab25a30186b86ad060eccd57e825de4d2dcda211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ada971d2ea2d5b690fbc2086938a74afe6d886f3535687706f46b2425e3116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mlglg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.028644 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.044393 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.057420 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.067841 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.080108 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.086824 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.087004 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.087090 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.087182 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.087267 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:07Z","lastTransitionTime":"2025-12-01T14:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.094782 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96824c21af3e98dc4565197e8c1e036e7a99ce8855d92ff01b023cea04e7d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:35:05Z\\\",\\\"message\\\":\\\"2025-12-01T14:34:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e86a7ce5-d3be-4bf0-8ca0-792f1d2b97b0\\\\n2025-12-01T14:34:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e86a7ce5-d3be-4bf0-8ca0-792f1d2b97b0 to /host/opt/cni/bin/\\\\n2025-12-01T14:34:19Z [verbose] multus-daemon started\\\\n2025-12-01T14:34:19Z [verbose] Readiness Indicator file check\\\\n2025-12-01T14:35:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.106966 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bd862d7-0562-4f45-a529-96b05cc7b1c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46594b36eb97f9787facc7ee07e0d3e2aecd45816488438a8af9e619ff37aeb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f7016800cb62ec155217c88e1d0968f86b6f807ab7a1e410579713f15741a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b24a67bb57bd282f48b4185451f120aaebab8373d0e368a70c97596b53a95213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.122596 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.138135 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.150431 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.171916 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:50Z\\\",\\\"message\\\":\\\"\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 14:34:50.479725 6482 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 14:34:50.479773 6482 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 14:34:50.479808 6482 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.182057 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jlwwg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d379bcb2-a35d-470f-894c-8629d5023df8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jlwwg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:07Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.190660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.190717 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.190737 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.190764 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.190784 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:07Z","lastTransitionTime":"2025-12-01T14:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.293543 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.293591 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.293604 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.293624 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.293636 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:07Z","lastTransitionTime":"2025-12-01T14:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.396842 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.396891 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.396904 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.396924 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.396937 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:07Z","lastTransitionTime":"2025-12-01T14:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.498366 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.498412 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.498427 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.498445 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.498461 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:07Z","lastTransitionTime":"2025-12-01T14:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.601288 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.601333 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.601348 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.601371 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.601387 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:07Z","lastTransitionTime":"2025-12-01T14:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.704755 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.704823 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.704833 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.704848 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.704859 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:07Z","lastTransitionTime":"2025-12-01T14:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.807338 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.807381 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.807389 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.807403 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.807412 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:07Z","lastTransitionTime":"2025-12-01T14:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.910515 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.910565 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.910577 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.910594 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:07 crc kubenswrapper[4810]: I1201 14:35:07.910606 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:07Z","lastTransitionTime":"2025-12-01T14:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.012880 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.012955 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.012978 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.013010 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.013033 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:08Z","lastTransitionTime":"2025-12-01T14:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.032129 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.032297 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.032399 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.032539 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.032637 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:08Z","lastTransitionTime":"2025-12-01T14:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:08 crc kubenswrapper[4810]: E1201 14:35:08.043848 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.047238 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.047261 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.047272 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.047288 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.047299 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:08Z","lastTransitionTime":"2025-12-01T14:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:08 crc kubenswrapper[4810]: E1201 14:35:08.060843 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.064951 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.065090 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.065175 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.065267 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.065345 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:08Z","lastTransitionTime":"2025-12-01T14:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:08 crc kubenswrapper[4810]: E1201 14:35:08.077934 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.082505 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.082564 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.082579 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.082600 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.082619 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:08Z","lastTransitionTime":"2025-12-01T14:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:08 crc kubenswrapper[4810]: E1201 14:35:08.098619 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.102388 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.102729 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.102805 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.102874 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.102931 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:08Z","lastTransitionTime":"2025-12-01T14:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:08 crc kubenswrapper[4810]: E1201 14:35:08.116412 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:08Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:08 crc kubenswrapper[4810]: E1201 14:35:08.116784 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.118003 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.118090 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.118149 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.118209 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.118271 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:08Z","lastTransitionTime":"2025-12-01T14:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.220526 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.220769 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.220853 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.220938 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.221052 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:08Z","lastTransitionTime":"2025-12-01T14:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.323444 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.323751 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.323853 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.324364 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.324523 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:08Z","lastTransitionTime":"2025-12-01T14:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.427259 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.427541 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.427784 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.428093 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.428307 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:08Z","lastTransitionTime":"2025-12-01T14:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.491821 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.491853 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:08 crc kubenswrapper[4810]: E1201 14:35:08.491993 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.492037 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.492096 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:08 crc kubenswrapper[4810]: E1201 14:35:08.492247 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:08 crc kubenswrapper[4810]: E1201 14:35:08.492297 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:08 crc kubenswrapper[4810]: E1201 14:35:08.492382 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.531075 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.531144 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.531161 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.531186 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.531203 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:08Z","lastTransitionTime":"2025-12-01T14:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.633959 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.634005 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.634015 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.634032 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.634041 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:08Z","lastTransitionTime":"2025-12-01T14:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.737234 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.737276 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.737289 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.737308 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.737323 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:08Z","lastTransitionTime":"2025-12-01T14:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.840317 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.840391 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.840415 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.840440 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.840460 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:08Z","lastTransitionTime":"2025-12-01T14:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.943739 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.944227 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.944407 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.944658 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:08 crc kubenswrapper[4810]: I1201 14:35:08.944802 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:08Z","lastTransitionTime":"2025-12-01T14:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.048025 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.048369 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.048648 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.048693 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.048712 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:09Z","lastTransitionTime":"2025-12-01T14:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.151062 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.151131 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.151151 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.151178 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.151200 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:09Z","lastTransitionTime":"2025-12-01T14:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.254377 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.254410 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.254422 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.254439 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.254451 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:09Z","lastTransitionTime":"2025-12-01T14:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.356679 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.356718 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.356730 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.356747 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.356759 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:09Z","lastTransitionTime":"2025-12-01T14:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.460158 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.460217 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.460233 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.460255 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.460274 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:09Z","lastTransitionTime":"2025-12-01T14:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.563061 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.563202 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.563228 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.563253 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.563271 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:09Z","lastTransitionTime":"2025-12-01T14:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.666649 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.666719 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.666737 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.666763 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.666782 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:09Z","lastTransitionTime":"2025-12-01T14:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.769391 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.769430 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.769446 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.769501 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.769519 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:09Z","lastTransitionTime":"2025-12-01T14:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.872987 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.873046 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.873058 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.873078 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.873091 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:09Z","lastTransitionTime":"2025-12-01T14:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.975901 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.975969 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.975983 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.976002 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:09 crc kubenswrapper[4810]: I1201 14:35:09.976017 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:09Z","lastTransitionTime":"2025-12-01T14:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.078865 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.078917 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.078932 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.078951 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.078967 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:10Z","lastTransitionTime":"2025-12-01T14:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.181721 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.181781 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.181798 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.181822 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.181840 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:10Z","lastTransitionTime":"2025-12-01T14:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.284092 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.284130 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.284139 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.284156 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.284176 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:10Z","lastTransitionTime":"2025-12-01T14:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.386155 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.386183 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.386192 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.386207 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.386215 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:10Z","lastTransitionTime":"2025-12-01T14:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.488875 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.488904 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.488913 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.488926 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.488935 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:10Z","lastTransitionTime":"2025-12-01T14:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.490295 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:10 crc kubenswrapper[4810]: E1201 14:35:10.490505 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.490592 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:10 crc kubenswrapper[4810]: E1201 14:35:10.490710 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.490783 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.490833 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:10 crc kubenswrapper[4810]: E1201 14:35:10.490976 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:10 crc kubenswrapper[4810]: E1201 14:35:10.491078 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.592254 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.592314 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.592323 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.592335 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.592344 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:10Z","lastTransitionTime":"2025-12-01T14:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.695204 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.695242 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.695250 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.695264 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.695274 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:10Z","lastTransitionTime":"2025-12-01T14:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.797397 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.797444 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.797455 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.797487 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.797498 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:10Z","lastTransitionTime":"2025-12-01T14:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.899442 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.899508 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.899521 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.899539 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:10 crc kubenswrapper[4810]: I1201 14:35:10.899551 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:10Z","lastTransitionTime":"2025-12-01T14:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.001677 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.001714 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.001724 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.001740 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.001751 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:11Z","lastTransitionTime":"2025-12-01T14:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.104265 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.104297 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.104306 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.104318 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.104328 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:11Z","lastTransitionTime":"2025-12-01T14:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.206351 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.206402 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.206416 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.206433 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.206444 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:11Z","lastTransitionTime":"2025-12-01T14:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.309125 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.309175 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.309185 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.309199 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.309208 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:11Z","lastTransitionTime":"2025-12-01T14:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.412020 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.412058 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.412066 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.412079 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.412090 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:11Z","lastTransitionTime":"2025-12-01T14:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.514557 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.514593 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.514604 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.514619 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.514629 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:11Z","lastTransitionTime":"2025-12-01T14:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.618248 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.618328 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.618344 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.618362 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.618374 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:11Z","lastTransitionTime":"2025-12-01T14:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.721326 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.721381 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.721395 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.721416 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.721430 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:11Z","lastTransitionTime":"2025-12-01T14:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.824428 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.824600 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.824653 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.824672 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.824685 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:11Z","lastTransitionTime":"2025-12-01T14:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.928093 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.928152 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.928169 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.928194 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:11 crc kubenswrapper[4810]: I1201 14:35:11.928212 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:11Z","lastTransitionTime":"2025-12-01T14:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.030833 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.030907 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.030926 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.030953 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.030971 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:12Z","lastTransitionTime":"2025-12-01T14:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.134017 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.134089 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.134103 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.134121 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.134135 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:12Z","lastTransitionTime":"2025-12-01T14:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.236209 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.236277 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.236289 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.236306 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.236318 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:12Z","lastTransitionTime":"2025-12-01T14:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.338309 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.338351 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.338359 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.338374 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.338382 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:12Z","lastTransitionTime":"2025-12-01T14:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.440655 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.440692 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.440701 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.440715 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.440724 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:12Z","lastTransitionTime":"2025-12-01T14:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.490060 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.490131 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.490089 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.490256 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:12 crc kubenswrapper[4810]: E1201 14:35:12.490357 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:12 crc kubenswrapper[4810]: E1201 14:35:12.490582 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:12 crc kubenswrapper[4810]: E1201 14:35:12.490679 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:12 crc kubenswrapper[4810]: E1201 14:35:12.490742 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.543750 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.543793 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.543804 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.543826 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.543877 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:12Z","lastTransitionTime":"2025-12-01T14:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.647242 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.647314 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.647339 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.647372 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.647399 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:12Z","lastTransitionTime":"2025-12-01T14:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.750828 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.750898 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.750910 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.750952 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.750965 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:12Z","lastTransitionTime":"2025-12-01T14:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.853520 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.853561 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.853572 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.853584 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.853593 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:12Z","lastTransitionTime":"2025-12-01T14:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.956807 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.956851 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.956863 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.956880 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:12 crc kubenswrapper[4810]: I1201 14:35:12.956890 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:12Z","lastTransitionTime":"2025-12-01T14:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.060297 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.060342 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.060354 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.060371 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.060383 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:13Z","lastTransitionTime":"2025-12-01T14:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.162949 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.163002 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.163019 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.163045 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.163062 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:13Z","lastTransitionTime":"2025-12-01T14:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.265521 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.265587 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.265606 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.265631 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.265654 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:13Z","lastTransitionTime":"2025-12-01T14:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.368694 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.368729 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.368737 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.368750 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.368758 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:13Z","lastTransitionTime":"2025-12-01T14:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.471211 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.471292 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.471311 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.471342 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.471363 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:13Z","lastTransitionTime":"2025-12-01T14:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.575224 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.575282 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.575297 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.575316 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.575328 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:13Z","lastTransitionTime":"2025-12-01T14:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.678573 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.678635 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.678653 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.678676 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.678695 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:13Z","lastTransitionTime":"2025-12-01T14:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.781105 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.781146 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.781155 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.781167 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.781176 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:13Z","lastTransitionTime":"2025-12-01T14:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.883740 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.883815 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.883832 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.883864 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.883882 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:13Z","lastTransitionTime":"2025-12-01T14:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.986618 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.986659 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.986670 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.986685 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:13 crc kubenswrapper[4810]: I1201 14:35:13.986695 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:13Z","lastTransitionTime":"2025-12-01T14:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.089202 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.089266 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.089284 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.089304 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.089319 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:14Z","lastTransitionTime":"2025-12-01T14:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.191651 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.191687 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.191698 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.191715 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.191727 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:14Z","lastTransitionTime":"2025-12-01T14:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.296826 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.296877 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.296892 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.296912 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.296926 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:14Z","lastTransitionTime":"2025-12-01T14:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.398799 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.398861 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.398884 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.398914 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.398953 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:14Z","lastTransitionTime":"2025-12-01T14:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.491326 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.491534 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.491630 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:14 crc kubenswrapper[4810]: E1201 14:35:14.491617 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:14 crc kubenswrapper[4810]: E1201 14:35:14.491920 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.491969 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:14 crc kubenswrapper[4810]: E1201 14:35:14.492032 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:14 crc kubenswrapper[4810]: E1201 14:35:14.492082 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.501045 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.501079 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.501091 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.501109 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.501122 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:14Z","lastTransitionTime":"2025-12-01T14:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.510601 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.524052 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.536741 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.553829 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.570127 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.582777 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.592871 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.602846 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.603892 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.603939 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.603954 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.603974 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.603987 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:14Z","lastTransitionTime":"2025-12-01T14:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.615899 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96824c21af3e98dc4565197e8c1e036e7a99ce8855d92ff01b023cea04e7d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:35:05Z\\\",\\\"message\\\":\\\"2025-12-01T14:34:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e86a7ce5-d3be-4bf0-8ca0-792f1d2b97b0\\\\n2025-12-01T14:34:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e86a7ce5-d3be-4bf0-8ca0-792f1d2b97b0 to /host/opt/cni/bin/\\\\n2025-12-01T14:34:19Z [verbose] multus-daemon started\\\\n2025-12-01T14:34:19Z [verbose] Readiness Indicator file check\\\\n2025-12-01T14:35:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.626615 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.637645 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ef3ab2-0451-4901-96d9-2d96faa37567\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8909387cd335182759cbebab25a30186b86ad060eccd57e825de4d2dcda211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ada971d2ea2d5b690fbc2086938a74afe6d886f3535687706f46b2425e3116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mlglg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.658405 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.672810 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.682297 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bd862d7-0562-4f45-a529-96b05cc7b1c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46594b36eb97f9787facc7ee07e0d3e2aecd45816488438a8af9e619ff37aeb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f7016800cb62ec155217c88e1d0968f86b6f807ab7a1e410579713f15741a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b24a67bb57bd282f48b4185451f120aaebab8373d0e368a70c97596b53a95213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.693819 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.706993 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.707028 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.707037 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.707051 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.707060 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:14Z","lastTransitionTime":"2025-12-01T14:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.707028 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.726392 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:50Z\\\",\\\"message\\\":\\\"\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 14:34:50.479725 6482 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 14:34:50.479773 6482 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 14:34:50.479808 6482 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.738990 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jlwwg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d379bcb2-a35d-470f-894c-8629d5023df8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jlwwg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:14Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.809375 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.809414 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.809426 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.809443 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.809457 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:14Z","lastTransitionTime":"2025-12-01T14:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.912104 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.912145 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.912156 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.912171 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:14 crc kubenswrapper[4810]: I1201 14:35:14.912292 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:14Z","lastTransitionTime":"2025-12-01T14:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.015005 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.015044 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.015055 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.015069 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.015080 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:15Z","lastTransitionTime":"2025-12-01T14:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.117701 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.117744 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.117755 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.117772 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.117784 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:15Z","lastTransitionTime":"2025-12-01T14:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.220267 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.220384 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.220399 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.220416 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.220429 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:15Z","lastTransitionTime":"2025-12-01T14:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.323513 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.323568 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.323587 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.323609 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.323625 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:15Z","lastTransitionTime":"2025-12-01T14:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.425993 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.426040 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.426055 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.426077 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.426092 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:15Z","lastTransitionTime":"2025-12-01T14:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.529558 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.529618 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.529636 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.529660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.529677 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:15Z","lastTransitionTime":"2025-12-01T14:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.632958 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.633040 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.633065 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.633095 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.633135 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:15Z","lastTransitionTime":"2025-12-01T14:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.736793 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.736864 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.736886 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.736916 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.736938 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:15Z","lastTransitionTime":"2025-12-01T14:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.840005 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.840300 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.840434 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.840581 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.840697 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:15Z","lastTransitionTime":"2025-12-01T14:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.943725 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.943809 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.943829 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.943862 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:15 crc kubenswrapper[4810]: I1201 14:35:15.943887 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:15Z","lastTransitionTime":"2025-12-01T14:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.046681 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.046730 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.046748 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.046773 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.046790 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:16Z","lastTransitionTime":"2025-12-01T14:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.150119 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.150198 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.150217 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.150243 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.150261 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:16Z","lastTransitionTime":"2025-12-01T14:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.253320 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.253435 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.253454 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.253505 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.253523 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:16Z","lastTransitionTime":"2025-12-01T14:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.356693 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.356757 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.356774 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.356802 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.356820 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:16Z","lastTransitionTime":"2025-12-01T14:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.460451 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.460592 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.460626 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.460657 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.460681 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:16Z","lastTransitionTime":"2025-12-01T14:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.490306 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:16 crc kubenswrapper[4810]: E1201 14:35:16.490566 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.490587 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.490693 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:16 crc kubenswrapper[4810]: E1201 14:35:16.490769 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.490603 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:16 crc kubenswrapper[4810]: E1201 14:35:16.490876 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:16 crc kubenswrapper[4810]: E1201 14:35:16.490982 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.563936 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.564003 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.564028 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.564061 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.564085 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:16Z","lastTransitionTime":"2025-12-01T14:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.667332 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.667415 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.667440 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.667508 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.667531 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:16Z","lastTransitionTime":"2025-12-01T14:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.769795 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.769891 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.769918 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.769953 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.769977 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:16Z","lastTransitionTime":"2025-12-01T14:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.873622 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.873675 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.873685 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.873713 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.873726 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:16Z","lastTransitionTime":"2025-12-01T14:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.976543 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.976604 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.976622 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.976649 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:16 crc kubenswrapper[4810]: I1201 14:35:16.976668 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:16Z","lastTransitionTime":"2025-12-01T14:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.079984 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.080017 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.080027 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.080043 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.080055 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:17Z","lastTransitionTime":"2025-12-01T14:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.109990 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.110149 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:17 crc kubenswrapper[4810]: E1201 14:35:17.110253 4810 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 14:35:17 crc kubenswrapper[4810]: E1201 14:35:17.110331 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 14:36:21.110314368 +0000 UTC m=+146.873823981 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 14:35:17 crc kubenswrapper[4810]: E1201 14:35:17.110449 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:21.11038749 +0000 UTC m=+146.873897133 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.183708 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.183773 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.183791 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.183820 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.183839 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:17Z","lastTransitionTime":"2025-12-01T14:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.211720 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.211804 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.211851 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:17 crc kubenswrapper[4810]: E1201 14:35:17.211979 4810 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 14:35:17 crc kubenswrapper[4810]: E1201 14:35:17.212047 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 14:35:17 crc kubenswrapper[4810]: E1201 14:35:17.212082 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 14:35:17 crc kubenswrapper[4810]: E1201 14:35:17.212104 4810 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:35:17 crc kubenswrapper[4810]: E1201 14:35:17.212114 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 14:35:17 crc kubenswrapper[4810]: E1201 14:35:17.212169 4810 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 14:35:17 crc kubenswrapper[4810]: E1201 14:35:17.212189 4810 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:35:17 crc kubenswrapper[4810]: E1201 14:35:17.212138 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 14:36:21.212103471 +0000 UTC m=+146.975613104 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 14:35:17 crc kubenswrapper[4810]: E1201 14:35:17.212292 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 14:36:21.212260075 +0000 UTC m=+146.975769858 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:35:17 crc kubenswrapper[4810]: E1201 14:35:17.212319 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 14:36:21.212304946 +0000 UTC m=+146.975814829 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.287097 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.287232 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.287261 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.287293 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.287312 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:17Z","lastTransitionTime":"2025-12-01T14:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.391778 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.391847 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.391865 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.391890 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.391912 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:17Z","lastTransitionTime":"2025-12-01T14:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.491740 4810 scope.go:117] "RemoveContainer" containerID="a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.494326 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.494380 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.494397 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.494425 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.494442 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:17Z","lastTransitionTime":"2025-12-01T14:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.597434 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.597618 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.597649 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.597679 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.597702 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:17Z","lastTransitionTime":"2025-12-01T14:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.701905 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.701946 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.701957 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.701973 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.701984 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:17Z","lastTransitionTime":"2025-12-01T14:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.806112 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.806177 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.806194 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.806220 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.806241 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:17Z","lastTransitionTime":"2025-12-01T14:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.908965 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.909016 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.909027 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.909045 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:17 crc kubenswrapper[4810]: I1201 14:35:17.909057 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:17Z","lastTransitionTime":"2025-12-01T14:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.012456 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.013049 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.013090 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.013114 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.013140 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:18Z","lastTransitionTime":"2025-12-01T14:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.116318 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.116363 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.116378 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.116400 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.116413 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:18Z","lastTransitionTime":"2025-12-01T14:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.219997 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.220066 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.220086 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.220113 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.220134 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:18Z","lastTransitionTime":"2025-12-01T14:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.323350 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.323436 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.323501 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.323541 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.323584 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:18Z","lastTransitionTime":"2025-12-01T14:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.427577 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.427639 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.427658 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.427678 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.427692 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:18Z","lastTransitionTime":"2025-12-01T14:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.487826 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.487872 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.487882 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.487897 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.487913 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:18Z","lastTransitionTime":"2025-12-01T14:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.490394 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.490601 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:18 crc kubenswrapper[4810]: E1201 14:35:18.490780 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.490824 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.490847 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:18 crc kubenswrapper[4810]: E1201 14:35:18.490998 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:18 crc kubenswrapper[4810]: E1201 14:35:18.491238 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:18 crc kubenswrapper[4810]: E1201 14:35:18.491394 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:18 crc kubenswrapper[4810]: E1201 14:35:18.514129 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.522635 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.522700 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.522715 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.522737 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.522754 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:18Z","lastTransitionTime":"2025-12-01T14:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:18 crc kubenswrapper[4810]: E1201 14:35:18.534943 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.538805 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.538854 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.538865 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.538882 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.538892 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:18Z","lastTransitionTime":"2025-12-01T14:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:18 crc kubenswrapper[4810]: E1201 14:35:18.552570 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.555878 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.556070 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.556167 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.556248 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.556307 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:18Z","lastTransitionTime":"2025-12-01T14:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:18 crc kubenswrapper[4810]: E1201 14:35:18.568621 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.572407 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.572456 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.572498 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.572523 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.572562 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:18Z","lastTransitionTime":"2025-12-01T14:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:18 crc kubenswrapper[4810]: E1201 14:35:18.584684 4810 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4654003b-86dd-4016-8fc1-f1d206961e79\\\",\\\"systemUUID\\\":\\\"4b632de5-2300-4f1e-a112-9403c29ed772\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:18 crc kubenswrapper[4810]: E1201 14:35:18.584845 4810 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.586913 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.586949 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.586959 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.586977 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.586988 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:18Z","lastTransitionTime":"2025-12-01T14:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.689771 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.689855 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.689873 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.689898 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.689917 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:18Z","lastTransitionTime":"2025-12-01T14:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.792791 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.793007 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.793068 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.793150 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.793211 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:18Z","lastTransitionTime":"2025-12-01T14:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.896589 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.896638 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.896648 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.896664 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.896678 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:18Z","lastTransitionTime":"2025-12-01T14:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.953076 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovnkube-controller/2.log" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.957153 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerStarted","Data":"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe"} Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.957926 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.975461 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bd862d7-0562-4f45-a529-96b05cc7b1c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46594b36eb97f9787facc7ee07e0d3e2aecd45816488438a8af9e619ff37aeb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f7016800cb62ec155217c88e1d0968f86b6f807ab7a1e410579713f15741a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b24a67bb57bd282f48b4185451f120aaebab8373d0e368a70c97596b53a95213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.996402 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:18Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.999557 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.999595 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.999607 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.999625 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:18 crc kubenswrapper[4810]: I1201 14:35:18.999638 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:18Z","lastTransitionTime":"2025-12-01T14:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.014432 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.032214 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.064061 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:50Z\\\",\\\"message\\\":\\\"\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 14:34:50.479725 6482 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 14:34:50.479773 6482 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 14:34:50.479808 6482 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:35:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.080741 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jlwwg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d379bcb2-a35d-470f-894c-8629d5023df8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jlwwg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.097965 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.105046 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.105187 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.105270 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.105348 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.105458 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:19Z","lastTransitionTime":"2025-12-01T14:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.117079 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.136366 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.150902 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.176501 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.202092 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.208274 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.208551 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.208643 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.208755 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.208847 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:19Z","lastTransitionTime":"2025-12-01T14:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.217501 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.236095 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.250850 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.268095 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96824c21af3e98dc4565197e8c1e036e7a99ce8855d92ff01b023cea04e7d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:35:05Z\\\",\\\"message\\\":\\\"2025-12-01T14:34:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e86a7ce5-d3be-4bf0-8ca0-792f1d2b97b0\\\\n2025-12-01T14:34:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e86a7ce5-d3be-4bf0-8ca0-792f1d2b97b0 to /host/opt/cni/bin/\\\\n2025-12-01T14:34:19Z [verbose] multus-daemon started\\\\n2025-12-01T14:34:19Z [verbose] Readiness Indicator file check\\\\n2025-12-01T14:35:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.282498 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.297585 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ef3ab2-0451-4901-96d9-2d96faa37567\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8909387cd335182759cbebab25a30186b86ad060eccd57e825de4d2dcda211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ada971d2ea2d5b690fbc2086938a74afe6d886f3535687706f46b2425e3116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mlglg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.311922 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.311973 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.311983 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.312002 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.312015 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:19Z","lastTransitionTime":"2025-12-01T14:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.419735 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.419828 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.419847 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.419878 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.419901 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:19Z","lastTransitionTime":"2025-12-01T14:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.521718 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.521766 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.521780 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.521798 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.521811 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:19Z","lastTransitionTime":"2025-12-01T14:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.627382 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.628004 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.628025 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.628056 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.628076 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:19Z","lastTransitionTime":"2025-12-01T14:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.731296 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.731368 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.731394 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.731422 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.731442 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:19Z","lastTransitionTime":"2025-12-01T14:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.834488 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.834532 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.834543 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.834559 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.834571 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:19Z","lastTransitionTime":"2025-12-01T14:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.937343 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.937386 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.937394 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.937407 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.937418 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:19Z","lastTransitionTime":"2025-12-01T14:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.961753 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovnkube-controller/3.log" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.962285 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovnkube-controller/2.log" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.964651 4810 generic.go:334] "Generic (PLEG): container finished" podID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerID="d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe" exitCode=1 Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.964687 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerDied","Data":"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe"} Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.964721 4810 scope.go:117] "RemoveContainer" containerID="a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.965332 4810 scope.go:117] "RemoveContainer" containerID="d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe" Dec 01 14:35:19 crc kubenswrapper[4810]: E1201 14:35:19.965486 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.982917 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:19 crc kubenswrapper[4810]: I1201 14:35:19.996340 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:19Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.010646 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96824c21af3e98dc4565197e8c1e036e7a99ce8855d92ff01b023cea04e7d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:35:05Z\\\",\\\"message\\\":\\\"2025-12-01T14:34:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e86a7ce5-d3be-4bf0-8ca0-792f1d2b97b0\\\\n2025-12-01T14:34:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e86a7ce5-d3be-4bf0-8ca0-792f1d2b97b0 to /host/opt/cni/bin/\\\\n2025-12-01T14:34:19Z [verbose] multus-daemon started\\\\n2025-12-01T14:34:19Z [verbose] Readiness Indicator file check\\\\n2025-12-01T14:35:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.020256 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.031438 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ef3ab2-0451-4901-96d9-2d96faa37567\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8909387cd335182759cbebab25a30186b86ad060eccd57e825de4d2dcda211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ada971d2ea2d5b690fbc2086938a74afe6d886f3535687706f46b2425e3116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mlglg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.040620 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.040653 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.040663 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.040677 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.040688 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:20Z","lastTransitionTime":"2025-12-01T14:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.059889 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.073990 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.090158 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.104491 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bd862d7-0562-4f45-a529-96b05cc7b1c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46594b36eb97f9787facc7ee07e0d3e2aecd45816488438a8af9e619ff37aeb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f7016800cb62ec155217c88e1d0968f86b6f807ab7a1e410579713f15741a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b24a67bb57bd282f48b4185451f120aaebab8373d0e368a70c97596b53a95213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.119401 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.142012 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8ef6ddb80f40b0910d4f59db23f8bf4927769e7de10c875d7b8b623152708e6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:34:50Z\\\",\\\"message\\\":\\\"\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 14:34:50.479725 6482 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1201 14:34:50.479773 6482 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1201 14:34:50.479808 6482 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:35:19Z\\\",\\\"message\\\":\\\"ft-network-operator/iptables-alerter-4ln5h\\\\nI1201 14:35:18.916152 6853 services_controller.go:452] Built service openshift-image-registry/image-registry per-node LB for network=default: []services.LB{}\\\\nI1201 14:35:18.916155 6853 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1201 14:35:18.916160 6853 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-rbt4q\\\\nI1201 14:35:18.916168 6853 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-kbfbn\\\\nF1201 14:35:18.916053 6853 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:35:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.143724 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.143783 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.143800 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.143827 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.143843 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:20Z","lastTransitionTime":"2025-12-01T14:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.154369 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jlwwg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d379bcb2-a35d-470f-894c-8629d5023df8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jlwwg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.172854 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.193348 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.207397 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.225189 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.245336 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.246580 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.246622 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.246636 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.246653 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.246666 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:20Z","lastTransitionTime":"2025-12-01T14:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.262913 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.350049 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.350093 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.350109 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.350127 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.350137 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:20Z","lastTransitionTime":"2025-12-01T14:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.452097 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.452158 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.452175 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.452199 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.452217 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:20Z","lastTransitionTime":"2025-12-01T14:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.490499 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.490557 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.490562 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.490690 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:20 crc kubenswrapper[4810]: E1201 14:35:20.490703 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:20 crc kubenswrapper[4810]: E1201 14:35:20.490809 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:20 crc kubenswrapper[4810]: E1201 14:35:20.490895 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:20 crc kubenswrapper[4810]: E1201 14:35:20.490994 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.554211 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.554271 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.554282 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.554297 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.554339 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:20Z","lastTransitionTime":"2025-12-01T14:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.657438 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.657526 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.657538 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.657553 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.657594 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:20Z","lastTransitionTime":"2025-12-01T14:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.760277 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.760346 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.760363 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.760386 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.760403 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:20Z","lastTransitionTime":"2025-12-01T14:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.862820 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.862864 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.862876 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.862922 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.862936 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:20Z","lastTransitionTime":"2025-12-01T14:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.965643 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.965687 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.965699 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.965723 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.965737 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:20Z","lastTransitionTime":"2025-12-01T14:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.968948 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovnkube-controller/3.log" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.972507 4810 scope.go:117] "RemoveContainer" containerID="d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe" Dec 01 14:35:20 crc kubenswrapper[4810]: E1201 14:35:20.972702 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" Dec 01 14:35:20 crc kubenswrapper[4810]: I1201 14:35:20.993722 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41bbeb3b-624e-4b84-896f-ecd37a5cb7f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"message\\\":\\\"g.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197867 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 14:34:13.197872 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 14:34:13.197875 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 14:34:13.197878 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 14:34:13.197881 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 14:34:13.198022 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1201 14:34:13.202259 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202314 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1201 14:34:13.202360 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202377 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1201 14:34:13.202399 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1201 14:34:13.202413 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1201 14:34:13.202555 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1201 14:34:13.202574 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nF1201 14:34:13.202560 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:20Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.007866 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75c68a3a-18c6-40db-a980-0141d5edb1a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://011fa7016c38b4a235219bc2e3d0e3f328b8f5caa35c4d22c433e3dfbdfc010c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cd134c2b759a6852fd1260531aa22c395541588dd88294ebdd98510a194460\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e3128b93aa1107357ccf49c1ffff0c122d99fbc1f2fdbf352864afde622c769\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.020861 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.035761 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c275317-c741-4d61-a399-d196f37f1745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dc6728f115a60624f45ef04be18cda87edfab14d2ffd9fa3d9bd0959de6544f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mwkk7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rbt4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.051266 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"42ef3ab2-0451-4901-96d9-2d96faa37567\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da8909387cd335182759cbebab25a30186b86ad060eccd57e825de4d2dcda211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6ada971d2ea2d5b690fbc2086938a74afe6d886f3535687706f46b2425e3116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpfv5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:31Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-mlglg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.068798 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.068844 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.068856 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.068876 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.068890 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:21Z","lastTransitionTime":"2025-12-01T14:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.081898 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b73ce030-8a45-423c-a76e-f7d42d13a8a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d836e9519bfa4f8fb92997b22304bb7beefc7f2345a80cf369d708ed27fed401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01bf8d30b7b5e164a8b9d3d095ce0d4cb172abe71387f0f87067083303572ffd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb7eb8ebb296d404187f867b94931f315364d18addcd1369301d4248de7fbd51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4fc738cd291cd4dec687d14846ebcbba7ffa26dc2fb41306648903bb190f413\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2212c1c72c184723a932d44321cd52943e680550e36a06426bb62e953a01684\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae79ef61ffd852b8b1b66ef26f0c8b6bd46f8bfc7a2ff17076b1e5320df84179\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4370308690637726dcb4fac875493049a63608aec8fce767c3e3534b8413cec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1849e88064d463478f72f4314ac9f7ce50e432ba1a207dbea20af65cd5fdac5c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.099857 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2d144e1d49ec0150a2bea99203b90da1889c6595ec6a7a9aa0587870ea54d995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.116653 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.133680 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:16Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://947cd32b793e095cf30cacca196be40c8c43ef184e00fff7a5813e822a882aaf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.148857 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-v7dlj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293d2ee5-5bc6-40ca-a176-6ea87131aa9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1265c7da88427eb21b3810af05ac9f2230499ed3be4cfde07b3f537adb2cd678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4xws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-v7dlj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.170742 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-s4cc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df7ecc53-6478-49cd-bfb6-6ee80e850a19\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd96824c21af3e98dc4565197e8c1e036e7a99ce8855d92ff01b023cea04e7d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:35:05Z\\\",\\\"message\\\":\\\"2025-12-01T14:34:19+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e86a7ce5-d3be-4bf0-8ca0-792f1d2b97b0\\\\n2025-12-01T14:34:19+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e86a7ce5-d3be-4bf0-8ca0-792f1d2b97b0 to /host/opt/cni/bin/\\\\n2025-12-01T14:34:19Z [verbose] multus-daemon started\\\\n2025-12-01T14:34:19Z [verbose] Readiness Indicator file check\\\\n2025-12-01T14:35:04Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jnqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-s4cc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.172300 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.172378 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.172404 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.172435 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.172459 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:21Z","lastTransitionTime":"2025-12-01T14:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.188353 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h8m66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87c01a68-ffa6-46eb-97b2-65fc457d8a79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a37327c609f13c5b725d831516c8ed5cc181908fab57fee25448144cb5b97dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcs69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h8m66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.202658 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0bd862d7-0562-4f45-a529-96b05cc7b1c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:33:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46594b36eb97f9787facc7ee07e0d3e2aecd45816488438a8af9e619ff37aeb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f7016800cb62ec155217c88e1d0968f86b6f807ab7a1e410579713f15741a4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b24a67bb57bd282f48b4185451f120aaebab8373d0e368a70c97596b53a95213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:33:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1a3491cc3c359ca48e8a059ea3b89b0421a4476ddc47206a6fa4ef25acb31acf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:33:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:33:55Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:33:54Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.233705 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"48f9858e-b927-4c2f-ab39-ded37b102dcd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101d52374be3398c3c0f335e98e355397174a1a15362dc164a89f5d71824d71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9dc0f258c1ebcb123dbb3993a0cfed00557efa119357bb1b46db11374f8a71a8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda06a79d6d488fbe7e820512d8fe93f098674f3e3f53b922d4fb70f11236c26\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3210cc60ed8177b0d00f8b1442868f590b7f5c72c812838dcf40f639ebb5748\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df7f6d77466fe3ba182b4347524494e45e2e35519d9be6210c1b625c35f1fd75\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f9fd97d6a2c8d91f1fcc0b76d45c6b31320c45d5a3c0deebb7d2c1b388262ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bd4b250b435ed663d7efa52b205a811acbc7b9090d10a3ba7f8510f82be54de\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4tg8g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:18Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kbfbn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.274791 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.274838 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.274851 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.274887 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.274900 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:21Z","lastTransitionTime":"2025-12-01T14:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.285694 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.306648 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:13Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b002cdbb47688668d222d41a55a3000c67337e687adeb797dae313384cc7b4ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2576b711b4aa149c5678438ed89540acccdeecbf384b5a403f4a68a482013b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.330908 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efba5556-97da-447f-9dda-40c2ea6d3e3d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:19Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T14:35:19Z\\\",\\\"message\\\":\\\"ft-network-operator/iptables-alerter-4ln5h\\\\nI1201 14:35:18.916152 6853 services_controller.go:452] Built service openshift-image-registry/image-registry per-node LB for network=default: []services.LB{}\\\\nI1201 14:35:18.916155 6853 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1201 14:35:18.916160 6853 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-rbt4q\\\\nI1201 14:35:18.916168 6853 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-kbfbn\\\\nF1201 14:35:18.916053 6853 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T14:35:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T14:34:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T14:34:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T14:34:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7xmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-p7pvm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.341747 4810 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jlwwg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d379bcb2-a35d-470f-894c-8629d5023df8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T14:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-djdpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T14:34:32Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jlwwg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T14:35:21Z is after 2025-08-24T17:21:41Z" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.377134 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.377171 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.377180 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.377192 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.377202 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:21Z","lastTransitionTime":"2025-12-01T14:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.480354 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.480417 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.480435 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.480457 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.480503 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:21Z","lastTransitionTime":"2025-12-01T14:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.582576 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.582657 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.582675 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.582699 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.582717 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:21Z","lastTransitionTime":"2025-12-01T14:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.685335 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.685372 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.685381 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.685395 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.685404 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:21Z","lastTransitionTime":"2025-12-01T14:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.788603 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.788660 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.788677 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.788703 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.788722 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:21Z","lastTransitionTime":"2025-12-01T14:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.892104 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.892201 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.892217 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.892241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.892257 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:21Z","lastTransitionTime":"2025-12-01T14:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.995174 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.995273 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.995294 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.995315 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:21 crc kubenswrapper[4810]: I1201 14:35:21.995332 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:21Z","lastTransitionTime":"2025-12-01T14:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.098027 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.098066 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.098074 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.098088 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.098099 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:22Z","lastTransitionTime":"2025-12-01T14:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.200756 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.200816 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.200830 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.200842 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.200851 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:22Z","lastTransitionTime":"2025-12-01T14:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.308341 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.309111 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.309142 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.309162 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.309175 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:22Z","lastTransitionTime":"2025-12-01T14:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.412740 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.412803 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.412820 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.412842 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.412855 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:22Z","lastTransitionTime":"2025-12-01T14:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.491062 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.491134 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.491162 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.491325 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:22 crc kubenswrapper[4810]: E1201 14:35:22.491452 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:22 crc kubenswrapper[4810]: E1201 14:35:22.491703 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:22 crc kubenswrapper[4810]: E1201 14:35:22.491939 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:22 crc kubenswrapper[4810]: E1201 14:35:22.491982 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.516595 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.516656 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.516675 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.516700 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.516718 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:22Z","lastTransitionTime":"2025-12-01T14:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.619444 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.619597 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.619620 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.619645 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.619663 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:22Z","lastTransitionTime":"2025-12-01T14:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.723443 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.723592 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.723619 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.723643 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.723666 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:22Z","lastTransitionTime":"2025-12-01T14:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.826442 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.826500 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.826512 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.826527 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.826540 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:22Z","lastTransitionTime":"2025-12-01T14:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.929239 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.929287 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.929298 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.929313 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:22 crc kubenswrapper[4810]: I1201 14:35:22.929325 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:22Z","lastTransitionTime":"2025-12-01T14:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.031903 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.031950 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.031962 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.031977 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.031989 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:23Z","lastTransitionTime":"2025-12-01T14:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.134600 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.134659 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.134678 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.134701 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.134772 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:23Z","lastTransitionTime":"2025-12-01T14:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.237072 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.237106 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.237115 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.237128 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.237138 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:23Z","lastTransitionTime":"2025-12-01T14:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.339673 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.339714 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.339725 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.339739 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.339749 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:23Z","lastTransitionTime":"2025-12-01T14:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.441365 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.441439 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.441457 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.441508 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.441527 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:23Z","lastTransitionTime":"2025-12-01T14:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.543606 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.543641 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.543653 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.543667 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.543676 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:23Z","lastTransitionTime":"2025-12-01T14:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.646227 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.646275 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.646292 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.646311 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.646323 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:23Z","lastTransitionTime":"2025-12-01T14:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.748872 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.748912 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.748924 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.748940 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.748954 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:23Z","lastTransitionTime":"2025-12-01T14:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.851240 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.851294 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.851309 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.851330 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.851347 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:23Z","lastTransitionTime":"2025-12-01T14:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.956674 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.956757 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.956778 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.956801 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:23 crc kubenswrapper[4810]: I1201 14:35:23.956818 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:23Z","lastTransitionTime":"2025-12-01T14:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.059678 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.059737 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.059756 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.059773 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.059783 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:24Z","lastTransitionTime":"2025-12-01T14:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.162398 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.162459 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.162500 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.162522 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.162541 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:24Z","lastTransitionTime":"2025-12-01T14:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.265162 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.265241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.265263 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.265295 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.265321 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:24Z","lastTransitionTime":"2025-12-01T14:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.368066 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.368125 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.368142 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.368165 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.368182 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:24Z","lastTransitionTime":"2025-12-01T14:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.470417 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.470452 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.470460 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.470487 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.470496 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:24Z","lastTransitionTime":"2025-12-01T14:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.490755 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.490828 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.490881 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:24 crc kubenswrapper[4810]: E1201 14:35:24.491083 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.491098 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:24 crc kubenswrapper[4810]: E1201 14:35:24.491326 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:24 crc kubenswrapper[4810]: E1201 14:35:24.491356 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:24 crc kubenswrapper[4810]: E1201 14:35:24.491789 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.509902 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.529350 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=71.529332692 podStartE2EDuration="1m11.529332692s" podCreationTimestamp="2025-12-01 14:34:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:35:24.515780342 +0000 UTC m=+90.279289955" watchObservedRunningTime="2025-12-01 14:35:24.529332692 +0000 UTC m=+90.292842295" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.529559 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=71.529555618 podStartE2EDuration="1m11.529555618s" podCreationTimestamp="2025-12-01 14:34:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:35:24.52883248 +0000 UTC m=+90.292342073" watchObservedRunningTime="2025-12-01 14:35:24.529555618 +0000 UTC m=+90.293065211" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.572216 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.572250 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.572259 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.572275 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.572285 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:24Z","lastTransitionTime":"2025-12-01T14:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.596665 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podStartSLOduration=66.596642466 podStartE2EDuration="1m6.596642466s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:35:24.561785847 +0000 UTC m=+90.325295440" watchObservedRunningTime="2025-12-01 14:35:24.596642466 +0000 UTC m=+90.360152069" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.616311 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=68.616269259 podStartE2EDuration="1m8.616269259s" podCreationTimestamp="2025-12-01 14:34:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:35:24.596785 +0000 UTC m=+90.360294623" watchObservedRunningTime="2025-12-01 14:35:24.616269259 +0000 UTC m=+90.379778922" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.668203 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-v7dlj" podStartSLOduration=66.668179012 podStartE2EDuration="1m6.668179012s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:35:24.654523719 +0000 UTC m=+90.418033342" watchObservedRunningTime="2025-12-01 14:35:24.668179012 +0000 UTC m=+90.431688635" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.674663 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.674701 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.674711 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.674743 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.674756 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:24Z","lastTransitionTime":"2025-12-01T14:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.679863 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-s4cc7" podStartSLOduration=66.679847334 podStartE2EDuration="1m6.679847334s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:35:24.66843266 +0000 UTC m=+90.431942263" watchObservedRunningTime="2025-12-01 14:35:24.679847334 +0000 UTC m=+90.443356957" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.693903 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-mlglg" podStartSLOduration=66.693885488 podStartE2EDuration="1m6.693885488s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:35:24.69321404 +0000 UTC m=+90.456723643" watchObservedRunningTime="2025-12-01 14:35:24.693885488 +0000 UTC m=+90.457395091" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.694063 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-h8m66" podStartSLOduration=66.694057742 podStartE2EDuration="1m6.694057742s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:35:24.680311336 +0000 UTC m=+90.443820949" watchObservedRunningTime="2025-12-01 14:35:24.694057742 +0000 UTC m=+90.457567345" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.705971 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=44.705953999 podStartE2EDuration="44.705953999s" podCreationTimestamp="2025-12-01 14:34:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:35:24.705925598 +0000 UTC m=+90.469435201" watchObservedRunningTime="2025-12-01 14:35:24.705953999 +0000 UTC m=+90.469463602" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.726338 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-kbfbn" podStartSLOduration=66.726322442 podStartE2EDuration="1m6.726322442s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:35:24.725869071 +0000 UTC m=+90.489378674" watchObservedRunningTime="2025-12-01 14:35:24.726322442 +0000 UTC m=+90.489832045" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.776661 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.776927 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.777029 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.777116 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.777191 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:24Z","lastTransitionTime":"2025-12-01T14:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.879523 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.879836 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.879933 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.880022 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.880111 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:24Z","lastTransitionTime":"2025-12-01T14:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.981746 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.981782 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.981792 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.981807 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:24 crc kubenswrapper[4810]: I1201 14:35:24.981817 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:24Z","lastTransitionTime":"2025-12-01T14:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.085030 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.085084 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.085097 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.085115 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.085129 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:25Z","lastTransitionTime":"2025-12-01T14:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.187368 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.187406 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.187414 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.187427 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.187436 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:25Z","lastTransitionTime":"2025-12-01T14:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.290762 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.290808 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.290824 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.290846 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.290863 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:25Z","lastTransitionTime":"2025-12-01T14:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.393063 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.393104 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.393116 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.393134 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.393147 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:25Z","lastTransitionTime":"2025-12-01T14:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.495209 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.495241 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.495250 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.495263 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.495275 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:25Z","lastTransitionTime":"2025-12-01T14:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.597534 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.597571 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.597582 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.597600 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.597613 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:25Z","lastTransitionTime":"2025-12-01T14:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.699749 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.699808 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.699827 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.699851 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.699868 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:25Z","lastTransitionTime":"2025-12-01T14:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.802735 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.802813 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.802833 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.802858 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.802877 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:25Z","lastTransitionTime":"2025-12-01T14:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.905302 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.905386 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.905412 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.905443 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:25 crc kubenswrapper[4810]: I1201 14:35:25.905466 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:25Z","lastTransitionTime":"2025-12-01T14:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.008909 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.008992 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.009015 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.009045 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.009077 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:26Z","lastTransitionTime":"2025-12-01T14:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.111612 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.111652 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.111662 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.111676 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.111685 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:26Z","lastTransitionTime":"2025-12-01T14:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.214155 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.214185 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.214193 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.214209 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.214217 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:26Z","lastTransitionTime":"2025-12-01T14:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.316083 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.316127 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.316138 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.316153 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.316167 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:26Z","lastTransitionTime":"2025-12-01T14:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.418345 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.418378 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.418389 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.418401 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.418409 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:26Z","lastTransitionTime":"2025-12-01T14:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.490456 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.490505 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.490538 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.490484 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:26 crc kubenswrapper[4810]: E1201 14:35:26.490604 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:26 crc kubenswrapper[4810]: E1201 14:35:26.490655 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:26 crc kubenswrapper[4810]: E1201 14:35:26.490697 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:26 crc kubenswrapper[4810]: E1201 14:35:26.490763 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.520196 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.520243 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.520252 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.520266 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.520274 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:26Z","lastTransitionTime":"2025-12-01T14:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.621997 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.622031 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.622040 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.622053 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.622062 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:26Z","lastTransitionTime":"2025-12-01T14:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.724401 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.724454 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.724497 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.724518 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.724531 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:26Z","lastTransitionTime":"2025-12-01T14:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.826620 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.826687 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.826705 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.826727 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.826745 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:26Z","lastTransitionTime":"2025-12-01T14:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.928860 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.928900 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.928909 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.928921 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:26 crc kubenswrapper[4810]: I1201 14:35:26.928930 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:26Z","lastTransitionTime":"2025-12-01T14:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.030646 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.030942 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.031010 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.031077 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.031153 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:27Z","lastTransitionTime":"2025-12-01T14:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.133808 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.134101 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.134166 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.134235 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.134303 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:27Z","lastTransitionTime":"2025-12-01T14:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.236629 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.236971 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.237128 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.237296 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.237433 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:27Z","lastTransitionTime":"2025-12-01T14:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.340133 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.340193 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.340205 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.340219 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.340229 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:27Z","lastTransitionTime":"2025-12-01T14:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.443182 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.443783 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.443850 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.443915 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.443982 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:27Z","lastTransitionTime":"2025-12-01T14:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.546395 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.546439 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.546450 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.546466 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.546501 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:27Z","lastTransitionTime":"2025-12-01T14:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.649675 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.649734 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.649753 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.649782 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.649805 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:27Z","lastTransitionTime":"2025-12-01T14:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.753134 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.753175 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.753189 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.753209 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.753223 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:27Z","lastTransitionTime":"2025-12-01T14:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.855947 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.856004 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.856021 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.856046 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.856064 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:27Z","lastTransitionTime":"2025-12-01T14:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.958849 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.958925 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.958948 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.958980 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:27 crc kubenswrapper[4810]: I1201 14:35:27.959006 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:27Z","lastTransitionTime":"2025-12-01T14:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.060917 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.060961 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.060975 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.060990 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.061004 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:28Z","lastTransitionTime":"2025-12-01T14:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.164005 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.164051 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.164065 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.164084 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.164099 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:28Z","lastTransitionTime":"2025-12-01T14:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.266970 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.267013 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.267025 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.267043 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.267056 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:28Z","lastTransitionTime":"2025-12-01T14:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.370154 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.370207 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.370224 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.370248 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.370267 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:28Z","lastTransitionTime":"2025-12-01T14:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.473743 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.473801 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.473811 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.473829 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.473841 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:28Z","lastTransitionTime":"2025-12-01T14:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.491102 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.491161 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.491175 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.491104 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:28 crc kubenswrapper[4810]: E1201 14:35:28.491243 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:28 crc kubenswrapper[4810]: E1201 14:35:28.491386 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:28 crc kubenswrapper[4810]: E1201 14:35:28.491728 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:28 crc kubenswrapper[4810]: E1201 14:35:28.491817 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.577024 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.577092 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.577105 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.577128 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.577144 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:28Z","lastTransitionTime":"2025-12-01T14:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.647546 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.647586 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.647597 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.647614 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.647626 4810 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T14:35:28Z","lastTransitionTime":"2025-12-01T14:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.706297 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67"] Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.706733 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.708851 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.708870 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.710289 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.711681 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.719637 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=4.719612037 podStartE2EDuration="4.719612037s" podCreationTimestamp="2025-12-01 14:35:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:35:28.718124078 +0000 UTC m=+94.481633751" watchObservedRunningTime="2025-12-01 14:35:28.719612037 +0000 UTC m=+94.483121680" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.833758 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/eb60c491-5a4a-41a1-a6a2-88204689b4ff-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ddb67\" (UID: \"eb60c491-5a4a-41a1-a6a2-88204689b4ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.833846 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb60c491-5a4a-41a1-a6a2-88204689b4ff-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ddb67\" (UID: \"eb60c491-5a4a-41a1-a6a2-88204689b4ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.833966 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/eb60c491-5a4a-41a1-a6a2-88204689b4ff-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ddb67\" (UID: \"eb60c491-5a4a-41a1-a6a2-88204689b4ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.833999 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/eb60c491-5a4a-41a1-a6a2-88204689b4ff-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ddb67\" (UID: \"eb60c491-5a4a-41a1-a6a2-88204689b4ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.834034 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb60c491-5a4a-41a1-a6a2-88204689b4ff-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ddb67\" (UID: \"eb60c491-5a4a-41a1-a6a2-88204689b4ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.935461 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/eb60c491-5a4a-41a1-a6a2-88204689b4ff-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ddb67\" (UID: \"eb60c491-5a4a-41a1-a6a2-88204689b4ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.935528 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/eb60c491-5a4a-41a1-a6a2-88204689b4ff-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ddb67\" (UID: \"eb60c491-5a4a-41a1-a6a2-88204689b4ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.935550 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb60c491-5a4a-41a1-a6a2-88204689b4ff-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ddb67\" (UID: \"eb60c491-5a4a-41a1-a6a2-88204689b4ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.935582 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/eb60c491-5a4a-41a1-a6a2-88204689b4ff-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ddb67\" (UID: \"eb60c491-5a4a-41a1-a6a2-88204689b4ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.935613 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb60c491-5a4a-41a1-a6a2-88204689b4ff-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ddb67\" (UID: \"eb60c491-5a4a-41a1-a6a2-88204689b4ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.935758 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/eb60c491-5a4a-41a1-a6a2-88204689b4ff-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ddb67\" (UID: \"eb60c491-5a4a-41a1-a6a2-88204689b4ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.936282 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/eb60c491-5a4a-41a1-a6a2-88204689b4ff-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ddb67\" (UID: \"eb60c491-5a4a-41a1-a6a2-88204689b4ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.937533 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/eb60c491-5a4a-41a1-a6a2-88204689b4ff-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ddb67\" (UID: \"eb60c491-5a4a-41a1-a6a2-88204689b4ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.944619 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb60c491-5a4a-41a1-a6a2-88204689b4ff-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ddb67\" (UID: \"eb60c491-5a4a-41a1-a6a2-88204689b4ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" Dec 01 14:35:28 crc kubenswrapper[4810]: I1201 14:35:28.958105 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb60c491-5a4a-41a1-a6a2-88204689b4ff-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ddb67\" (UID: \"eb60c491-5a4a-41a1-a6a2-88204689b4ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" Dec 01 14:35:29 crc kubenswrapper[4810]: I1201 14:35:29.028433 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" Dec 01 14:35:30 crc kubenswrapper[4810]: I1201 14:35:30.003184 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" event={"ID":"eb60c491-5a4a-41a1-a6a2-88204689b4ff","Type":"ContainerStarted","Data":"4aad3269e2acf2762932cd8915c895744589f3fa2e65db6e64ffa3ffd96de7cc"} Dec 01 14:35:30 crc kubenswrapper[4810]: I1201 14:35:30.003236 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" event={"ID":"eb60c491-5a4a-41a1-a6a2-88204689b4ff","Type":"ContainerStarted","Data":"d5ca3835044f38d18762ec1e4249b2e96f9c8a3d068a723f66ce804421b81287"} Dec 01 14:35:30 crc kubenswrapper[4810]: I1201 14:35:30.023137 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ddb67" podStartSLOduration=72.023120954 podStartE2EDuration="1m12.023120954s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:35:30.022864627 +0000 UTC m=+95.786374240" watchObservedRunningTime="2025-12-01 14:35:30.023120954 +0000 UTC m=+95.786630557" Dec 01 14:35:30 crc kubenswrapper[4810]: I1201 14:35:30.490158 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:30 crc kubenswrapper[4810]: I1201 14:35:30.490298 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:30 crc kubenswrapper[4810]: E1201 14:35:30.490532 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:30 crc kubenswrapper[4810]: I1201 14:35:30.490612 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:30 crc kubenswrapper[4810]: E1201 14:35:30.490333 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:30 crc kubenswrapper[4810]: E1201 14:35:30.490717 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:30 crc kubenswrapper[4810]: I1201 14:35:30.490812 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:30 crc kubenswrapper[4810]: E1201 14:35:30.490967 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:32 crc kubenswrapper[4810]: I1201 14:35:32.491276 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:32 crc kubenswrapper[4810]: I1201 14:35:32.491353 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:32 crc kubenswrapper[4810]: I1201 14:35:32.491449 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:32 crc kubenswrapper[4810]: I1201 14:35:32.491363 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:32 crc kubenswrapper[4810]: E1201 14:35:32.491645 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:32 crc kubenswrapper[4810]: E1201 14:35:32.491791 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:32 crc kubenswrapper[4810]: E1201 14:35:32.491941 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:32 crc kubenswrapper[4810]: E1201 14:35:32.492086 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:34 crc kubenswrapper[4810]: I1201 14:35:34.492862 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:34 crc kubenswrapper[4810]: I1201 14:35:34.492920 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:34 crc kubenswrapper[4810]: I1201 14:35:34.492983 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:34 crc kubenswrapper[4810]: I1201 14:35:34.492938 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:34 crc kubenswrapper[4810]: E1201 14:35:34.493105 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:34 crc kubenswrapper[4810]: E1201 14:35:34.493258 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:34 crc kubenswrapper[4810]: E1201 14:35:34.493681 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:34 crc kubenswrapper[4810]: E1201 14:35:34.493754 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:34 crc kubenswrapper[4810]: I1201 14:35:34.494012 4810 scope.go:117] "RemoveContainer" containerID="d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe" Dec 01 14:35:34 crc kubenswrapper[4810]: E1201 14:35:34.494232 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" Dec 01 14:35:36 crc kubenswrapper[4810]: I1201 14:35:36.490625 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:36 crc kubenswrapper[4810]: I1201 14:35:36.490762 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:36 crc kubenswrapper[4810]: I1201 14:35:36.490815 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:36 crc kubenswrapper[4810]: I1201 14:35:36.490673 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:36 crc kubenswrapper[4810]: E1201 14:35:36.490921 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:36 crc kubenswrapper[4810]: E1201 14:35:36.491150 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:36 crc kubenswrapper[4810]: E1201 14:35:36.491267 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:36 crc kubenswrapper[4810]: E1201 14:35:36.491429 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:36 crc kubenswrapper[4810]: I1201 14:35:36.844644 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs\") pod \"network-metrics-daemon-jlwwg\" (UID: \"d379bcb2-a35d-470f-894c-8629d5023df8\") " pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:36 crc kubenswrapper[4810]: E1201 14:35:36.844827 4810 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 14:35:36 crc kubenswrapper[4810]: E1201 14:35:36.844901 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs podName:d379bcb2-a35d-470f-894c-8629d5023df8 nodeName:}" failed. No retries permitted until 2025-12-01 14:36:40.844885144 +0000 UTC m=+166.608394747 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs") pod "network-metrics-daemon-jlwwg" (UID: "d379bcb2-a35d-470f-894c-8629d5023df8") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 14:35:38 crc kubenswrapper[4810]: I1201 14:35:38.490752 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:38 crc kubenswrapper[4810]: E1201 14:35:38.490980 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:38 crc kubenswrapper[4810]: I1201 14:35:38.491057 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:38 crc kubenswrapper[4810]: I1201 14:35:38.491161 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:38 crc kubenswrapper[4810]: I1201 14:35:38.491203 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:38 crc kubenswrapper[4810]: E1201 14:35:38.491809 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:38 crc kubenswrapper[4810]: E1201 14:35:38.491582 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:38 crc kubenswrapper[4810]: E1201 14:35:38.491376 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:40 crc kubenswrapper[4810]: I1201 14:35:40.491054 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:40 crc kubenswrapper[4810]: E1201 14:35:40.491167 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:40 crc kubenswrapper[4810]: I1201 14:35:40.491337 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:40 crc kubenswrapper[4810]: E1201 14:35:40.491378 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:40 crc kubenswrapper[4810]: I1201 14:35:40.491506 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:40 crc kubenswrapper[4810]: E1201 14:35:40.491553 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:40 crc kubenswrapper[4810]: I1201 14:35:40.492780 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:40 crc kubenswrapper[4810]: E1201 14:35:40.492990 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:42 crc kubenswrapper[4810]: I1201 14:35:42.490834 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:42 crc kubenswrapper[4810]: I1201 14:35:42.490902 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:42 crc kubenswrapper[4810]: I1201 14:35:42.490958 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:42 crc kubenswrapper[4810]: E1201 14:35:42.491036 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:42 crc kubenswrapper[4810]: I1201 14:35:42.491134 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:42 crc kubenswrapper[4810]: E1201 14:35:42.491336 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:42 crc kubenswrapper[4810]: E1201 14:35:42.491466 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:42 crc kubenswrapper[4810]: E1201 14:35:42.491842 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:44 crc kubenswrapper[4810]: I1201 14:35:44.490506 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:44 crc kubenswrapper[4810]: I1201 14:35:44.490572 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:44 crc kubenswrapper[4810]: E1201 14:35:44.490700 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:44 crc kubenswrapper[4810]: I1201 14:35:44.490749 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:44 crc kubenswrapper[4810]: I1201 14:35:44.490759 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:44 crc kubenswrapper[4810]: E1201 14:35:44.490881 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:44 crc kubenswrapper[4810]: E1201 14:35:44.504072 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:44 crc kubenswrapper[4810]: E1201 14:35:44.505730 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:46 crc kubenswrapper[4810]: I1201 14:35:46.491172 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:46 crc kubenswrapper[4810]: I1201 14:35:46.491183 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:46 crc kubenswrapper[4810]: E1201 14:35:46.491397 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:46 crc kubenswrapper[4810]: I1201 14:35:46.491200 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:46 crc kubenswrapper[4810]: E1201 14:35:46.491602 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:46 crc kubenswrapper[4810]: E1201 14:35:46.491749 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:46 crc kubenswrapper[4810]: I1201 14:35:46.492640 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:46 crc kubenswrapper[4810]: E1201 14:35:46.493010 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:47 crc kubenswrapper[4810]: I1201 14:35:47.492163 4810 scope.go:117] "RemoveContainer" containerID="d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe" Dec 01 14:35:47 crc kubenswrapper[4810]: E1201 14:35:47.492567 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" Dec 01 14:35:48 crc kubenswrapper[4810]: I1201 14:35:48.490823 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:48 crc kubenswrapper[4810]: I1201 14:35:48.490895 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:48 crc kubenswrapper[4810]: I1201 14:35:48.490938 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:48 crc kubenswrapper[4810]: I1201 14:35:48.490995 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:48 crc kubenswrapper[4810]: E1201 14:35:48.491114 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:48 crc kubenswrapper[4810]: E1201 14:35:48.491190 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:48 crc kubenswrapper[4810]: E1201 14:35:48.491412 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:48 crc kubenswrapper[4810]: E1201 14:35:48.491507 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:50 crc kubenswrapper[4810]: I1201 14:35:50.491123 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:50 crc kubenswrapper[4810]: E1201 14:35:50.491400 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:50 crc kubenswrapper[4810]: I1201 14:35:50.492400 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:50 crc kubenswrapper[4810]: I1201 14:35:50.492625 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:50 crc kubenswrapper[4810]: I1201 14:35:50.492663 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:50 crc kubenswrapper[4810]: E1201 14:35:50.493437 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:50 crc kubenswrapper[4810]: E1201 14:35:50.493546 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:50 crc kubenswrapper[4810]: E1201 14:35:50.493643 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:52 crc kubenswrapper[4810]: I1201 14:35:52.080533 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-s4cc7_df7ecc53-6478-49cd-bfb6-6ee80e850a19/kube-multus/1.log" Dec 01 14:35:52 crc kubenswrapper[4810]: I1201 14:35:52.080983 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-s4cc7_df7ecc53-6478-49cd-bfb6-6ee80e850a19/kube-multus/0.log" Dec 01 14:35:52 crc kubenswrapper[4810]: I1201 14:35:52.081033 4810 generic.go:334] "Generic (PLEG): container finished" podID="df7ecc53-6478-49cd-bfb6-6ee80e850a19" containerID="bd96824c21af3e98dc4565197e8c1e036e7a99ce8855d92ff01b023cea04e7d5" exitCode=1 Dec 01 14:35:52 crc kubenswrapper[4810]: I1201 14:35:52.081080 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-s4cc7" event={"ID":"df7ecc53-6478-49cd-bfb6-6ee80e850a19","Type":"ContainerDied","Data":"bd96824c21af3e98dc4565197e8c1e036e7a99ce8855d92ff01b023cea04e7d5"} Dec 01 14:35:52 crc kubenswrapper[4810]: I1201 14:35:52.081262 4810 scope.go:117] "RemoveContainer" containerID="70d32c80ad5756d75bb4087bfd2e472706ce1da7893164358aa01ba52d0246dd" Dec 01 14:35:52 crc kubenswrapper[4810]: I1201 14:35:52.081963 4810 scope.go:117] "RemoveContainer" containerID="bd96824c21af3e98dc4565197e8c1e036e7a99ce8855d92ff01b023cea04e7d5" Dec 01 14:35:52 crc kubenswrapper[4810]: E1201 14:35:52.082303 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-s4cc7_openshift-multus(df7ecc53-6478-49cd-bfb6-6ee80e850a19)\"" pod="openshift-multus/multus-s4cc7" podUID="df7ecc53-6478-49cd-bfb6-6ee80e850a19" Dec 01 14:35:52 crc kubenswrapper[4810]: I1201 14:35:52.490826 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:52 crc kubenswrapper[4810]: I1201 14:35:52.490903 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:52 crc kubenswrapper[4810]: I1201 14:35:52.490956 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:52 crc kubenswrapper[4810]: I1201 14:35:52.490978 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:52 crc kubenswrapper[4810]: E1201 14:35:52.491641 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:52 crc kubenswrapper[4810]: E1201 14:35:52.491910 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:52 crc kubenswrapper[4810]: E1201 14:35:52.492172 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:52 crc kubenswrapper[4810]: E1201 14:35:52.492426 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:53 crc kubenswrapper[4810]: I1201 14:35:53.086440 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-s4cc7_df7ecc53-6478-49cd-bfb6-6ee80e850a19/kube-multus/1.log" Dec 01 14:35:54 crc kubenswrapper[4810]: I1201 14:35:54.490620 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:54 crc kubenswrapper[4810]: I1201 14:35:54.490660 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:54 crc kubenswrapper[4810]: I1201 14:35:54.490734 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:54 crc kubenswrapper[4810]: E1201 14:35:54.491701 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:54 crc kubenswrapper[4810]: I1201 14:35:54.491774 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:54 crc kubenswrapper[4810]: E1201 14:35:54.491928 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:54 crc kubenswrapper[4810]: E1201 14:35:54.492059 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:54 crc kubenswrapper[4810]: E1201 14:35:54.492140 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:54 crc kubenswrapper[4810]: E1201 14:35:54.519680 4810 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 01 14:35:54 crc kubenswrapper[4810]: E1201 14:35:54.562417 4810 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 14:35:56 crc kubenswrapper[4810]: I1201 14:35:56.490578 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:56 crc kubenswrapper[4810]: I1201 14:35:56.490586 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:56 crc kubenswrapper[4810]: E1201 14:35:56.490988 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:56 crc kubenswrapper[4810]: I1201 14:35:56.490652 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:56 crc kubenswrapper[4810]: I1201 14:35:56.490613 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:56 crc kubenswrapper[4810]: E1201 14:35:56.491094 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:56 crc kubenswrapper[4810]: E1201 14:35:56.491231 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:56 crc kubenswrapper[4810]: E1201 14:35:56.491339 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:58 crc kubenswrapper[4810]: I1201 14:35:58.490945 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:35:58 crc kubenswrapper[4810]: I1201 14:35:58.491018 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:35:58 crc kubenswrapper[4810]: I1201 14:35:58.491070 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:35:58 crc kubenswrapper[4810]: I1201 14:35:58.490956 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:35:58 crc kubenswrapper[4810]: E1201 14:35:58.491194 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:35:58 crc kubenswrapper[4810]: E1201 14:35:58.491344 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:35:58 crc kubenswrapper[4810]: E1201 14:35:58.492075 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:35:58 crc kubenswrapper[4810]: E1201 14:35:58.492362 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:35:58 crc kubenswrapper[4810]: I1201 14:35:58.492732 4810 scope.go:117] "RemoveContainer" containerID="d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe" Dec 01 14:35:58 crc kubenswrapper[4810]: E1201 14:35:58.493013 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-p7pvm_openshift-ovn-kubernetes(efba5556-97da-447f-9dda-40c2ea6d3e3d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" Dec 01 14:35:59 crc kubenswrapper[4810]: E1201 14:35:59.563686 4810 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 14:36:00 crc kubenswrapper[4810]: I1201 14:36:00.490708 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:36:00 crc kubenswrapper[4810]: I1201 14:36:00.490779 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:36:00 crc kubenswrapper[4810]: I1201 14:36:00.490842 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:36:00 crc kubenswrapper[4810]: I1201 14:36:00.490741 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:36:00 crc kubenswrapper[4810]: E1201 14:36:00.490917 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:36:00 crc kubenswrapper[4810]: E1201 14:36:00.491064 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:36:00 crc kubenswrapper[4810]: E1201 14:36:00.491189 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:36:00 crc kubenswrapper[4810]: E1201 14:36:00.491273 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:36:02 crc kubenswrapper[4810]: I1201 14:36:02.491053 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:36:02 crc kubenswrapper[4810]: I1201 14:36:02.491135 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:36:02 crc kubenswrapper[4810]: I1201 14:36:02.491088 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:36:02 crc kubenswrapper[4810]: I1201 14:36:02.491085 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:36:02 crc kubenswrapper[4810]: E1201 14:36:02.491257 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:36:02 crc kubenswrapper[4810]: E1201 14:36:02.491344 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:36:02 crc kubenswrapper[4810]: E1201 14:36:02.491432 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:36:02 crc kubenswrapper[4810]: E1201 14:36:02.491549 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:36:04 crc kubenswrapper[4810]: I1201 14:36:04.491065 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:36:04 crc kubenswrapper[4810]: I1201 14:36:04.491122 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:36:04 crc kubenswrapper[4810]: E1201 14:36:04.491289 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:36:04 crc kubenswrapper[4810]: I1201 14:36:04.491319 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:36:04 crc kubenswrapper[4810]: I1201 14:36:04.491359 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:36:04 crc kubenswrapper[4810]: E1201 14:36:04.491722 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:36:04 crc kubenswrapper[4810]: E1201 14:36:04.493947 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:36:04 crc kubenswrapper[4810]: E1201 14:36:04.494046 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:36:04 crc kubenswrapper[4810]: E1201 14:36:04.564681 4810 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 14:36:06 crc kubenswrapper[4810]: I1201 14:36:06.491027 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:36:06 crc kubenswrapper[4810]: I1201 14:36:06.491075 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:36:06 crc kubenswrapper[4810]: I1201 14:36:06.491130 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:36:06 crc kubenswrapper[4810]: E1201 14:36:06.491163 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:36:06 crc kubenswrapper[4810]: E1201 14:36:06.491317 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:36:06 crc kubenswrapper[4810]: I1201 14:36:06.491368 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:36:06 crc kubenswrapper[4810]: E1201 14:36:06.491429 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:36:06 crc kubenswrapper[4810]: E1201 14:36:06.491618 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:36:06 crc kubenswrapper[4810]: I1201 14:36:06.491952 4810 scope.go:117] "RemoveContainer" containerID="bd96824c21af3e98dc4565197e8c1e036e7a99ce8855d92ff01b023cea04e7d5" Dec 01 14:36:07 crc kubenswrapper[4810]: I1201 14:36:07.141887 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-s4cc7_df7ecc53-6478-49cd-bfb6-6ee80e850a19/kube-multus/1.log" Dec 01 14:36:07 crc kubenswrapper[4810]: I1201 14:36:07.141948 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-s4cc7" event={"ID":"df7ecc53-6478-49cd-bfb6-6ee80e850a19","Type":"ContainerStarted","Data":"13444fe245e31b72033991aab5b46b4bb2bbb61d1f5f07f09473175dfa7c1464"} Dec 01 14:36:08 crc kubenswrapper[4810]: I1201 14:36:08.490511 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:36:08 crc kubenswrapper[4810]: I1201 14:36:08.490561 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:36:08 crc kubenswrapper[4810]: E1201 14:36:08.490718 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:36:08 crc kubenswrapper[4810]: I1201 14:36:08.490744 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:36:08 crc kubenswrapper[4810]: I1201 14:36:08.490752 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:36:08 crc kubenswrapper[4810]: E1201 14:36:08.490847 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:36:08 crc kubenswrapper[4810]: E1201 14:36:08.490945 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:36:08 crc kubenswrapper[4810]: E1201 14:36:08.491089 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:36:09 crc kubenswrapper[4810]: E1201 14:36:09.566024 4810 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 14:36:10 crc kubenswrapper[4810]: I1201 14:36:10.490873 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:36:10 crc kubenswrapper[4810]: I1201 14:36:10.491074 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:36:10 crc kubenswrapper[4810]: I1201 14:36:10.491513 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:36:10 crc kubenswrapper[4810]: I1201 14:36:10.491531 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:36:10 crc kubenswrapper[4810]: E1201 14:36:10.491673 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:36:10 crc kubenswrapper[4810]: E1201 14:36:10.491710 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:36:10 crc kubenswrapper[4810]: E1201 14:36:10.491755 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:36:10 crc kubenswrapper[4810]: E1201 14:36:10.491798 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:36:10 crc kubenswrapper[4810]: I1201 14:36:10.492276 4810 scope.go:117] "RemoveContainer" containerID="d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe" Dec 01 14:36:11 crc kubenswrapper[4810]: I1201 14:36:11.156299 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovnkube-controller/3.log" Dec 01 14:36:11 crc kubenswrapper[4810]: I1201 14:36:11.157956 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerStarted","Data":"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4"} Dec 01 14:36:11 crc kubenswrapper[4810]: I1201 14:36:11.158329 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:36:11 crc kubenswrapper[4810]: I1201 14:36:11.186811 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podStartSLOduration=113.186790649 podStartE2EDuration="1m53.186790649s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:11.185634478 +0000 UTC m=+136.949144091" watchObservedRunningTime="2025-12-01 14:36:11.186790649 +0000 UTC m=+136.950300262" Dec 01 14:36:11 crc kubenswrapper[4810]: I1201 14:36:11.247730 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-jlwwg"] Dec 01 14:36:11 crc kubenswrapper[4810]: I1201 14:36:11.247847 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:36:11 crc kubenswrapper[4810]: E1201 14:36:11.247945 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:36:12 crc kubenswrapper[4810]: I1201 14:36:12.490593 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:36:12 crc kubenswrapper[4810]: I1201 14:36:12.490687 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:36:12 crc kubenswrapper[4810]: E1201 14:36:12.490772 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:36:12 crc kubenswrapper[4810]: I1201 14:36:12.490849 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:36:12 crc kubenswrapper[4810]: I1201 14:36:12.490900 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:36:12 crc kubenswrapper[4810]: E1201 14:36:12.490995 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:36:12 crc kubenswrapper[4810]: E1201 14:36:12.491121 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:36:12 crc kubenswrapper[4810]: E1201 14:36:12.491258 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:36:14 crc kubenswrapper[4810]: I1201 14:36:14.490923 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:36:14 crc kubenswrapper[4810]: I1201 14:36:14.490976 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:36:14 crc kubenswrapper[4810]: I1201 14:36:14.491066 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:36:14 crc kubenswrapper[4810]: I1201 14:36:14.491167 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:36:14 crc kubenswrapper[4810]: E1201 14:36:14.492349 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 14:36:14 crc kubenswrapper[4810]: E1201 14:36:14.492394 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 14:36:14 crc kubenswrapper[4810]: E1201 14:36:14.492409 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jlwwg" podUID="d379bcb2-a35d-470f-894c-8629d5023df8" Dec 01 14:36:14 crc kubenswrapper[4810]: E1201 14:36:14.492424 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 14:36:16 crc kubenswrapper[4810]: I1201 14:36:16.491107 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:36:16 crc kubenswrapper[4810]: I1201 14:36:16.491205 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:36:16 crc kubenswrapper[4810]: I1201 14:36:16.491228 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:36:16 crc kubenswrapper[4810]: I1201 14:36:16.491314 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:36:16 crc kubenswrapper[4810]: I1201 14:36:16.498124 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 01 14:36:16 crc kubenswrapper[4810]: I1201 14:36:16.498265 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 01 14:36:16 crc kubenswrapper[4810]: I1201 14:36:16.498675 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 01 14:36:16 crc kubenswrapper[4810]: I1201 14:36:16.498804 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 01 14:36:16 crc kubenswrapper[4810]: I1201 14:36:16.499662 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 01 14:36:16 crc kubenswrapper[4810]: I1201 14:36:16.500099 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.003707 4810 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.041072 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pnv2c"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.041894 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.043664 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.044058 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.045486 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-r2vhz"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.045858 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-r2vhz" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.047094 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.052028 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.052092 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.052118 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.052129 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.052210 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.052278 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.052441 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.052453 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.052545 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.052715 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.052754 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.053026 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.055194 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.055449 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.055563 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.055580 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.056092 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4rgpg"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.056736 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.058641 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.058816 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.060324 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.060997 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.061291 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fcdlt"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.062076 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.065884 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.075670 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.076504 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.079411 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.091584 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.091740 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.091826 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.092044 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.095090 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-rf6tj"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.095424 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xxkbc"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.095650 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5267z"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.095781 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-rf6tj" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.095989 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5267z" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.096283 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xxkbc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.097743 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.098011 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.098445 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.098580 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.099232 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.099919 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.099936 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qvm7d"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.101088 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.103341 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2mxj4"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.104052 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-2mxj4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.104713 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.105269 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.105530 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-pjt5m"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.107168 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.112317 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fdwcc"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.112601 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-8tcbx"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.112827 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.113123 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.113563 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.113633 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pjt5m" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.113635 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.116897 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.116992 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.117029 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.116904 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.117129 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.117297 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.117311 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.117405 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.117502 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.117626 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.117661 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.117723 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.117779 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.117867 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.118277 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.118849 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.119207 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.119350 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.119704 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.119910 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.120251 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.120432 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.120791 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.121721 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n6q5s"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.122709 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n6q5s" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.127369 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.127621 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.127896 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6qzb"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.128699 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9r9fc"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.128975 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.129335 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.129361 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.129463 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.129740 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.129852 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.129971 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.130077 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.130176 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.130219 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.130338 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.130456 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.130625 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.130857 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.131160 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6qzb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.131354 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.131445 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wvgpw"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.131384 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9r9fc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.131950 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.132080 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.132567 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.132714 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.132661 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.139192 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.139363 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.139617 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.139781 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.139869 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2781f96f-4018-4ccf-adb8-04512de46450-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.139919 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2781f96f-4018-4ccf-adb8-04512de46450-serving-cert\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.139988 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2781f96f-4018-4ccf-adb8-04512de46450-config\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.140015 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2781f96f-4018-4ccf-adb8-04512de46450-etcd-client\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.140058 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2781f96f-4018-4ccf-adb8-04512de46450-etcd-serving-ca\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.140081 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2781f96f-4018-4ccf-adb8-04512de46450-audit-dir\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.140097 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkf9f\" (UniqueName: \"kubernetes.io/projected/2781f96f-4018-4ccf-adb8-04512de46450-kube-api-access-fkf9f\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.140115 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2781f96f-4018-4ccf-adb8-04512de46450-audit\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.140122 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.140149 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2781f96f-4018-4ccf-adb8-04512de46450-encryption-config\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.140173 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2781f96f-4018-4ccf-adb8-04512de46450-node-pullsecrets\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.140269 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2781f96f-4018-4ccf-adb8-04512de46450-image-import-ca\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.140763 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.141001 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.141022 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cgzqr"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.141047 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.141433 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.141653 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.141866 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.141596 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wvgpw" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.142169 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.142589 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.142860 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.143043 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.143064 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.143114 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.143160 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.154214 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.154439 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.159223 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.168497 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-h54rq"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.173076 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-hsnbk"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.173539 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.173794 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h54rq" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.173854 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-tlqrt"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.174362 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-tlqrt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.177885 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.189835 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.191306 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n67g2"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.192163 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-t7pxb"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.192668 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.192684 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n67g2" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.197039 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.201613 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-rf6tj"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.203607 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.204655 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-s5rbr"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.204748 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.205242 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s5rbr" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.205333 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.205971 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.206191 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.207621 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5hpvd"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.207937 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.208098 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.211343 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.218207 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.218767 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.219110 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zkr9c"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.219266 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-5hpvd" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.219347 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.219923 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-hqhng"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.221623 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.221947 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-hqhng" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.222354 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zkr9c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.226019 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jnrsh"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.226666 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.226687 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xtf78"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.226704 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.226810 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jnrsh" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.254754 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xtf78" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.256272 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257027 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2781f96f-4018-4ccf-adb8-04512de46450-serving-cert\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257078 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d2540546-e6b2-421d-8680-327f9cdeca7d-images\") pod \"machine-config-operator-74547568cd-xlw6n\" (UID: \"d2540546-e6b2-421d-8680-327f9cdeca7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257104 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2zhn\" (UniqueName: \"kubernetes.io/projected/ef1bca29-62d7-422b-890a-f93b965c0934-kube-api-access-x2zhn\") pod \"controller-manager-879f6c89f-4rgpg\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257127 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54284225-c472-482c-93b2-2426031f0957-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9r9fc\" (UID: \"54284225-c472-482c-93b2-2426031f0957\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9r9fc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257168 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257188 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s666l\" (UniqueName: \"kubernetes.io/projected/1394009a-f690-4e2f-8086-87b9483fabbd-kube-api-access-s666l\") pod \"olm-operator-6b444d44fb-2zf5q\" (UID: \"1394009a-f690-4e2f-8086-87b9483fabbd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257206 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2781f96f-4018-4ccf-adb8-04512de46450-config\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257224 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257241 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5537fef3-e146-4eac-bc97-3edc211a2b3d-config\") pod \"machine-approver-56656f9798-84qz9\" (UID: \"5537fef3-e146-4eac-bc97-3edc211a2b3d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257258 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6855b17f-5a68-45c7-a1b4-0c9cacc6ed14-metrics-tls\") pod \"ingress-operator-5b745b69d9-l56lb\" (UID: \"6855b17f-5a68-45c7-a1b4-0c9cacc6ed14\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257276 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f010e33-46e5-4080-a24f-47850bee04cd-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6qzb\" (UID: \"4f010e33-46e5-4080-a24f-47850bee04cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6qzb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257291 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0fa71e97-6be0-4300-bef2-b2dffbf30076-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cgzqr\" (UID: \"0fa71e97-6be0-4300-bef2-b2dffbf30076\") " pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257305 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddafebe9-9403-44c2-9d32-2867e6c212ce-config\") pod \"console-operator-58897d9998-2mxj4\" (UID: \"ddafebe9-9403-44c2-9d32-2867e6c212ce\") " pod="openshift-console-operator/console-operator-58897d9998-2mxj4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257320 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzm8b\" (UniqueName: \"kubernetes.io/projected/5537fef3-e146-4eac-bc97-3edc211a2b3d-kube-api-access-xzm8b\") pod \"machine-approver-56656f9798-84qz9\" (UID: \"5537fef3-e146-4eac-bc97-3edc211a2b3d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257334 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54284225-c472-482c-93b2-2426031f0957-config\") pod \"kube-controller-manager-operator-78b949d7b-9r9fc\" (UID: \"54284225-c472-482c-93b2-2426031f0957\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9r9fc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257351 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2781f96f-4018-4ccf-adb8-04512de46450-etcd-client\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257367 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257386 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws2bl\" (UniqueName: \"kubernetes.io/projected/88f6bb4e-362c-425c-9f67-81f76ed05e42-kube-api-access-ws2bl\") pod \"downloads-7954f5f757-rf6tj\" (UID: \"88f6bb4e-362c-425c-9f67-81f76ed05e42\") " pod="openshift-console/downloads-7954f5f757-rf6tj" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257406 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2781f96f-4018-4ccf-adb8-04512de46450-etcd-serving-ca\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257422 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/16c6f469-3ac7-46e3-84f3-a42187626e2b-etcd-client\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257440 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-audit-policies\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257458 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7e7e3fa5-5c91-4694-95f4-67a503c3f42c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-r2vhz\" (UID: \"7e7e3fa5-5c91-4694-95f4-67a503c3f42c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r2vhz" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257504 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6855b17f-5a68-45c7-a1b4-0c9cacc6ed14-trusted-ca\") pod \"ingress-operator-5b745b69d9-l56lb\" (UID: \"6855b17f-5a68-45c7-a1b4-0c9cacc6ed14\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257538 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2781f96f-4018-4ccf-adb8-04512de46450-encryption-config\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257563 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/367566c7-8e32-444f-866a-d64b98a2f596-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-5267z\" (UID: \"367566c7-8e32-444f-866a-d64b98a2f596\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5267z" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257580 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-serving-cert\") pod \"route-controller-manager-6576b87f9c-gn5f4\" (UID: \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257596 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2781f96f-4018-4ccf-adb8-04512de46450-node-pullsecrets\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257613 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2781f96f-4018-4ccf-adb8-04512de46450-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257626 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrz5s\" (UniqueName: \"kubernetes.io/projected/367566c7-8e32-444f-866a-d64b98a2f596-kube-api-access-vrz5s\") pod \"cluster-samples-operator-665b6dd947-5267z\" (UID: \"367566c7-8e32-444f-866a-d64b98a2f596\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5267z" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257641 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257657 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0fa71e97-6be0-4300-bef2-b2dffbf30076-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cgzqr\" (UID: \"0fa71e97-6be0-4300-bef2-b2dffbf30076\") " pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257673 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3e6111a1-b6ac-4a66-b8f6-efef4463b6df-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-pjt5m\" (UID: \"3e6111a1-b6ac-4a66-b8f6-efef4463b6df\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pjt5m" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257687 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a543c4e2-99d9-4b47-a9c6-927c8737ed49-config\") pod \"kube-apiserver-operator-766d6c64bb-n6q5s\" (UID: \"a543c4e2-99d9-4b47-a9c6-927c8737ed49\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n6q5s" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257702 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5537fef3-e146-4eac-bc97-3edc211a2b3d-auth-proxy-config\") pod \"machine-approver-56656f9798-84qz9\" (UID: \"5537fef3-e146-4eac-bc97-3edc211a2b3d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257716 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr6t8\" (UniqueName: \"kubernetes.io/projected/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-kube-api-access-sr6t8\") pod \"route-controller-manager-6576b87f9c-gn5f4\" (UID: \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257735 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e7e3fa5-5c91-4694-95f4-67a503c3f42c-config\") pod \"machine-api-operator-5694c8668f-r2vhz\" (UID: \"7e7e3fa5-5c91-4694-95f4-67a503c3f42c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r2vhz" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257750 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddafebe9-9403-44c2-9d32-2867e6c212ce-serving-cert\") pod \"console-operator-58897d9998-2mxj4\" (UID: \"ddafebe9-9403-44c2-9d32-2867e6c212ce\") " pod="openshift-console-operator/console-operator-58897d9998-2mxj4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257766 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-client-ca\") pod \"route-controller-manager-6576b87f9c-gn5f4\" (UID: \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257781 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e00f933-d775-4766-ae08-0d300d09fa4a-serving-cert\") pod \"authentication-operator-69f744f599-fcdlt\" (UID: \"8e00f933-d775-4766-ae08-0d300d09fa4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257796 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5537fef3-e146-4eac-bc97-3edc211a2b3d-machine-approver-tls\") pod \"machine-approver-56656f9798-84qz9\" (UID: \"5537fef3-e146-4eac-bc97-3edc211a2b3d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257811 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ft22\" (UniqueName: \"kubernetes.io/projected/6855b17f-5a68-45c7-a1b4-0c9cacc6ed14-kube-api-access-5ft22\") pod \"ingress-operator-5b745b69d9-l56lb\" (UID: \"6855b17f-5a68-45c7-a1b4-0c9cacc6ed14\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257828 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257844 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257860 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrkzf\" (UniqueName: \"kubernetes.io/projected/0fa71e97-6be0-4300-bef2-b2dffbf30076-kube-api-access-lrkzf\") pod \"marketplace-operator-79b997595-cgzqr\" (UID: \"0fa71e97-6be0-4300-bef2-b2dffbf30076\") " pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257876 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef1bca29-62d7-422b-890a-f93b965c0934-serving-cert\") pod \"controller-manager-879f6c89f-4rgpg\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257891 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7e7e3fa5-5c91-4694-95f4-67a503c3f42c-images\") pod \"machine-api-operator-5694c8668f-r2vhz\" (UID: \"7e7e3fa5-5c91-4694-95f4-67a503c3f42c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r2vhz" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257908 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257924 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ef1bca29-62d7-422b-890a-f93b965c0934-client-ca\") pod \"controller-manager-879f6c89f-4rgpg\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257938 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4f010e33-46e5-4080-a24f-47850bee04cd-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6qzb\" (UID: \"4f010e33-46e5-4080-a24f-47850bee04cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6qzb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257961 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8jpx\" (UniqueName: \"kubernetes.io/projected/8e00f933-d775-4766-ae08-0d300d09fa4a-kube-api-access-z8jpx\") pod \"authentication-operator-69f744f599-fcdlt\" (UID: \"8e00f933-d775-4766-ae08-0d300d09fa4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257979 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jpfm\" (UniqueName: \"kubernetes.io/projected/d2540546-e6b2-421d-8680-327f9cdeca7d-kube-api-access-2jpfm\") pod \"machine-config-operator-74547568cd-xlw6n\" (UID: \"d2540546-e6b2-421d-8680-327f9cdeca7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257996 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a543c4e2-99d9-4b47-a9c6-927c8737ed49-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-n6q5s\" (UID: \"a543c4e2-99d9-4b47-a9c6-927c8737ed49\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n6q5s" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258012 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-k6w5z\" (UID: \"ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258027 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqm5b\" (UniqueName: \"kubernetes.io/projected/7e7e3fa5-5c91-4694-95f4-67a503c3f42c-kube-api-access-fqm5b\") pod \"machine-api-operator-5694c8668f-r2vhz\" (UID: \"7e7e3fa5-5c91-4694-95f4-67a503c3f42c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r2vhz" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258042 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/16c6f469-3ac7-46e3-84f3-a42187626e2b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258056 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ps7xs\" (UniqueName: \"kubernetes.io/projected/3e6111a1-b6ac-4a66-b8f6-efef4463b6df-kube-api-access-ps7xs\") pod \"machine-config-controller-84d6567774-pjt5m\" (UID: \"3e6111a1-b6ac-4a66-b8f6-efef4463b6df\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pjt5m" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258078 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hbgk\" (UniqueName: \"kubernetes.io/projected/ddafebe9-9403-44c2-9d32-2867e6c212ce-kube-api-access-2hbgk\") pod \"console-operator-58897d9998-2mxj4\" (UID: \"ddafebe9-9403-44c2-9d32-2867e6c212ce\") " pod="openshift-console-operator/console-operator-58897d9998-2mxj4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258093 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/16c6f469-3ac7-46e3-84f3-a42187626e2b-audit-policies\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258113 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv2tf\" (UniqueName: \"kubernetes.io/projected/93b94518-da01-439a-9f46-2329f86020a4-kube-api-access-nv2tf\") pod \"openshift-apiserver-operator-796bbdcf4f-xxkbc\" (UID: \"93b94518-da01-439a-9f46-2329f86020a4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xxkbc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258135 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e00f933-d775-4766-ae08-0d300d09fa4a-service-ca-bundle\") pod \"authentication-operator-69f744f599-fcdlt\" (UID: \"8e00f933-d775-4766-ae08-0d300d09fa4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258150 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258164 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/16c6f469-3ac7-46e3-84f3-a42187626e2b-encryption-config\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258180 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2781f96f-4018-4ccf-adb8-04512de46450-audit-dir\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258195 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e00f933-d775-4766-ae08-0d300d09fa4a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fcdlt\" (UID: \"8e00f933-d775-4766-ae08-0d300d09fa4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258212 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258227 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-k6w5z\" (UID: \"ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258283 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-config\") pod \"route-controller-manager-6576b87f9c-gn5f4\" (UID: \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258302 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkf9f\" (UniqueName: \"kubernetes.io/projected/2781f96f-4018-4ccf-adb8-04512de46450-kube-api-access-fkf9f\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258317 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258340 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1394009a-f690-4e2f-8086-87b9483fabbd-srv-cert\") pod \"olm-operator-6b444d44fb-2zf5q\" (UID: \"1394009a-f690-4e2f-8086-87b9483fabbd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258355 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2781f96f-4018-4ccf-adb8-04512de46450-audit\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258376 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef1bca29-62d7-422b-890a-f93b965c0934-config\") pod \"controller-manager-879f6c89f-4rgpg\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258392 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2781f96f-4018-4ccf-adb8-04512de46450-image-import-ca\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258412 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258432 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54284225-c472-482c-93b2-2426031f0957-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9r9fc\" (UID: \"54284225-c472-482c-93b2-2426031f0957\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9r9fc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258453 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93b94518-da01-439a-9f46-2329f86020a4-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xxkbc\" (UID: \"93b94518-da01-439a-9f46-2329f86020a4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xxkbc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258486 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6855b17f-5a68-45c7-a1b4-0c9cacc6ed14-bound-sa-token\") pod \"ingress-operator-5b745b69d9-l56lb\" (UID: \"6855b17f-5a68-45c7-a1b4-0c9cacc6ed14\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258504 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a543c4e2-99d9-4b47-a9c6-927c8737ed49-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-n6q5s\" (UID: \"a543c4e2-99d9-4b47-a9c6-927c8737ed49\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n6q5s" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258521 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/16c6f469-3ac7-46e3-84f3-a42187626e2b-audit-dir\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258538 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s956t\" (UniqueName: \"kubernetes.io/projected/16c6f469-3ac7-46e3-84f3-a42187626e2b-kube-api-access-s956t\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258552 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f010e33-46e5-4080-a24f-47850bee04cd-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6qzb\" (UID: \"4f010e33-46e5-4080-a24f-47850bee04cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6qzb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258567 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/16c6f469-3ac7-46e3-84f3-a42187626e2b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258584 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3e6111a1-b6ac-4a66-b8f6-efef4463b6df-proxy-tls\") pod \"machine-config-controller-84d6567774-pjt5m\" (UID: \"3e6111a1-b6ac-4a66-b8f6-efef4463b6df\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pjt5m" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258622 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d2540546-e6b2-421d-8680-327f9cdeca7d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xlw6n\" (UID: \"d2540546-e6b2-421d-8680-327f9cdeca7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258638 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ef1bca29-62d7-422b-890a-f93b965c0934-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4rgpg\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258656 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncfbq\" (UniqueName: \"kubernetes.io/projected/ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19-kube-api-access-ncfbq\") pod \"cluster-image-registry-operator-dc59b4c8b-k6w5z\" (UID: \"ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258672 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93b94518-da01-439a-9f46-2329f86020a4-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xxkbc\" (UID: \"93b94518-da01-439a-9f46-2329f86020a4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xxkbc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258687 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-audit-dir\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258704 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5xpn\" (UniqueName: \"kubernetes.io/projected/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-kube-api-access-j5xpn\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258720 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1394009a-f690-4e2f-8086-87b9483fabbd-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2zf5q\" (UID: \"1394009a-f690-4e2f-8086-87b9483fabbd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258741 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d2540546-e6b2-421d-8680-327f9cdeca7d-proxy-tls\") pod \"machine-config-operator-74547568cd-xlw6n\" (UID: \"d2540546-e6b2-421d-8680-327f9cdeca7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258762 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ddafebe9-9403-44c2-9d32-2867e6c212ce-trusted-ca\") pod \"console-operator-58897d9998-2mxj4\" (UID: \"ddafebe9-9403-44c2-9d32-2867e6c212ce\") " pod="openshift-console-operator/console-operator-58897d9998-2mxj4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258777 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16c6f469-3ac7-46e3-84f3-a42187626e2b-serving-cert\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258792 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-k6w5z\" (UID: \"ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.258816 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e00f933-d775-4766-ae08-0d300d09fa4a-config\") pod \"authentication-operator-69f744f599-fcdlt\" (UID: \"8e00f933-d775-4766-ae08-0d300d09fa4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.259564 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.259862 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2781f96f-4018-4ccf-adb8-04512de46450-node-pullsecrets\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.257592 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.260711 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2781f96f-4018-4ccf-adb8-04512de46450-audit\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.261069 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2781f96f-4018-4ccf-adb8-04512de46450-audit-dir\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.261347 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5267z"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.261368 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4rgpg"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.261432 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.261784 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2781f96f-4018-4ccf-adb8-04512de46450-image-import-ca\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.262366 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2781f96f-4018-4ccf-adb8-04512de46450-etcd-serving-ca\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.263198 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2781f96f-4018-4ccf-adb8-04512de46450-trusted-ca-bundle\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.263911 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2781f96f-4018-4ccf-adb8-04512de46450-config\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.263515 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-dt9dw"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.265411 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9r9fc"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.265434 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xxkbc"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.265540 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dt9dw" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.269606 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.269744 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2mxj4"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.270605 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2781f96f-4018-4ccf-adb8-04512de46450-etcd-client\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.270671 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pnv2c"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.273908 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-pjt5m"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.273954 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qvm7d"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.273965 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fdwcc"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.275252 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.275863 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2781f96f-4018-4ccf-adb8-04512de46450-serving-cert\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.278570 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.278600 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-r2vhz"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.283452 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.283509 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.285090 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pdv9r"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.286881 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-tlqrt"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.287262 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.287260 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n67g2"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.287358 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.289522 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-s5rbr"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.289792 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5hpvd"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.291490 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2781f96f-4018-4ccf-adb8-04512de46450-encryption-config\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.292600 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-t7pxb"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.295431 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fcdlt"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.295483 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dt9dw"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.295496 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.299904 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.299950 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-h54rq"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.300917 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-hqhng"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.302804 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-88hcl"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.303391 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-88hcl" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.303560 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.304802 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wvgpw"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.306769 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-5rxzl"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.307511 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-5rxzl" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.308723 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.310691 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6qzb"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.311812 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n6q5s"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.313082 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zkr9c"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.314183 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-8tcbx"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.315695 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.317322 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cgzqr"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.318596 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jnrsh"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.319842 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xtf78"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.321313 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pdv9r"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.323211 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.323496 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.324383 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-88hcl"] Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.345614 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.359231 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4f010e33-46e5-4080-a24f-47850bee04cd-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6qzb\" (UID: \"4f010e33-46e5-4080-a24f-47850bee04cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6qzb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.359415 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ef1bca29-62d7-422b-890a-f93b965c0934-client-ca\") pod \"controller-manager-879f6c89f-4rgpg\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.359548 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8jpx\" (UniqueName: \"kubernetes.io/projected/8e00f933-d775-4766-ae08-0d300d09fa4a-kube-api-access-z8jpx\") pod \"authentication-operator-69f744f599-fcdlt\" (UID: \"8e00f933-d775-4766-ae08-0d300d09fa4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.359631 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jpfm\" (UniqueName: \"kubernetes.io/projected/d2540546-e6b2-421d-8680-327f9cdeca7d-kube-api-access-2jpfm\") pod \"machine-config-operator-74547568cd-xlw6n\" (UID: \"d2540546-e6b2-421d-8680-327f9cdeca7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.359699 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a543c4e2-99d9-4b47-a9c6-927c8737ed49-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-n6q5s\" (UID: \"a543c4e2-99d9-4b47-a9c6-927c8737ed49\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n6q5s" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.359782 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-k6w5z\" (UID: \"ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.359894 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqm5b\" (UniqueName: \"kubernetes.io/projected/7e7e3fa5-5c91-4694-95f4-67a503c3f42c-kube-api-access-fqm5b\") pod \"machine-api-operator-5694c8668f-r2vhz\" (UID: \"7e7e3fa5-5c91-4694-95f4-67a503c3f42c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r2vhz" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.360000 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/16c6f469-3ac7-46e3-84f3-a42187626e2b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.360101 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ps7xs\" (UniqueName: \"kubernetes.io/projected/3e6111a1-b6ac-4a66-b8f6-efef4463b6df-kube-api-access-ps7xs\") pod \"machine-config-controller-84d6567774-pjt5m\" (UID: \"3e6111a1-b6ac-4a66-b8f6-efef4463b6df\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pjt5m" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.360209 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hbgk\" (UniqueName: \"kubernetes.io/projected/ddafebe9-9403-44c2-9d32-2867e6c212ce-kube-api-access-2hbgk\") pod \"console-operator-58897d9998-2mxj4\" (UID: \"ddafebe9-9403-44c2-9d32-2867e6c212ce\") " pod="openshift-console-operator/console-operator-58897d9998-2mxj4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.360301 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/16c6f469-3ac7-46e3-84f3-a42187626e2b-audit-policies\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.360380 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv2tf\" (UniqueName: \"kubernetes.io/projected/93b94518-da01-439a-9f46-2329f86020a4-kube-api-access-nv2tf\") pod \"openshift-apiserver-operator-796bbdcf4f-xxkbc\" (UID: \"93b94518-da01-439a-9f46-2329f86020a4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xxkbc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.360455 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e00f933-d775-4766-ae08-0d300d09fa4a-service-ca-bundle\") pod \"authentication-operator-69f744f599-fcdlt\" (UID: \"8e00f933-d775-4766-ae08-0d300d09fa4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.360566 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.360642 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/16c6f469-3ac7-46e3-84f3-a42187626e2b-encryption-config\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.360716 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e00f933-d775-4766-ae08-0d300d09fa4a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fcdlt\" (UID: \"8e00f933-d775-4766-ae08-0d300d09fa4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.360793 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/16c6f469-3ac7-46e3-84f3-a42187626e2b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.360421 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ef1bca29-62d7-422b-890a-f93b965c0934-client-ca\") pod \"controller-manager-879f6c89f-4rgpg\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.361384 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/16c6f469-3ac7-46e3-84f3-a42187626e2b-audit-policies\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.361467 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e00f933-d775-4766-ae08-0d300d09fa4a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fcdlt\" (UID: \"8e00f933-d775-4766-ae08-0d300d09fa4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.361660 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.361799 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-k6w5z\" (UID: \"ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.361918 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.361695 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e00f933-d775-4766-ae08-0d300d09fa4a-service-ca-bundle\") pod \"authentication-operator-69f744f599-fcdlt\" (UID: \"8e00f933-d775-4766-ae08-0d300d09fa4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.361936 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-config\") pod \"route-controller-manager-6576b87f9c-gn5f4\" (UID: \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.362128 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-k6w5z\" (UID: \"ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.362264 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.362368 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1394009a-f690-4e2f-8086-87b9483fabbd-srv-cert\") pod \"olm-operator-6b444d44fb-2zf5q\" (UID: \"1394009a-f690-4e2f-8086-87b9483fabbd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.362515 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef1bca29-62d7-422b-890a-f93b965c0934-config\") pod \"controller-manager-879f6c89f-4rgpg\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.362627 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.362724 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54284225-c472-482c-93b2-2426031f0957-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9r9fc\" (UID: \"54284225-c472-482c-93b2-2426031f0957\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9r9fc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.362823 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93b94518-da01-439a-9f46-2329f86020a4-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xxkbc\" (UID: \"93b94518-da01-439a-9f46-2329f86020a4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xxkbc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363184 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6855b17f-5a68-45c7-a1b4-0c9cacc6ed14-bound-sa-token\") pod \"ingress-operator-5b745b69d9-l56lb\" (UID: \"6855b17f-5a68-45c7-a1b4-0c9cacc6ed14\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.362785 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a543c4e2-99d9-4b47-a9c6-927c8737ed49-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-n6q5s\" (UID: \"a543c4e2-99d9-4b47-a9c6-927c8737ed49\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n6q5s" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363317 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a543c4e2-99d9-4b47-a9c6-927c8737ed49-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-n6q5s\" (UID: \"a543c4e2-99d9-4b47-a9c6-927c8737ed49\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n6q5s" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363393 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/16c6f469-3ac7-46e3-84f3-a42187626e2b-audit-dir\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363446 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/16c6f469-3ac7-46e3-84f3-a42187626e2b-encryption-config\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363461 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s956t\" (UniqueName: \"kubernetes.io/projected/16c6f469-3ac7-46e3-84f3-a42187626e2b-kube-api-access-s956t\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363461 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/16c6f469-3ac7-46e3-84f3-a42187626e2b-audit-dir\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363505 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f010e33-46e5-4080-a24f-47850bee04cd-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6qzb\" (UID: \"4f010e33-46e5-4080-a24f-47850bee04cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6qzb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363535 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/16c6f469-3ac7-46e3-84f3-a42187626e2b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363557 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3e6111a1-b6ac-4a66-b8f6-efef4463b6df-proxy-tls\") pod \"machine-config-controller-84d6567774-pjt5m\" (UID: \"3e6111a1-b6ac-4a66-b8f6-efef4463b6df\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pjt5m" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363577 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d2540546-e6b2-421d-8680-327f9cdeca7d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xlw6n\" (UID: \"d2540546-e6b2-421d-8680-327f9cdeca7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363597 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ef1bca29-62d7-422b-890a-f93b965c0934-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4rgpg\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363615 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncfbq\" (UniqueName: \"kubernetes.io/projected/ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19-kube-api-access-ncfbq\") pod \"cluster-image-registry-operator-dc59b4c8b-k6w5z\" (UID: \"ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363631 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93b94518-da01-439a-9f46-2329f86020a4-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xxkbc\" (UID: \"93b94518-da01-439a-9f46-2329f86020a4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xxkbc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363648 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d2540546-e6b2-421d-8680-327f9cdeca7d-proxy-tls\") pod \"machine-config-operator-74547568cd-xlw6n\" (UID: \"d2540546-e6b2-421d-8680-327f9cdeca7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363664 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-audit-dir\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363680 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5xpn\" (UniqueName: \"kubernetes.io/projected/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-kube-api-access-j5xpn\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363697 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1394009a-f690-4e2f-8086-87b9483fabbd-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2zf5q\" (UID: \"1394009a-f690-4e2f-8086-87b9483fabbd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363723 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e00f933-d775-4766-ae08-0d300d09fa4a-config\") pod \"authentication-operator-69f744f599-fcdlt\" (UID: \"8e00f933-d775-4766-ae08-0d300d09fa4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363739 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ddafebe9-9403-44c2-9d32-2867e6c212ce-trusted-ca\") pod \"console-operator-58897d9998-2mxj4\" (UID: \"ddafebe9-9403-44c2-9d32-2867e6c212ce\") " pod="openshift-console-operator/console-operator-58897d9998-2mxj4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363756 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16c6f469-3ac7-46e3-84f3-a42187626e2b-serving-cert\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363773 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-k6w5z\" (UID: \"ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363795 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d2540546-e6b2-421d-8680-327f9cdeca7d-images\") pod \"machine-config-operator-74547568cd-xlw6n\" (UID: \"d2540546-e6b2-421d-8680-327f9cdeca7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363811 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2zhn\" (UniqueName: \"kubernetes.io/projected/ef1bca29-62d7-422b-890a-f93b965c0934-kube-api-access-x2zhn\") pod \"controller-manager-879f6c89f-4rgpg\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363827 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54284225-c472-482c-93b2-2426031f0957-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9r9fc\" (UID: \"54284225-c472-482c-93b2-2426031f0957\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9r9fc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363848 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363865 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s666l\" (UniqueName: \"kubernetes.io/projected/1394009a-f690-4e2f-8086-87b9483fabbd-kube-api-access-s666l\") pod \"olm-operator-6b444d44fb-2zf5q\" (UID: \"1394009a-f690-4e2f-8086-87b9483fabbd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363885 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363917 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5537fef3-e146-4eac-bc97-3edc211a2b3d-config\") pod \"machine-approver-56656f9798-84qz9\" (UID: \"5537fef3-e146-4eac-bc97-3edc211a2b3d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363934 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6855b17f-5a68-45c7-a1b4-0c9cacc6ed14-metrics-tls\") pod \"ingress-operator-5b745b69d9-l56lb\" (UID: \"6855b17f-5a68-45c7-a1b4-0c9cacc6ed14\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363954 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363969 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f010e33-46e5-4080-a24f-47850bee04cd-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6qzb\" (UID: \"4f010e33-46e5-4080-a24f-47850bee04cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6qzb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.363985 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0fa71e97-6be0-4300-bef2-b2dffbf30076-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cgzqr\" (UID: \"0fa71e97-6be0-4300-bef2-b2dffbf30076\") " pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364001 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddafebe9-9403-44c2-9d32-2867e6c212ce-config\") pod \"console-operator-58897d9998-2mxj4\" (UID: \"ddafebe9-9403-44c2-9d32-2867e6c212ce\") " pod="openshift-console-operator/console-operator-58897d9998-2mxj4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364017 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzm8b\" (UniqueName: \"kubernetes.io/projected/5537fef3-e146-4eac-bc97-3edc211a2b3d-kube-api-access-xzm8b\") pod \"machine-approver-56656f9798-84qz9\" (UID: \"5537fef3-e146-4eac-bc97-3edc211a2b3d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364033 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54284225-c472-482c-93b2-2426031f0957-config\") pod \"kube-controller-manager-operator-78b949d7b-9r9fc\" (UID: \"54284225-c472-482c-93b2-2426031f0957\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9r9fc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364060 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws2bl\" (UniqueName: \"kubernetes.io/projected/88f6bb4e-362c-425c-9f67-81f76ed05e42-kube-api-access-ws2bl\") pod \"downloads-7954f5f757-rf6tj\" (UID: \"88f6bb4e-362c-425c-9f67-81f76ed05e42\") " pod="openshift-console/downloads-7954f5f757-rf6tj" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364075 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/16c6f469-3ac7-46e3-84f3-a42187626e2b-etcd-client\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364089 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/16c6f469-3ac7-46e3-84f3-a42187626e2b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364093 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-audit-policies\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364145 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7e7e3fa5-5c91-4694-95f4-67a503c3f42c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-r2vhz\" (UID: \"7e7e3fa5-5c91-4694-95f4-67a503c3f42c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r2vhz" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364164 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6855b17f-5a68-45c7-a1b4-0c9cacc6ed14-trusted-ca\") pod \"ingress-operator-5b745b69d9-l56lb\" (UID: \"6855b17f-5a68-45c7-a1b4-0c9cacc6ed14\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364199 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/367566c7-8e32-444f-866a-d64b98a2f596-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-5267z\" (UID: \"367566c7-8e32-444f-866a-d64b98a2f596\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5267z" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364217 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-serving-cert\") pod \"route-controller-manager-6576b87f9c-gn5f4\" (UID: \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364236 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrz5s\" (UniqueName: \"kubernetes.io/projected/367566c7-8e32-444f-866a-d64b98a2f596-kube-api-access-vrz5s\") pod \"cluster-samples-operator-665b6dd947-5267z\" (UID: \"367566c7-8e32-444f-866a-d64b98a2f596\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5267z" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364252 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364268 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0fa71e97-6be0-4300-bef2-b2dffbf30076-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cgzqr\" (UID: \"0fa71e97-6be0-4300-bef2-b2dffbf30076\") " pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364287 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a543c4e2-99d9-4b47-a9c6-927c8737ed49-config\") pod \"kube-apiserver-operator-766d6c64bb-n6q5s\" (UID: \"a543c4e2-99d9-4b47-a9c6-927c8737ed49\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n6q5s" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364302 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5537fef3-e146-4eac-bc97-3edc211a2b3d-auth-proxy-config\") pod \"machine-approver-56656f9798-84qz9\" (UID: \"5537fef3-e146-4eac-bc97-3edc211a2b3d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364317 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr6t8\" (UniqueName: \"kubernetes.io/projected/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-kube-api-access-sr6t8\") pod \"route-controller-manager-6576b87f9c-gn5f4\" (UID: \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364335 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3e6111a1-b6ac-4a66-b8f6-efef4463b6df-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-pjt5m\" (UID: \"3e6111a1-b6ac-4a66-b8f6-efef4463b6df\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pjt5m" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364287 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d2540546-e6b2-421d-8680-327f9cdeca7d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xlw6n\" (UID: \"d2540546-e6b2-421d-8680-327f9cdeca7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364354 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e7e3fa5-5c91-4694-95f4-67a503c3f42c-config\") pod \"machine-api-operator-5694c8668f-r2vhz\" (UID: \"7e7e3fa5-5c91-4694-95f4-67a503c3f42c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r2vhz" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364401 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddafebe9-9403-44c2-9d32-2867e6c212ce-serving-cert\") pod \"console-operator-58897d9998-2mxj4\" (UID: \"ddafebe9-9403-44c2-9d32-2867e6c212ce\") " pod="openshift-console-operator/console-operator-58897d9998-2mxj4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364414 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef1bca29-62d7-422b-890a-f93b965c0934-config\") pod \"controller-manager-879f6c89f-4rgpg\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364428 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e00f933-d775-4766-ae08-0d300d09fa4a-serving-cert\") pod \"authentication-operator-69f744f599-fcdlt\" (UID: \"8e00f933-d775-4766-ae08-0d300d09fa4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364453 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-client-ca\") pod \"route-controller-manager-6576b87f9c-gn5f4\" (UID: \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364491 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5537fef3-e146-4eac-bc97-3edc211a2b3d-machine-approver-tls\") pod \"machine-approver-56656f9798-84qz9\" (UID: \"5537fef3-e146-4eac-bc97-3edc211a2b3d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364514 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ft22\" (UniqueName: \"kubernetes.io/projected/6855b17f-5a68-45c7-a1b4-0c9cacc6ed14-kube-api-access-5ft22\") pod \"ingress-operator-5b745b69d9-l56lb\" (UID: \"6855b17f-5a68-45c7-a1b4-0c9cacc6ed14\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364528 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-config\") pod \"route-controller-manager-6576b87f9c-gn5f4\" (UID: \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364542 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7e7e3fa5-5c91-4694-95f4-67a503c3f42c-images\") pod \"machine-api-operator-5694c8668f-r2vhz\" (UID: \"7e7e3fa5-5c91-4694-95f4-67a503c3f42c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r2vhz" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364568 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364595 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364620 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364644 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrkzf\" (UniqueName: \"kubernetes.io/projected/0fa71e97-6be0-4300-bef2-b2dffbf30076-kube-api-access-lrkzf\") pod \"marketplace-operator-79b997595-cgzqr\" (UID: \"0fa71e97-6be0-4300-bef2-b2dffbf30076\") " pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364668 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef1bca29-62d7-422b-890a-f93b965c0934-serving-cert\") pod \"controller-manager-879f6c89f-4rgpg\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364707 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-audit-policies\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.364748 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d2540546-e6b2-421d-8680-327f9cdeca7d-images\") pod \"machine-config-operator-74547568cd-xlw6n\" (UID: \"d2540546-e6b2-421d-8680-327f9cdeca7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.365043 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e7e3fa5-5c91-4694-95f4-67a503c3f42c-config\") pod \"machine-api-operator-5694c8668f-r2vhz\" (UID: \"7e7e3fa5-5c91-4694-95f4-67a503c3f42c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r2vhz" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.365099 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.365556 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ef1bca29-62d7-422b-890a-f93b965c0934-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4rgpg\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.365623 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddafebe9-9403-44c2-9d32-2867e6c212ce-config\") pod \"console-operator-58897d9998-2mxj4\" (UID: \"ddafebe9-9403-44c2-9d32-2867e6c212ce\") " pod="openshift-console-operator/console-operator-58897d9998-2mxj4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.365784 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.366363 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3e6111a1-b6ac-4a66-b8f6-efef4463b6df-proxy-tls\") pod \"machine-config-controller-84d6567774-pjt5m\" (UID: \"3e6111a1-b6ac-4a66-b8f6-efef4463b6df\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pjt5m" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.366759 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.367098 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.367932 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7e7e3fa5-5c91-4694-95f4-67a503c3f42c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-r2vhz\" (UID: \"7e7e3fa5-5c91-4694-95f4-67a503c3f42c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r2vhz" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.368423 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a543c4e2-99d9-4b47-a9c6-927c8737ed49-config\") pod \"kube-apiserver-operator-766d6c64bb-n6q5s\" (UID: \"a543c4e2-99d9-4b47-a9c6-927c8737ed49\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n6q5s" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.368438 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.368873 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.368943 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5537fef3-e146-4eac-bc97-3edc211a2b3d-auth-proxy-config\") pod \"machine-approver-56656f9798-84qz9\" (UID: \"5537fef3-e146-4eac-bc97-3edc211a2b3d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.369212 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93b94518-da01-439a-9f46-2329f86020a4-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xxkbc\" (UID: \"93b94518-da01-439a-9f46-2329f86020a4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xxkbc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.369233 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1394009a-f690-4e2f-8086-87b9483fabbd-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2zf5q\" (UID: \"1394009a-f690-4e2f-8086-87b9483fabbd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.369259 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-audit-dir\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.369523 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.369862 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3e6111a1-b6ac-4a66-b8f6-efef4463b6df-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-pjt5m\" (UID: \"3e6111a1-b6ac-4a66-b8f6-efef4463b6df\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pjt5m" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.370164 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.370343 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.370395 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ddafebe9-9403-44c2-9d32-2867e6c212ce-trusted-ca\") pod \"console-operator-58897d9998-2mxj4\" (UID: \"ddafebe9-9403-44c2-9d32-2867e6c212ce\") " pod="openshift-console-operator/console-operator-58897d9998-2mxj4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.371214 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5537fef3-e146-4eac-bc97-3edc211a2b3d-config\") pod \"machine-approver-56656f9798-84qz9\" (UID: \"5537fef3-e146-4eac-bc97-3edc211a2b3d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.371419 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7e7e3fa5-5c91-4694-95f4-67a503c3f42c-images\") pod \"machine-api-operator-5694c8668f-r2vhz\" (UID: \"7e7e3fa5-5c91-4694-95f4-67a503c3f42c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r2vhz" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.371717 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddafebe9-9403-44c2-9d32-2867e6c212ce-serving-cert\") pod \"console-operator-58897d9998-2mxj4\" (UID: \"ddafebe9-9403-44c2-9d32-2867e6c212ce\") " pod="openshift-console-operator/console-operator-58897d9998-2mxj4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.371919 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e00f933-d775-4766-ae08-0d300d09fa4a-config\") pod \"authentication-operator-69f744f599-fcdlt\" (UID: \"8e00f933-d775-4766-ae08-0d300d09fa4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.371960 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6855b17f-5a68-45c7-a1b4-0c9cacc6ed14-trusted-ca\") pod \"ingress-operator-5b745b69d9-l56lb\" (UID: \"6855b17f-5a68-45c7-a1b4-0c9cacc6ed14\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.372591 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-client-ca\") pod \"route-controller-manager-6576b87f9c-gn5f4\" (UID: \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.372617 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d2540546-e6b2-421d-8680-327f9cdeca7d-proxy-tls\") pod \"machine-config-operator-74547568cd-xlw6n\" (UID: \"d2540546-e6b2-421d-8680-327f9cdeca7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.373433 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5537fef3-e146-4eac-bc97-3edc211a2b3d-machine-approver-tls\") pod \"machine-approver-56656f9798-84qz9\" (UID: \"5537fef3-e146-4eac-bc97-3edc211a2b3d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.373685 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-k6w5z\" (UID: \"ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.373714 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6855b17f-5a68-45c7-a1b4-0c9cacc6ed14-metrics-tls\") pod \"ingress-operator-5b745b69d9-l56lb\" (UID: \"6855b17f-5a68-45c7-a1b4-0c9cacc6ed14\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.373962 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-serving-cert\") pod \"route-controller-manager-6576b87f9c-gn5f4\" (UID: \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.374170 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/367566c7-8e32-444f-866a-d64b98a2f596-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-5267z\" (UID: \"367566c7-8e32-444f-866a-d64b98a2f596\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5267z" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.374174 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93b94518-da01-439a-9f46-2329f86020a4-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xxkbc\" (UID: \"93b94518-da01-439a-9f46-2329f86020a4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xxkbc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.374316 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.374617 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16c6f469-3ac7-46e3-84f3-a42187626e2b-serving-cert\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.374965 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e00f933-d775-4766-ae08-0d300d09fa4a-serving-cert\") pod \"authentication-operator-69f744f599-fcdlt\" (UID: \"8e00f933-d775-4766-ae08-0d300d09fa4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.375102 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.375363 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/16c6f469-3ac7-46e3-84f3-a42187626e2b-etcd-client\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.375721 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef1bca29-62d7-422b-890a-f93b965c0934-serving-cert\") pod \"controller-manager-879f6c89f-4rgpg\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.384671 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.396535 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1394009a-f690-4e2f-8086-87b9483fabbd-srv-cert\") pod \"olm-operator-6b444d44fb-2zf5q\" (UID: \"1394009a-f690-4e2f-8086-87b9483fabbd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.405153 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.425086 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.444079 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.464172 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.484411 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.496679 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f010e33-46e5-4080-a24f-47850bee04cd-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6qzb\" (UID: \"4f010e33-46e5-4080-a24f-47850bee04cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6qzb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.504706 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.511255 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f010e33-46e5-4080-a24f-47850bee04cd-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6qzb\" (UID: \"4f010e33-46e5-4080-a24f-47850bee04cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6qzb" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.524092 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.544086 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.546139 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54284225-c472-482c-93b2-2426031f0957-config\") pod \"kube-controller-manager-operator-78b949d7b-9r9fc\" (UID: \"54284225-c472-482c-93b2-2426031f0957\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9r9fc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.564198 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.583895 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.587819 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54284225-c472-482c-93b2-2426031f0957-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9r9fc\" (UID: \"54284225-c472-482c-93b2-2426031f0957\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9r9fc" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.603890 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.623816 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.645738 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.653866 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0fa71e97-6be0-4300-bef2-b2dffbf30076-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-cgzqr\" (UID: \"0fa71e97-6be0-4300-bef2-b2dffbf30076\") " pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.672974 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.679560 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0fa71e97-6be0-4300-bef2-b2dffbf30076-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-cgzqr\" (UID: \"0fa71e97-6be0-4300-bef2-b2dffbf30076\") " pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.685038 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.704400 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.724224 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.744928 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.765284 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.784886 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.825168 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.844466 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.864670 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.884519 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.904390 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.924512 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.944996 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.963810 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 01 14:36:20 crc kubenswrapper[4810]: I1201 14:36:20.985081 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.004210 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.027418 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.044786 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.064509 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.084908 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.104649 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.144734 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.164635 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.176961 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:21 crc kubenswrapper[4810]: E1201 14:36:21.177079 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:38:23.177056421 +0000 UTC m=+268.940566044 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.177121 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.178276 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.185052 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.203247 4810 request.go:700] Waited for 1.010286324s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-etcd-operator/secrets?fieldSelector=metadata.name%3Detcd-client&limit=500&resourceVersion=0 Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.205456 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.224618 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.244356 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.264457 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.278674 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.278750 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.278868 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.281953 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.281984 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.283771 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.284402 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.304681 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.324343 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.324503 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.335880 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.344695 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.350700 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.364628 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.384694 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.405659 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.424632 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.445401 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.464213 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.484394 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.504595 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.525254 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 01 14:36:21 crc kubenswrapper[4810]: W1201 14:36:21.530840 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-1fb92c3589a0a8adf48e9a06fc3c362aed29937450b8d51d993e0a1859adafe3 WatchSource:0}: Error finding container 1fb92c3589a0a8adf48e9a06fc3c362aed29937450b8d51d993e0a1859adafe3: Status 404 returned error can't find the container with id 1fb92c3589a0a8adf48e9a06fc3c362aed29937450b8d51d993e0a1859adafe3 Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.544696 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 01 14:36:21 crc kubenswrapper[4810]: W1201 14:36:21.551595 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-b884db075e2fd8456f77cd63b88706c78b152d180b876f6e562e770bf57e9e3f WatchSource:0}: Error finding container b884db075e2fd8456f77cd63b88706c78b152d180b876f6e562e770bf57e9e3f: Status 404 returned error can't find the container with id b884db075e2fd8456f77cd63b88706c78b152d180b876f6e562e770bf57e9e3f Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.564812 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.584602 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.604446 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.625042 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.644444 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.663910 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.685163 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.704941 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.724060 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.743832 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.764261 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.784167 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.825115 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.844450 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.865355 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.884608 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.904491 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.924011 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.944315 4810 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.964777 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 01 14:36:21 crc kubenswrapper[4810]: I1201 14:36:21.984856 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.004567 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.023756 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.045136 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.052743 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkf9f\" (UniqueName: \"kubernetes.io/projected/2781f96f-4018-4ccf-adb8-04512de46450-kube-api-access-fkf9f\") pod \"apiserver-76f77b778f-pnv2c\" (UID: \"2781f96f-4018-4ccf-adb8-04512de46450\") " pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.064619 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.085248 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.105038 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.142218 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4f010e33-46e5-4080-a24f-47850bee04cd-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6qzb\" (UID: \"4f010e33-46e5-4080-a24f-47850bee04cd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6qzb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.160568 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8jpx\" (UniqueName: \"kubernetes.io/projected/8e00f933-d775-4766-ae08-0d300d09fa4a-kube-api-access-z8jpx\") pod \"authentication-operator-69f744f599-fcdlt\" (UID: \"8e00f933-d775-4766-ae08-0d300d09fa4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.168665 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.183537 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jpfm\" (UniqueName: \"kubernetes.io/projected/d2540546-e6b2-421d-8680-327f9cdeca7d-kube-api-access-2jpfm\") pod \"machine-config-operator-74547568cd-xlw6n\" (UID: \"d2540546-e6b2-421d-8680-327f9cdeca7d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.198943 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"3b1b016d63707742d6d21e34bf271e1f843e4ac7f0e67c309290d45e4b35e1e0"} Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.198982 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"1fb92c3589a0a8adf48e9a06fc3c362aed29937450b8d51d993e0a1859adafe3"} Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.200640 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqm5b\" (UniqueName: \"kubernetes.io/projected/7e7e3fa5-5c91-4694-95f4-67a503c3f42c-kube-api-access-fqm5b\") pod \"machine-api-operator-5694c8668f-r2vhz\" (UID: \"7e7e3fa5-5c91-4694-95f4-67a503c3f42c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r2vhz" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.201542 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"f43a41aaa2840dcef2389928fc95a909d50054e1d10cd393ed1b1f393fe28e6f"} Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.201586 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"b884db075e2fd8456f77cd63b88706c78b152d180b876f6e562e770bf57e9e3f"} Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.203188 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a83c5fcedbe7dd64c788e99372907ac51480671c1fafd6ee26c5211c9c046c3e"} Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.203232 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d7e5adfb25d9db9f31bfb5b959b62c8739312e024058e5ab9fd89620634d9072"} Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.203761 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.207304 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-r2vhz" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.217608 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ps7xs\" (UniqueName: \"kubernetes.io/projected/3e6111a1-b6ac-4a66-b8f6-efef4463b6df-kube-api-access-ps7xs\") pod \"machine-config-controller-84d6567774-pjt5m\" (UID: \"3e6111a1-b6ac-4a66-b8f6-efef4463b6df\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pjt5m" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.221092 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pjt5m" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.222897 4810 request.go:700] Waited for 1.862344977s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver-operator/serviceaccounts/openshift-apiserver-operator/token Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.240872 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv2tf\" (UniqueName: \"kubernetes.io/projected/93b94518-da01-439a-9f46-2329f86020a4-kube-api-access-nv2tf\") pod \"openshift-apiserver-operator-796bbdcf4f-xxkbc\" (UID: \"93b94518-da01-439a-9f46-2329f86020a4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xxkbc" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.254656 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.261796 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hbgk\" (UniqueName: \"kubernetes.io/projected/ddafebe9-9403-44c2-9d32-2867e6c212ce-kube-api-access-2hbgk\") pod \"console-operator-58897d9998-2mxj4\" (UID: \"ddafebe9-9403-44c2-9d32-2867e6c212ce\") " pod="openshift-console-operator/console-operator-58897d9998-2mxj4" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.284435 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-k6w5z\" (UID: \"ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.299379 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6qzb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.304208 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54284225-c472-482c-93b2-2426031f0957-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9r9fc\" (UID: \"54284225-c472-482c-93b2-2426031f0957\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9r9fc" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.327174 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6855b17f-5a68-45c7-a1b4-0c9cacc6ed14-bound-sa-token\") pod \"ingress-operator-5b745b69d9-l56lb\" (UID: \"6855b17f-5a68-45c7-a1b4-0c9cacc6ed14\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.346448 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s956t\" (UniqueName: \"kubernetes.io/projected/16c6f469-3ac7-46e3-84f3-a42187626e2b-kube-api-access-s956t\") pod \"apiserver-7bbb656c7d-rvtjn\" (UID: \"16c6f469-3ac7-46e3-84f3-a42187626e2b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.367448 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5xpn\" (UniqueName: \"kubernetes.io/projected/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-kube-api-access-j5xpn\") pod \"oauth-openshift-558db77b4-fdwcc\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.369082 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-pnv2c"] Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.379293 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzm8b\" (UniqueName: \"kubernetes.io/projected/5537fef3-e146-4eac-bc97-3edc211a2b3d-kube-api-access-xzm8b\") pod \"machine-approver-56656f9798-84qz9\" (UID: \"5537fef3-e146-4eac-bc97-3edc211a2b3d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.410286 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xxkbc" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.417399 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2zhn\" (UniqueName: \"kubernetes.io/projected/ef1bca29-62d7-422b-890a-f93b965c0934-kube-api-access-x2zhn\") pod \"controller-manager-879f6c89f-4rgpg\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.426411 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncfbq\" (UniqueName: \"kubernetes.io/projected/ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19-kube-api-access-ncfbq\") pod \"cluster-image-registry-operator-dc59b4c8b-k6w5z\" (UID: \"ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.436198 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-2mxj4" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.438919 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a543c4e2-99d9-4b47-a9c6-927c8737ed49-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-n6q5s\" (UID: \"a543c4e2-99d9-4b47-a9c6-927c8737ed49\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n6q5s" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.449566 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.457987 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.458885 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrz5s\" (UniqueName: \"kubernetes.io/projected/367566c7-8e32-444f-866a-d64b98a2f596-kube-api-access-vrz5s\") pod \"cluster-samples-operator-665b6dd947-5267z\" (UID: \"367566c7-8e32-444f-866a-d64b98a2f596\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5267z" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.490239 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr6t8\" (UniqueName: \"kubernetes.io/projected/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-kube-api-access-sr6t8\") pod \"route-controller-manager-6576b87f9c-gn5f4\" (UID: \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.496154 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.504160 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws2bl\" (UniqueName: \"kubernetes.io/projected/88f6bb4e-362c-425c-9f67-81f76ed05e42-kube-api-access-ws2bl\") pod \"downloads-7954f5f757-rf6tj\" (UID: \"88f6bb4e-362c-425c-9f67-81f76ed05e42\") " pod="openshift-console/downloads-7954f5f757-rf6tj" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.511888 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.518759 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s666l\" (UniqueName: \"kubernetes.io/projected/1394009a-f690-4e2f-8086-87b9483fabbd-kube-api-access-s666l\") pod \"olm-operator-6b444d44fb-2zf5q\" (UID: \"1394009a-f690-4e2f-8086-87b9483fabbd\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.528582 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.540132 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n6q5s" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.542150 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ft22\" (UniqueName: \"kubernetes.io/projected/6855b17f-5a68-45c7-a1b4-0c9cacc6ed14-kube-api-access-5ft22\") pod \"ingress-operator-5b745b69d9-l56lb\" (UID: \"6855b17f-5a68-45c7-a1b4-0c9cacc6ed14\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.546292 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9r9fc" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.560876 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.567102 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrkzf\" (UniqueName: \"kubernetes.io/projected/0fa71e97-6be0-4300-bef2-b2dffbf30076-kube-api-access-lrkzf\") pod \"marketplace-operator-79b997595-cgzqr\" (UID: \"0fa71e97-6be0-4300-bef2-b2dffbf30076\") " pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.582900 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.583508 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fcdlt"] Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.600288 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e69cc34e-b76e-4a7e-a4b5-c7e8436167e7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wvgpw\" (UID: \"e69cc34e-b76e-4a7e-a4b5-c7e8436167e7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wvgpw" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.600342 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5a67f6b6-1881-4807-b719-40ca184bd5dc-bound-sa-token\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.600379 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f65c3828-a372-4660-97dd-e01d6efcb78c-console-oauth-config\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.600428 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv5zp\" (UniqueName: \"kubernetes.io/projected/5a67f6b6-1881-4807-b719-40ca184bd5dc-kube-api-access-bv5zp\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.600463 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5a67f6b6-1881-4807-b719-40ca184bd5dc-trusted-ca\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.600518 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxnbz\" (UniqueName: \"kubernetes.io/projected/f65c3828-a372-4660-97dd-e01d6efcb78c-kube-api-access-xxnbz\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.600558 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrbv7\" (UniqueName: \"kubernetes.io/projected/e69cc34e-b76e-4a7e-a4b5-c7e8436167e7-kube-api-access-lrbv7\") pod \"openshift-controller-manager-operator-756b6f6bc6-wvgpw\" (UID: \"e69cc34e-b76e-4a7e-a4b5-c7e8436167e7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wvgpw" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.600582 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-oauth-serving-cert\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.600618 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.600648 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-console-config\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.600676 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-service-ca\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.600704 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5a67f6b6-1881-4807-b719-40ca184bd5dc-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.600728 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5a67f6b6-1881-4807-b719-40ca184bd5dc-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.601034 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f65c3828-a372-4660-97dd-e01d6efcb78c-console-serving-cert\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.601074 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5a67f6b6-1881-4807-b719-40ca184bd5dc-registry-certificates\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.601136 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5a67f6b6-1881-4807-b719-40ca184bd5dc-registry-tls\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.601154 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-trusted-ca-bundle\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.601199 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e69cc34e-b76e-4a7e-a4b5-c7e8436167e7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wvgpw\" (UID: \"e69cc34e-b76e-4a7e-a4b5-c7e8436167e7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wvgpw" Dec 01 14:36:22 crc kubenswrapper[4810]: E1201 14:36:22.601249 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:23.101235699 +0000 UTC m=+148.864745302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.623534 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.652185 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6qzb"] Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.664565 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-rf6tj" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.692154 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5267z" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.706547 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.706812 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/17423ab4-d417-42d3-84f0-ca131f3dceeb-plugins-dir\") pod \"csi-hostpathplugin-pdv9r\" (UID: \"17423ab4-d417-42d3-84f0-ca131f3dceeb\") " pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.706874 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6705a032-c5d3-45be-acd4-94097a6c9ca6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-n67g2\" (UID: \"6705a032-c5d3-45be-acd4-94097a6c9ca6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n67g2" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.706904 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3863c075-f600-4d7b-b452-1ffa4b5c80f0-service-ca-bundle\") pod \"router-default-5444994796-hsnbk\" (UID: \"3863c075-f600-4d7b-b452-1ffa4b5c80f0\") " pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.706974 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4j7s\" (UniqueName: \"kubernetes.io/projected/985c8210-eb5f-41da-8b74-95d7a6e7e9c7-kube-api-access-d4j7s\") pod \"package-server-manager-789f6589d5-zkr9c\" (UID: \"985c8210-eb5f-41da-8b74-95d7a6e7e9c7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zkr9c" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707005 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxnbz\" (UniqueName: \"kubernetes.io/projected/f65c3828-a372-4660-97dd-e01d6efcb78c-kube-api-access-xxnbz\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707052 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6dbca0cc-145c-41df-823a-eec2e7359ace-webhook-cert\") pod \"packageserver-d55dfcdfc-qh9j7\" (UID: \"6dbca0cc-145c-41df-823a-eec2e7359ace\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707077 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xm5j\" (UniqueName: \"kubernetes.io/projected/6dbca0cc-145c-41df-823a-eec2e7359ace-kube-api-access-9xm5j\") pod \"packageserver-d55dfcdfc-qh9j7\" (UID: \"6dbca0cc-145c-41df-823a-eec2e7359ace\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707129 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/985c8210-eb5f-41da-8b74-95d7a6e7e9c7-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-zkr9c\" (UID: \"985c8210-eb5f-41da-8b74-95d7a6e7e9c7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zkr9c" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707154 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4dea83f8-89e3-4a33-a4f0-fbb0619ab1df-config\") pod \"service-ca-operator-777779d784-xtf78\" (UID: \"4dea83f8-89e3-4a33-a4f0-fbb0619ab1df\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xtf78" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707248 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d08fa7f1-84f4-4909-ada5-f264d2215ba9-node-bootstrap-token\") pod \"machine-config-server-5rxzl\" (UID: \"d08fa7f1-84f4-4909-ada5-f264d2215ba9\") " pod="openshift-machine-config-operator/machine-config-server-5rxzl" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707274 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/17423ab4-d417-42d3-84f0-ca131f3dceeb-registration-dir\") pod \"csi-hostpathplugin-pdv9r\" (UID: \"17423ab4-d417-42d3-84f0-ca131f3dceeb\") " pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707311 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/17423ab4-d417-42d3-84f0-ca131f3dceeb-socket-dir\") pod \"csi-hostpathplugin-pdv9r\" (UID: \"17423ab4-d417-42d3-84f0-ca131f3dceeb\") " pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707343 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/266c7a73-0bce-4fad-b8c3-494b90a7d9b8-serving-cert\") pod \"etcd-operator-b45778765-t7pxb\" (UID: \"266c7a73-0bce-4fad-b8c3-494b90a7d9b8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707368 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6dbca0cc-145c-41df-823a-eec2e7359ace-tmpfs\") pod \"packageserver-d55dfcdfc-qh9j7\" (UID: \"6dbca0cc-145c-41df-823a-eec2e7359ace\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707408 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52pls\" (UniqueName: \"kubernetes.io/projected/62d665ff-bba5-4ed3-b0f2-471be90260b7-kube-api-access-52pls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jnrsh\" (UID: \"62d665ff-bba5-4ed3-b0f2-471be90260b7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jnrsh" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707434 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/484c6c08-536f-4374-971e-1fb4ef397b64-secret-volume\") pod \"collect-profiles-29409990-xjzld\" (UID: \"484c6c08-536f-4374-971e-1fb4ef397b64\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707522 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/17423ab4-d417-42d3-84f0-ca131f3dceeb-mountpoint-dir\") pod \"csi-hostpathplugin-pdv9r\" (UID: \"17423ab4-d417-42d3-84f0-ca131f3dceeb\") " pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707549 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40767ced-66c6-41a6-8e24-46e1bfb30c48-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5hpvd\" (UID: \"40767ced-66c6-41a6-8e24-46e1bfb30c48\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5hpvd" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707590 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5a67f6b6-1881-4807-b719-40ca184bd5dc-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707614 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2qvw\" (UniqueName: \"kubernetes.io/projected/484c6c08-536f-4374-971e-1fb4ef397b64-kube-api-access-m2qvw\") pod \"collect-profiles-29409990-xjzld\" (UID: \"484c6c08-536f-4374-971e-1fb4ef397b64\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707652 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phh78\" (UniqueName: \"kubernetes.io/projected/40767ced-66c6-41a6-8e24-46e1bfb30c48-kube-api-access-phh78\") pod \"multus-admission-controller-857f4d67dd-5hpvd\" (UID: \"40767ced-66c6-41a6-8e24-46e1bfb30c48\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5hpvd" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707717 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/246199f6-4473-41ee-ba76-804ee04e8aa6-available-featuregates\") pod \"openshift-config-operator-7777fb866f-h54rq\" (UID: \"246199f6-4473-41ee-ba76-804ee04e8aa6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h54rq" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707741 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4g8r\" (UniqueName: \"kubernetes.io/projected/cd5edc05-8e37-4d29-a7e5-34fa7e39bfa4-kube-api-access-c4g8r\") pod \"migrator-59844c95c7-s5rbr\" (UID: \"cd5edc05-8e37-4d29-a7e5-34fa7e39bfa4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s5rbr" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707841 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbwc5\" (UniqueName: \"kubernetes.io/projected/246199f6-4473-41ee-ba76-804ee04e8aa6-kube-api-access-sbwc5\") pod \"openshift-config-operator-7777fb866f-h54rq\" (UID: \"246199f6-4473-41ee-ba76-804ee04e8aa6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h54rq" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707889 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfwbz\" (UniqueName: \"kubernetes.io/projected/17423ab4-d417-42d3-84f0-ca131f3dceeb-kube-api-access-qfwbz\") pod \"csi-hostpathplugin-pdv9r\" (UID: \"17423ab4-d417-42d3-84f0-ca131f3dceeb\") " pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707922 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5a67f6b6-1881-4807-b719-40ca184bd5dc-registry-tls\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707948 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-trusted-ca-bundle\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.707969 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5b8p\" (UniqueName: \"kubernetes.io/projected/cf666da3-0465-4de2-9cea-7355839709f7-kube-api-access-n5b8p\") pod \"catalog-operator-68c6474976-9x2jh\" (UID: \"cf666da3-0465-4de2-9cea-7355839709f7\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.708027 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gg82\" (UniqueName: \"kubernetes.io/projected/d08fa7f1-84f4-4909-ada5-f264d2215ba9-kube-api-access-6gg82\") pod \"machine-config-server-5rxzl\" (UID: \"d08fa7f1-84f4-4909-ada5-f264d2215ba9\") " pod="openshift-machine-config-operator/machine-config-server-5rxzl" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.708079 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e69cc34e-b76e-4a7e-a4b5-c7e8436167e7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wvgpw\" (UID: \"e69cc34e-b76e-4a7e-a4b5-c7e8436167e7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wvgpw" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.708151 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cf666da3-0465-4de2-9cea-7355839709f7-srv-cert\") pod \"catalog-operator-68c6474976-9x2jh\" (UID: \"cf666da3-0465-4de2-9cea-7355839709f7\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.708170 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/484c6c08-536f-4374-971e-1fb4ef397b64-config-volume\") pod \"collect-profiles-29409990-xjzld\" (UID: \"484c6c08-536f-4374-971e-1fb4ef397b64\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.708206 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f65c3828-a372-4660-97dd-e01d6efcb78c-console-oauth-config\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.708619 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/17423ab4-d417-42d3-84f0-ca131f3dceeb-csi-data-dir\") pod \"csi-hostpathplugin-pdv9r\" (UID: \"17423ab4-d417-42d3-84f0-ca131f3dceeb\") " pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.708687 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv5zp\" (UniqueName: \"kubernetes.io/projected/5a67f6b6-1881-4807-b719-40ca184bd5dc-kube-api-access-bv5zp\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.708709 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4dea83f8-89e3-4a33-a4f0-fbb0619ab1df-serving-cert\") pod \"service-ca-operator-777779d784-xtf78\" (UID: \"4dea83f8-89e3-4a33-a4f0-fbb0619ab1df\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xtf78" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.708762 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6dbca0cc-145c-41df-823a-eec2e7359ace-apiservice-cert\") pod \"packageserver-d55dfcdfc-qh9j7\" (UID: \"6dbca0cc-145c-41df-823a-eec2e7359ace\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.708800 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3863c075-f600-4d7b-b452-1ffa4b5c80f0-stats-auth\") pod \"router-default-5444994796-hsnbk\" (UID: \"3863c075-f600-4d7b-b452-1ffa4b5c80f0\") " pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.708952 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/afa8aca0-8b8b-49de-ac08-746a5936fb3a-metrics-tls\") pod \"dns-operator-744455d44c-tlqrt\" (UID: \"afa8aca0-8b8b-49de-ac08-746a5936fb3a\") " pod="openshift-dns-operator/dns-operator-744455d44c-tlqrt" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709055 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5a67f6b6-1881-4807-b719-40ca184bd5dc-trusted-ca\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709103 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h4sc\" (UniqueName: \"kubernetes.io/projected/e62e2a4f-5d0d-4ee2-af70-56dc93b8ea95-kube-api-access-8h4sc\") pod \"ingress-canary-88hcl\" (UID: \"e62e2a4f-5d0d-4ee2-af70-56dc93b8ea95\") " pod="openshift-ingress-canary/ingress-canary-88hcl" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709201 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/266c7a73-0bce-4fad-b8c3-494b90a7d9b8-config\") pod \"etcd-operator-b45778765-t7pxb\" (UID: \"266c7a73-0bce-4fad-b8c3-494b90a7d9b8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709284 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3863c075-f600-4d7b-b452-1ffa4b5c80f0-default-certificate\") pod \"router-default-5444994796-hsnbk\" (UID: \"3863c075-f600-4d7b-b452-1ffa4b5c80f0\") " pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709338 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwslp\" (UniqueName: \"kubernetes.io/projected/afa8aca0-8b8b-49de-ac08-746a5936fb3a-kube-api-access-vwslp\") pod \"dns-operator-744455d44c-tlqrt\" (UID: \"afa8aca0-8b8b-49de-ac08-746a5936fb3a\") " pod="openshift-dns-operator/dns-operator-744455d44c-tlqrt" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709410 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cf666da3-0465-4de2-9cea-7355839709f7-profile-collector-cert\") pod \"catalog-operator-68c6474976-9x2jh\" (UID: \"cf666da3-0465-4de2-9cea-7355839709f7\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709442 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/62d665ff-bba5-4ed3-b0f2-471be90260b7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jnrsh\" (UID: \"62d665ff-bba5-4ed3-b0f2-471be90260b7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jnrsh" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709553 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrbv7\" (UniqueName: \"kubernetes.io/projected/e69cc34e-b76e-4a7e-a4b5-c7e8436167e7-kube-api-access-lrbv7\") pod \"openshift-controller-manager-operator-756b6f6bc6-wvgpw\" (UID: \"e69cc34e-b76e-4a7e-a4b5-c7e8436167e7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wvgpw" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709591 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-oauth-serving-cert\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709626 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhm4w\" (UniqueName: \"kubernetes.io/projected/266c7a73-0bce-4fad-b8c3-494b90a7d9b8-kube-api-access-lhm4w\") pod \"etcd-operator-b45778765-t7pxb\" (UID: \"266c7a73-0bce-4fad-b8c3-494b90a7d9b8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709664 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a3c66172-24da-4de2-b16d-f09a56ece4d5-metrics-tls\") pod \"dns-default-dt9dw\" (UID: \"a3c66172-24da-4de2-b16d-f09a56ece4d5\") " pod="openshift-dns/dns-default-dt9dw" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709693 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/266c7a73-0bce-4fad-b8c3-494b90a7d9b8-etcd-client\") pod \"etcd-operator-b45778765-t7pxb\" (UID: \"266c7a73-0bce-4fad-b8c3-494b90a7d9b8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709754 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6705a032-c5d3-45be-acd4-94097a6c9ca6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-n67g2\" (UID: \"6705a032-c5d3-45be-acd4-94097a6c9ca6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n67g2" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709782 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zrls\" (UniqueName: \"kubernetes.io/projected/6705a032-c5d3-45be-acd4-94097a6c9ca6-kube-api-access-9zrls\") pod \"kube-storage-version-migrator-operator-b67b599dd-n67g2\" (UID: \"6705a032-c5d3-45be-acd4-94097a6c9ca6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n67g2" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709808 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/246199f6-4473-41ee-ba76-804ee04e8aa6-serving-cert\") pod \"openshift-config-operator-7777fb866f-h54rq\" (UID: \"246199f6-4473-41ee-ba76-804ee04e8aa6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h54rq" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709833 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d08fa7f1-84f4-4909-ada5-f264d2215ba9-certs\") pod \"machine-config-server-5rxzl\" (UID: \"d08fa7f1-84f4-4909-ada5-f264d2215ba9\") " pod="openshift-machine-config-operator/machine-config-server-5rxzl" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709880 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3c66172-24da-4de2-b16d-f09a56ece4d5-config-volume\") pod \"dns-default-dt9dw\" (UID: \"a3c66172-24da-4de2-b16d-f09a56ece4d5\") " pod="openshift-dns/dns-default-dt9dw" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709906 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-console-config\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.709985 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-service-ca\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.710021 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5a67f6b6-1881-4807-b719-40ca184bd5dc-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.710148 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f65c3828-a372-4660-97dd-e01d6efcb78c-console-serving-cert\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.710181 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5a67f6b6-1881-4807-b719-40ca184bd5dc-registry-certificates\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.710205 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c95ce2ca-f352-422f-ba4d-6de8dbdc14b7-signing-key\") pod \"service-ca-9c57cc56f-hqhng\" (UID: \"c95ce2ca-f352-422f-ba4d-6de8dbdc14b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-hqhng" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.710251 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3863c075-f600-4d7b-b452-1ffa4b5c80f0-metrics-certs\") pod \"router-default-5444994796-hsnbk\" (UID: \"3863c075-f600-4d7b-b452-1ffa4b5c80f0\") " pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.710345 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e69cc34e-b76e-4a7e-a4b5-c7e8436167e7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wvgpw\" (UID: \"e69cc34e-b76e-4a7e-a4b5-c7e8436167e7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wvgpw" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.710370 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/266c7a73-0bce-4fad-b8c3-494b90a7d9b8-etcd-service-ca\") pod \"etcd-operator-b45778765-t7pxb\" (UID: \"266c7a73-0bce-4fad-b8c3-494b90a7d9b8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.710412 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/266c7a73-0bce-4fad-b8c3-494b90a7d9b8-etcd-ca\") pod \"etcd-operator-b45778765-t7pxb\" (UID: \"266c7a73-0bce-4fad-b8c3-494b90a7d9b8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.710530 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5a67f6b6-1881-4807-b719-40ca184bd5dc-bound-sa-token\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.710557 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7sl2\" (UniqueName: \"kubernetes.io/projected/a3c66172-24da-4de2-b16d-f09a56ece4d5-kube-api-access-k7sl2\") pod \"dns-default-dt9dw\" (UID: \"a3c66172-24da-4de2-b16d-f09a56ece4d5\") " pod="openshift-dns/dns-default-dt9dw" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.710580 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ndhl\" (UniqueName: \"kubernetes.io/projected/4dea83f8-89e3-4a33-a4f0-fbb0619ab1df-kube-api-access-7ndhl\") pod \"service-ca-operator-777779d784-xtf78\" (UID: \"4dea83f8-89e3-4a33-a4f0-fbb0619ab1df\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xtf78" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.710604 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l45n4\" (UniqueName: \"kubernetes.io/projected/3863c075-f600-4d7b-b452-1ffa4b5c80f0-kube-api-access-l45n4\") pod \"router-default-5444994796-hsnbk\" (UID: \"3863c075-f600-4d7b-b452-1ffa4b5c80f0\") " pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.710629 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c95ce2ca-f352-422f-ba4d-6de8dbdc14b7-signing-cabundle\") pod \"service-ca-9c57cc56f-hqhng\" (UID: \"c95ce2ca-f352-422f-ba4d-6de8dbdc14b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-hqhng" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.710651 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh55f\" (UniqueName: \"kubernetes.io/projected/c95ce2ca-f352-422f-ba4d-6de8dbdc14b7-kube-api-access-kh55f\") pod \"service-ca-9c57cc56f-hqhng\" (UID: \"c95ce2ca-f352-422f-ba4d-6de8dbdc14b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-hqhng" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.710673 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e62e2a4f-5d0d-4ee2-af70-56dc93b8ea95-cert\") pod \"ingress-canary-88hcl\" (UID: \"e62e2a4f-5d0d-4ee2-af70-56dc93b8ea95\") " pod="openshift-ingress-canary/ingress-canary-88hcl" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.715008 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-oauth-serving-cert\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.715987 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-trusted-ca-bundle\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.716920 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-console-config\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.718360 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5a67f6b6-1881-4807-b719-40ca184bd5dc-trusted-ca\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.720953 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5a67f6b6-1881-4807-b719-40ca184bd5dc-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.721338 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-service-ca\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.723752 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5a67f6b6-1881-4807-b719-40ca184bd5dc-registry-certificates\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.725240 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.727076 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-r2vhz"] Dec 01 14:36:22 crc kubenswrapper[4810]: E1201 14:36:22.728108 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:23.228034575 +0000 UTC m=+148.991544328 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.728785 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-pjt5m"] Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.730339 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xxkbc"] Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.732155 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5a67f6b6-1881-4807-b719-40ca184bd5dc-registry-tls\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.733274 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e69cc34e-b76e-4a7e-a4b5-c7e8436167e7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wvgpw\" (UID: \"e69cc34e-b76e-4a7e-a4b5-c7e8436167e7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wvgpw" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.733775 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5a67f6b6-1881-4807-b719-40ca184bd5dc-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.734552 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e69cc34e-b76e-4a7e-a4b5-c7e8436167e7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wvgpw\" (UID: \"e69cc34e-b76e-4a7e-a4b5-c7e8436167e7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wvgpw" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.735640 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f65c3828-a372-4660-97dd-e01d6efcb78c-console-serving-cert\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.737494 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f65c3828-a372-4660-97dd-e01d6efcb78c-console-oauth-config\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.742910 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxnbz\" (UniqueName: \"kubernetes.io/projected/f65c3828-a372-4660-97dd-e01d6efcb78c-kube-api-access-xxnbz\") pod \"console-f9d7485db-8tcbx\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: W1201 14:36:22.762381 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e6111a1_b6ac_4a66_b8f6_efef4463b6df.slice/crio-1b412e2a445bb46b95a80420982c07808016be1ecf942a714ce451abbfe4363a WatchSource:0}: Error finding container 1b412e2a445bb46b95a80420982c07808016be1ecf942a714ce451abbfe4363a: Status 404 returned error can't find the container with id 1b412e2a445bb46b95a80420982c07808016be1ecf942a714ce451abbfe4363a Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.762918 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrbv7\" (UniqueName: \"kubernetes.io/projected/e69cc34e-b76e-4a7e-a4b5-c7e8436167e7-kube-api-access-lrbv7\") pod \"openshift-controller-manager-operator-756b6f6bc6-wvgpw\" (UID: \"e69cc34e-b76e-4a7e-a4b5-c7e8436167e7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wvgpw" Dec 01 14:36:22 crc kubenswrapper[4810]: W1201 14:36:22.776335 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e7e3fa5_5c91_4694_95f4_67a503c3f42c.slice/crio-a11efdbeb10a147281a7e4b6316f4c994d39ad308da0998978dda69a95e6887b WatchSource:0}: Error finding container a11efdbeb10a147281a7e4b6316f4c994d39ad308da0998978dda69a95e6887b: Status 404 returned error can't find the container with id a11efdbeb10a147281a7e4b6316f4c994d39ad308da0998978dda69a95e6887b Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.787068 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv5zp\" (UniqueName: \"kubernetes.io/projected/5a67f6b6-1881-4807-b719-40ca184bd5dc-kube-api-access-bv5zp\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.801116 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5a67f6b6-1881-4807-b719-40ca184bd5dc-bound-sa-token\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.812848 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cf666da3-0465-4de2-9cea-7355839709f7-srv-cert\") pod \"catalog-operator-68c6474976-9x2jh\" (UID: \"cf666da3-0465-4de2-9cea-7355839709f7\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.812898 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/484c6c08-536f-4374-971e-1fb4ef397b64-config-volume\") pod \"collect-profiles-29409990-xjzld\" (UID: \"484c6c08-536f-4374-971e-1fb4ef397b64\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.812948 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/17423ab4-d417-42d3-84f0-ca131f3dceeb-csi-data-dir\") pod \"csi-hostpathplugin-pdv9r\" (UID: \"17423ab4-d417-42d3-84f0-ca131f3dceeb\") " pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.812971 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4dea83f8-89e3-4a33-a4f0-fbb0619ab1df-serving-cert\") pod \"service-ca-operator-777779d784-xtf78\" (UID: \"4dea83f8-89e3-4a33-a4f0-fbb0619ab1df\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xtf78" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.812992 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6dbca0cc-145c-41df-823a-eec2e7359ace-apiservice-cert\") pod \"packageserver-d55dfcdfc-qh9j7\" (UID: \"6dbca0cc-145c-41df-823a-eec2e7359ace\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813012 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3863c075-f600-4d7b-b452-1ffa4b5c80f0-stats-auth\") pod \"router-default-5444994796-hsnbk\" (UID: \"3863c075-f600-4d7b-b452-1ffa4b5c80f0\") " pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813032 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/afa8aca0-8b8b-49de-ac08-746a5936fb3a-metrics-tls\") pod \"dns-operator-744455d44c-tlqrt\" (UID: \"afa8aca0-8b8b-49de-ac08-746a5936fb3a\") " pod="openshift-dns-operator/dns-operator-744455d44c-tlqrt" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813056 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h4sc\" (UniqueName: \"kubernetes.io/projected/e62e2a4f-5d0d-4ee2-af70-56dc93b8ea95-kube-api-access-8h4sc\") pod \"ingress-canary-88hcl\" (UID: \"e62e2a4f-5d0d-4ee2-af70-56dc93b8ea95\") " pod="openshift-ingress-canary/ingress-canary-88hcl" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813078 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/266c7a73-0bce-4fad-b8c3-494b90a7d9b8-config\") pod \"etcd-operator-b45778765-t7pxb\" (UID: \"266c7a73-0bce-4fad-b8c3-494b90a7d9b8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813097 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3863c075-f600-4d7b-b452-1ffa4b5c80f0-default-certificate\") pod \"router-default-5444994796-hsnbk\" (UID: \"3863c075-f600-4d7b-b452-1ffa4b5c80f0\") " pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813122 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwslp\" (UniqueName: \"kubernetes.io/projected/afa8aca0-8b8b-49de-ac08-746a5936fb3a-kube-api-access-vwslp\") pod \"dns-operator-744455d44c-tlqrt\" (UID: \"afa8aca0-8b8b-49de-ac08-746a5936fb3a\") " pod="openshift-dns-operator/dns-operator-744455d44c-tlqrt" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813146 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cf666da3-0465-4de2-9cea-7355839709f7-profile-collector-cert\") pod \"catalog-operator-68c6474976-9x2jh\" (UID: \"cf666da3-0465-4de2-9cea-7355839709f7\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813177 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/62d665ff-bba5-4ed3-b0f2-471be90260b7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jnrsh\" (UID: \"62d665ff-bba5-4ed3-b0f2-471be90260b7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jnrsh" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813205 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhm4w\" (UniqueName: \"kubernetes.io/projected/266c7a73-0bce-4fad-b8c3-494b90a7d9b8-kube-api-access-lhm4w\") pod \"etcd-operator-b45778765-t7pxb\" (UID: \"266c7a73-0bce-4fad-b8c3-494b90a7d9b8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813229 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/266c7a73-0bce-4fad-b8c3-494b90a7d9b8-etcd-client\") pod \"etcd-operator-b45778765-t7pxb\" (UID: \"266c7a73-0bce-4fad-b8c3-494b90a7d9b8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813253 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a3c66172-24da-4de2-b16d-f09a56ece4d5-metrics-tls\") pod \"dns-default-dt9dw\" (UID: \"a3c66172-24da-4de2-b16d-f09a56ece4d5\") " pod="openshift-dns/dns-default-dt9dw" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813277 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6705a032-c5d3-45be-acd4-94097a6c9ca6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-n67g2\" (UID: \"6705a032-c5d3-45be-acd4-94097a6c9ca6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n67g2" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813299 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zrls\" (UniqueName: \"kubernetes.io/projected/6705a032-c5d3-45be-acd4-94097a6c9ca6-kube-api-access-9zrls\") pod \"kube-storage-version-migrator-operator-b67b599dd-n67g2\" (UID: \"6705a032-c5d3-45be-acd4-94097a6c9ca6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n67g2" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813325 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/246199f6-4473-41ee-ba76-804ee04e8aa6-serving-cert\") pod \"openshift-config-operator-7777fb866f-h54rq\" (UID: \"246199f6-4473-41ee-ba76-804ee04e8aa6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h54rq" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813351 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d08fa7f1-84f4-4909-ada5-f264d2215ba9-certs\") pod \"machine-config-server-5rxzl\" (UID: \"d08fa7f1-84f4-4909-ada5-f264d2215ba9\") " pod="openshift-machine-config-operator/machine-config-server-5rxzl" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813374 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3c66172-24da-4de2-b16d-f09a56ece4d5-config-volume\") pod \"dns-default-dt9dw\" (UID: \"a3c66172-24da-4de2-b16d-f09a56ece4d5\") " pod="openshift-dns/dns-default-dt9dw" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813407 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c95ce2ca-f352-422f-ba4d-6de8dbdc14b7-signing-key\") pod \"service-ca-9c57cc56f-hqhng\" (UID: \"c95ce2ca-f352-422f-ba4d-6de8dbdc14b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-hqhng" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813428 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3863c075-f600-4d7b-b452-1ffa4b5c80f0-metrics-certs\") pod \"router-default-5444994796-hsnbk\" (UID: \"3863c075-f600-4d7b-b452-1ffa4b5c80f0\") " pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813452 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/266c7a73-0bce-4fad-b8c3-494b90a7d9b8-etcd-service-ca\") pod \"etcd-operator-b45778765-t7pxb\" (UID: \"266c7a73-0bce-4fad-b8c3-494b90a7d9b8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813500 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/266c7a73-0bce-4fad-b8c3-494b90a7d9b8-etcd-ca\") pod \"etcd-operator-b45778765-t7pxb\" (UID: \"266c7a73-0bce-4fad-b8c3-494b90a7d9b8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813524 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7sl2\" (UniqueName: \"kubernetes.io/projected/a3c66172-24da-4de2-b16d-f09a56ece4d5-kube-api-access-k7sl2\") pod \"dns-default-dt9dw\" (UID: \"a3c66172-24da-4de2-b16d-f09a56ece4d5\") " pod="openshift-dns/dns-default-dt9dw" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813546 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ndhl\" (UniqueName: \"kubernetes.io/projected/4dea83f8-89e3-4a33-a4f0-fbb0619ab1df-kube-api-access-7ndhl\") pod \"service-ca-operator-777779d784-xtf78\" (UID: \"4dea83f8-89e3-4a33-a4f0-fbb0619ab1df\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xtf78" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813567 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l45n4\" (UniqueName: \"kubernetes.io/projected/3863c075-f600-4d7b-b452-1ffa4b5c80f0-kube-api-access-l45n4\") pod \"router-default-5444994796-hsnbk\" (UID: \"3863c075-f600-4d7b-b452-1ffa4b5c80f0\") " pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813610 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c95ce2ca-f352-422f-ba4d-6de8dbdc14b7-signing-cabundle\") pod \"service-ca-9c57cc56f-hqhng\" (UID: \"c95ce2ca-f352-422f-ba4d-6de8dbdc14b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-hqhng" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813643 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh55f\" (UniqueName: \"kubernetes.io/projected/c95ce2ca-f352-422f-ba4d-6de8dbdc14b7-kube-api-access-kh55f\") pod \"service-ca-9c57cc56f-hqhng\" (UID: \"c95ce2ca-f352-422f-ba4d-6de8dbdc14b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-hqhng" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813667 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e62e2a4f-5d0d-4ee2-af70-56dc93b8ea95-cert\") pod \"ingress-canary-88hcl\" (UID: \"e62e2a4f-5d0d-4ee2-af70-56dc93b8ea95\") " pod="openshift-ingress-canary/ingress-canary-88hcl" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813691 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/17423ab4-d417-42d3-84f0-ca131f3dceeb-plugins-dir\") pod \"csi-hostpathplugin-pdv9r\" (UID: \"17423ab4-d417-42d3-84f0-ca131f3dceeb\") " pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813721 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6705a032-c5d3-45be-acd4-94097a6c9ca6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-n67g2\" (UID: \"6705a032-c5d3-45be-acd4-94097a6c9ca6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n67g2" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813753 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3863c075-f600-4d7b-b452-1ffa4b5c80f0-service-ca-bundle\") pod \"router-default-5444994796-hsnbk\" (UID: \"3863c075-f600-4d7b-b452-1ffa4b5c80f0\") " pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813780 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4j7s\" (UniqueName: \"kubernetes.io/projected/985c8210-eb5f-41da-8b74-95d7a6e7e9c7-kube-api-access-d4j7s\") pod \"package-server-manager-789f6589d5-zkr9c\" (UID: \"985c8210-eb5f-41da-8b74-95d7a6e7e9c7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zkr9c" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813805 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6dbca0cc-145c-41df-823a-eec2e7359ace-webhook-cert\") pod \"packageserver-d55dfcdfc-qh9j7\" (UID: \"6dbca0cc-145c-41df-823a-eec2e7359ace\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813851 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xm5j\" (UniqueName: \"kubernetes.io/projected/6dbca0cc-145c-41df-823a-eec2e7359ace-kube-api-access-9xm5j\") pod \"packageserver-d55dfcdfc-qh9j7\" (UID: \"6dbca0cc-145c-41df-823a-eec2e7359ace\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813881 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/985c8210-eb5f-41da-8b74-95d7a6e7e9c7-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-zkr9c\" (UID: \"985c8210-eb5f-41da-8b74-95d7a6e7e9c7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zkr9c" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813907 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4dea83f8-89e3-4a33-a4f0-fbb0619ab1df-config\") pod \"service-ca-operator-777779d784-xtf78\" (UID: \"4dea83f8-89e3-4a33-a4f0-fbb0619ab1df\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xtf78" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813929 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d08fa7f1-84f4-4909-ada5-f264d2215ba9-node-bootstrap-token\") pod \"machine-config-server-5rxzl\" (UID: \"d08fa7f1-84f4-4909-ada5-f264d2215ba9\") " pod="openshift-machine-config-operator/machine-config-server-5rxzl" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813954 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/17423ab4-d417-42d3-84f0-ca131f3dceeb-registration-dir\") pod \"csi-hostpathplugin-pdv9r\" (UID: \"17423ab4-d417-42d3-84f0-ca131f3dceeb\") " pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.813979 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/17423ab4-d417-42d3-84f0-ca131f3dceeb-socket-dir\") pod \"csi-hostpathplugin-pdv9r\" (UID: \"17423ab4-d417-42d3-84f0-ca131f3dceeb\") " pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.814011 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6dbca0cc-145c-41df-823a-eec2e7359ace-tmpfs\") pod \"packageserver-d55dfcdfc-qh9j7\" (UID: \"6dbca0cc-145c-41df-823a-eec2e7359ace\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.814069 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.814094 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/266c7a73-0bce-4fad-b8c3-494b90a7d9b8-serving-cert\") pod \"etcd-operator-b45778765-t7pxb\" (UID: \"266c7a73-0bce-4fad-b8c3-494b90a7d9b8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.814119 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/484c6c08-536f-4374-971e-1fb4ef397b64-secret-volume\") pod \"collect-profiles-29409990-xjzld\" (UID: \"484c6c08-536f-4374-971e-1fb4ef397b64\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.814147 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52pls\" (UniqueName: \"kubernetes.io/projected/62d665ff-bba5-4ed3-b0f2-471be90260b7-kube-api-access-52pls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jnrsh\" (UID: \"62d665ff-bba5-4ed3-b0f2-471be90260b7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jnrsh" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.814172 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/17423ab4-d417-42d3-84f0-ca131f3dceeb-mountpoint-dir\") pod \"csi-hostpathplugin-pdv9r\" (UID: \"17423ab4-d417-42d3-84f0-ca131f3dceeb\") " pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.814200 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40767ced-66c6-41a6-8e24-46e1bfb30c48-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5hpvd\" (UID: \"40767ced-66c6-41a6-8e24-46e1bfb30c48\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5hpvd" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.814229 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2qvw\" (UniqueName: \"kubernetes.io/projected/484c6c08-536f-4374-971e-1fb4ef397b64-kube-api-access-m2qvw\") pod \"collect-profiles-29409990-xjzld\" (UID: \"484c6c08-536f-4374-971e-1fb4ef397b64\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.814258 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phh78\" (UniqueName: \"kubernetes.io/projected/40767ced-66c6-41a6-8e24-46e1bfb30c48-kube-api-access-phh78\") pod \"multus-admission-controller-857f4d67dd-5hpvd\" (UID: \"40767ced-66c6-41a6-8e24-46e1bfb30c48\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5hpvd" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.814286 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/246199f6-4473-41ee-ba76-804ee04e8aa6-available-featuregates\") pod \"openshift-config-operator-7777fb866f-h54rq\" (UID: \"246199f6-4473-41ee-ba76-804ee04e8aa6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h54rq" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.814313 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4g8r\" (UniqueName: \"kubernetes.io/projected/cd5edc05-8e37-4d29-a7e5-34fa7e39bfa4-kube-api-access-c4g8r\") pod \"migrator-59844c95c7-s5rbr\" (UID: \"cd5edc05-8e37-4d29-a7e5-34fa7e39bfa4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s5rbr" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.814342 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbwc5\" (UniqueName: \"kubernetes.io/projected/246199f6-4473-41ee-ba76-804ee04e8aa6-kube-api-access-sbwc5\") pod \"openshift-config-operator-7777fb866f-h54rq\" (UID: \"246199f6-4473-41ee-ba76-804ee04e8aa6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h54rq" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.814370 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfwbz\" (UniqueName: \"kubernetes.io/projected/17423ab4-d417-42d3-84f0-ca131f3dceeb-kube-api-access-qfwbz\") pod \"csi-hostpathplugin-pdv9r\" (UID: \"17423ab4-d417-42d3-84f0-ca131f3dceeb\") " pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.814396 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5b8p\" (UniqueName: \"kubernetes.io/projected/cf666da3-0465-4de2-9cea-7355839709f7-kube-api-access-n5b8p\") pod \"catalog-operator-68c6474976-9x2jh\" (UID: \"cf666da3-0465-4de2-9cea-7355839709f7\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.814421 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gg82\" (UniqueName: \"kubernetes.io/projected/d08fa7f1-84f4-4909-ada5-f264d2215ba9-kube-api-access-6gg82\") pod \"machine-config-server-5rxzl\" (UID: \"d08fa7f1-84f4-4909-ada5-f264d2215ba9\") " pod="openshift-machine-config-operator/machine-config-server-5rxzl" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.814465 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6705a032-c5d3-45be-acd4-94097a6c9ca6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-n67g2\" (UID: \"6705a032-c5d3-45be-acd4-94097a6c9ca6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n67g2" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.815666 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3c66172-24da-4de2-b16d-f09a56ece4d5-config-volume\") pod \"dns-default-dt9dw\" (UID: \"a3c66172-24da-4de2-b16d-f09a56ece4d5\") " pod="openshift-dns/dns-default-dt9dw" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.819909 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/17423ab4-d417-42d3-84f0-ca131f3dceeb-csi-data-dir\") pod \"csi-hostpathplugin-pdv9r\" (UID: \"17423ab4-d417-42d3-84f0-ca131f3dceeb\") " pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.820671 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/484c6c08-536f-4374-971e-1fb4ef397b64-config-volume\") pod \"collect-profiles-29409990-xjzld\" (UID: \"484c6c08-536f-4374-971e-1fb4ef397b64\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.821267 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cf666da3-0465-4de2-9cea-7355839709f7-srv-cert\") pod \"catalog-operator-68c6474976-9x2jh\" (UID: \"cf666da3-0465-4de2-9cea-7355839709f7\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.821460 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4dea83f8-89e3-4a33-a4f0-fbb0619ab1df-config\") pod \"service-ca-operator-777779d784-xtf78\" (UID: \"4dea83f8-89e3-4a33-a4f0-fbb0619ab1df\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xtf78" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.821948 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cf666da3-0465-4de2-9cea-7355839709f7-profile-collector-cert\") pod \"catalog-operator-68c6474976-9x2jh\" (UID: \"cf666da3-0465-4de2-9cea-7355839709f7\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.822221 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/17423ab4-d417-42d3-84f0-ca131f3dceeb-mountpoint-dir\") pod \"csi-hostpathplugin-pdv9r\" (UID: \"17423ab4-d417-42d3-84f0-ca131f3dceeb\") " pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.822245 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/17423ab4-d417-42d3-84f0-ca131f3dceeb-plugins-dir\") pod \"csi-hostpathplugin-pdv9r\" (UID: \"17423ab4-d417-42d3-84f0-ca131f3dceeb\") " pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.822619 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/246199f6-4473-41ee-ba76-804ee04e8aa6-available-featuregates\") pod \"openshift-config-operator-7777fb866f-h54rq\" (UID: \"246199f6-4473-41ee-ba76-804ee04e8aa6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h54rq" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.824073 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3863c075-f600-4d7b-b452-1ffa4b5c80f0-default-certificate\") pod \"router-default-5444994796-hsnbk\" (UID: \"3863c075-f600-4d7b-b452-1ffa4b5c80f0\") " pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.824246 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3863c075-f600-4d7b-b452-1ffa4b5c80f0-service-ca-bundle\") pod \"router-default-5444994796-hsnbk\" (UID: \"3863c075-f600-4d7b-b452-1ffa4b5c80f0\") " pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.824805 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d08fa7f1-84f4-4909-ada5-f264d2215ba9-certs\") pod \"machine-config-server-5rxzl\" (UID: \"d08fa7f1-84f4-4909-ada5-f264d2215ba9\") " pod="openshift-machine-config-operator/machine-config-server-5rxzl" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.825154 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a3c66172-24da-4de2-b16d-f09a56ece4d5-metrics-tls\") pod \"dns-default-dt9dw\" (UID: \"a3c66172-24da-4de2-b16d-f09a56ece4d5\") " pod="openshift-dns/dns-default-dt9dw" Dec 01 14:36:22 crc kubenswrapper[4810]: E1201 14:36:22.825172 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:23.325149233 +0000 UTC m=+149.088659016 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.825486 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/266c7a73-0bce-4fad-b8c3-494b90a7d9b8-etcd-client\") pod \"etcd-operator-b45778765-t7pxb\" (UID: \"266c7a73-0bce-4fad-b8c3-494b90a7d9b8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.825860 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/266c7a73-0bce-4fad-b8c3-494b90a7d9b8-etcd-service-ca\") pod \"etcd-operator-b45778765-t7pxb\" (UID: \"266c7a73-0bce-4fad-b8c3-494b90a7d9b8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.826508 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6dbca0cc-145c-41df-823a-eec2e7359ace-tmpfs\") pod \"packageserver-d55dfcdfc-qh9j7\" (UID: \"6dbca0cc-145c-41df-823a-eec2e7359ace\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.826590 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/17423ab4-d417-42d3-84f0-ca131f3dceeb-socket-dir\") pod \"csi-hostpathplugin-pdv9r\" (UID: \"17423ab4-d417-42d3-84f0-ca131f3dceeb\") " pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.833588 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c95ce2ca-f352-422f-ba4d-6de8dbdc14b7-signing-cabundle\") pod \"service-ca-9c57cc56f-hqhng\" (UID: \"c95ce2ca-f352-422f-ba4d-6de8dbdc14b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-hqhng" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.835553 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/17423ab4-d417-42d3-84f0-ca131f3dceeb-registration-dir\") pod \"csi-hostpathplugin-pdv9r\" (UID: \"17423ab4-d417-42d3-84f0-ca131f3dceeb\") " pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.841457 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.845528 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6705a032-c5d3-45be-acd4-94097a6c9ca6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-n67g2\" (UID: \"6705a032-c5d3-45be-acd4-94097a6c9ca6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n67g2" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.846107 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4dea83f8-89e3-4a33-a4f0-fbb0619ab1df-serving-cert\") pod \"service-ca-operator-777779d784-xtf78\" (UID: \"4dea83f8-89e3-4a33-a4f0-fbb0619ab1df\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xtf78" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.846433 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d08fa7f1-84f4-4909-ada5-f264d2215ba9-node-bootstrap-token\") pod \"machine-config-server-5rxzl\" (UID: \"d08fa7f1-84f4-4909-ada5-f264d2215ba9\") " pod="openshift-machine-config-operator/machine-config-server-5rxzl" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.846462 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e62e2a4f-5d0d-4ee2-af70-56dc93b8ea95-cert\") pod \"ingress-canary-88hcl\" (UID: \"e62e2a4f-5d0d-4ee2-af70-56dc93b8ea95\") " pod="openshift-ingress-canary/ingress-canary-88hcl" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.846707 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/266c7a73-0bce-4fad-b8c3-494b90a7d9b8-config\") pod \"etcd-operator-b45778765-t7pxb\" (UID: \"266c7a73-0bce-4fad-b8c3-494b90a7d9b8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.846930 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/afa8aca0-8b8b-49de-ac08-746a5936fb3a-metrics-tls\") pod \"dns-operator-744455d44c-tlqrt\" (UID: \"afa8aca0-8b8b-49de-ac08-746a5936fb3a\") " pod="openshift-dns-operator/dns-operator-744455d44c-tlqrt" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.847643 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c95ce2ca-f352-422f-ba4d-6de8dbdc14b7-signing-key\") pod \"service-ca-9c57cc56f-hqhng\" (UID: \"c95ce2ca-f352-422f-ba4d-6de8dbdc14b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-hqhng" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.848979 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/246199f6-4473-41ee-ba76-804ee04e8aa6-serving-cert\") pod \"openshift-config-operator-7777fb866f-h54rq\" (UID: \"246199f6-4473-41ee-ba76-804ee04e8aa6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h54rq" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.849074 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/40767ced-66c6-41a6-8e24-46e1bfb30c48-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5hpvd\" (UID: \"40767ced-66c6-41a6-8e24-46e1bfb30c48\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5hpvd" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.849150 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/484c6c08-536f-4374-971e-1fb4ef397b64-secret-volume\") pod \"collect-profiles-29409990-xjzld\" (UID: \"484c6c08-536f-4374-971e-1fb4ef397b64\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.849535 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3863c075-f600-4d7b-b452-1ffa4b5c80f0-metrics-certs\") pod \"router-default-5444994796-hsnbk\" (UID: \"3863c075-f600-4d7b-b452-1ffa4b5c80f0\") " pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.849575 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/985c8210-eb5f-41da-8b74-95d7a6e7e9c7-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-zkr9c\" (UID: \"985c8210-eb5f-41da-8b74-95d7a6e7e9c7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zkr9c" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.849793 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/266c7a73-0bce-4fad-b8c3-494b90a7d9b8-serving-cert\") pod \"etcd-operator-b45778765-t7pxb\" (UID: \"266c7a73-0bce-4fad-b8c3-494b90a7d9b8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.853172 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3863c075-f600-4d7b-b452-1ffa4b5c80f0-stats-auth\") pod \"router-default-5444994796-hsnbk\" (UID: \"3863c075-f600-4d7b-b452-1ffa4b5c80f0\") " pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.854092 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/62d665ff-bba5-4ed3-b0f2-471be90260b7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jnrsh\" (UID: \"62d665ff-bba5-4ed3-b0f2-471be90260b7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jnrsh" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.854978 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6dbca0cc-145c-41df-823a-eec2e7359ace-apiservice-cert\") pod \"packageserver-d55dfcdfc-qh9j7\" (UID: \"6dbca0cc-145c-41df-823a-eec2e7359ace\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.863409 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/266c7a73-0bce-4fad-b8c3-494b90a7d9b8-etcd-ca\") pod \"etcd-operator-b45778765-t7pxb\" (UID: \"266c7a73-0bce-4fad-b8c3-494b90a7d9b8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.867495 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6dbca0cc-145c-41df-823a-eec2e7359ace-webhook-cert\") pod \"packageserver-d55dfcdfc-qh9j7\" (UID: \"6dbca0cc-145c-41df-823a-eec2e7359ace\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.869421 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.871820 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wvgpw" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.874569 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gg82\" (UniqueName: \"kubernetes.io/projected/d08fa7f1-84f4-4909-ada5-f264d2215ba9-kube-api-access-6gg82\") pod \"machine-config-server-5rxzl\" (UID: \"d08fa7f1-84f4-4909-ada5-f264d2215ba9\") " pod="openshift-machine-config-operator/machine-config-server-5rxzl" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.891650 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zrls\" (UniqueName: \"kubernetes.io/projected/6705a032-c5d3-45be-acd4-94097a6c9ca6-kube-api-access-9zrls\") pod \"kube-storage-version-migrator-operator-b67b599dd-n67g2\" (UID: \"6705a032-c5d3-45be-acd4-94097a6c9ca6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n67g2" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.902769 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhm4w\" (UniqueName: \"kubernetes.io/projected/266c7a73-0bce-4fad-b8c3-494b90a7d9b8-kube-api-access-lhm4w\") pod \"etcd-operator-b45778765-t7pxb\" (UID: \"266c7a73-0bce-4fad-b8c3-494b90a7d9b8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.909810 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.912639 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n67g2" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.915969 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:22 crc kubenswrapper[4810]: E1201 14:36:22.916379 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:23.416363851 +0000 UTC m=+149.179873454 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.937733 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwslp\" (UniqueName: \"kubernetes.io/projected/afa8aca0-8b8b-49de-ac08-746a5936fb3a-kube-api-access-vwslp\") pod \"dns-operator-744455d44c-tlqrt\" (UID: \"afa8aca0-8b8b-49de-ac08-746a5936fb3a\") " pod="openshift-dns-operator/dns-operator-744455d44c-tlqrt" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.948122 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h4sc\" (UniqueName: \"kubernetes.io/projected/e62e2a4f-5d0d-4ee2-af70-56dc93b8ea95-kube-api-access-8h4sc\") pod \"ingress-canary-88hcl\" (UID: \"e62e2a4f-5d0d-4ee2-af70-56dc93b8ea95\") " pod="openshift-ingress-canary/ingress-canary-88hcl" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.965938 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh55f\" (UniqueName: \"kubernetes.io/projected/c95ce2ca-f352-422f-ba4d-6de8dbdc14b7-kube-api-access-kh55f\") pod \"service-ca-9c57cc56f-hqhng\" (UID: \"c95ce2ca-f352-422f-ba4d-6de8dbdc14b7\") " pod="openshift-service-ca/service-ca-9c57cc56f-hqhng" Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.966516 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2mxj4"] Dec 01 14:36:22 crc kubenswrapper[4810]: I1201 14:36:22.998411 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4"] Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.003775 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xm5j\" (UniqueName: \"kubernetes.io/projected/6dbca0cc-145c-41df-823a-eec2e7359ace-kube-api-access-9xm5j\") pod \"packageserver-d55dfcdfc-qh9j7\" (UID: \"6dbca0cc-145c-41df-823a-eec2e7359ace\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.008058 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4g8r\" (UniqueName: \"kubernetes.io/projected/cd5edc05-8e37-4d29-a7e5-34fa7e39bfa4-kube-api-access-c4g8r\") pod \"migrator-59844c95c7-s5rbr\" (UID: \"cd5edc05-8e37-4d29-a7e5-34fa7e39bfa4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s5rbr" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.016831 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-88hcl" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.017184 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:23 crc kubenswrapper[4810]: E1201 14:36:23.017563 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:23.517548027 +0000 UTC m=+149.281057630 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.024191 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-5rxzl" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.043760 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbwc5\" (UniqueName: \"kubernetes.io/projected/246199f6-4473-41ee-ba76-804ee04e8aa6-kube-api-access-sbwc5\") pod \"openshift-config-operator-7777fb866f-h54rq\" (UID: \"246199f6-4473-41ee-ba76-804ee04e8aa6\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h54rq" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.044531 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfwbz\" (UniqueName: \"kubernetes.io/projected/17423ab4-d417-42d3-84f0-ca131f3dceeb-kube-api-access-qfwbz\") pod \"csi-hostpathplugin-pdv9r\" (UID: \"17423ab4-d417-42d3-84f0-ca131f3dceeb\") " pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:23 crc kubenswrapper[4810]: W1201 14:36:23.050932 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9c93c51_49ca_4557_a26b_0ba162d4dbc2.slice/crio-373ee90aacb1d2f9d9c242b48cf4b84d90dedb67cea3df6ad832c47d82b302ea WatchSource:0}: Error finding container 373ee90aacb1d2f9d9c242b48cf4b84d90dedb67cea3df6ad832c47d82b302ea: Status 404 returned error can't find the container with id 373ee90aacb1d2f9d9c242b48cf4b84d90dedb67cea3df6ad832c47d82b302ea Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.092389 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n"] Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.099886 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2qvw\" (UniqueName: \"kubernetes.io/projected/484c6c08-536f-4374-971e-1fb4ef397b64-kube-api-access-m2qvw\") pod \"collect-profiles-29409990-xjzld\" (UID: \"484c6c08-536f-4374-971e-1fb4ef397b64\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.101820 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z"] Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.110441 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4j7s\" (UniqueName: \"kubernetes.io/projected/985c8210-eb5f-41da-8b74-95d7a6e7e9c7-kube-api-access-d4j7s\") pod \"package-server-manager-789f6589d5-zkr9c\" (UID: \"985c8210-eb5f-41da-8b74-95d7a6e7e9c7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zkr9c" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.110757 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5b8p\" (UniqueName: \"kubernetes.io/projected/cf666da3-0465-4de2-9cea-7355839709f7-kube-api-access-n5b8p\") pod \"catalog-operator-68c6474976-9x2jh\" (UID: \"cf666da3-0465-4de2-9cea-7355839709f7\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.119205 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:23 crc kubenswrapper[4810]: E1201 14:36:23.119704 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:23.619681219 +0000 UTC m=+149.383190822 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.120908 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fdwcc"] Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.123935 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7sl2\" (UniqueName: \"kubernetes.io/projected/a3c66172-24da-4de2-b16d-f09a56ece4d5-kube-api-access-k7sl2\") pod \"dns-default-dt9dw\" (UID: \"a3c66172-24da-4de2-b16d-f09a56ece4d5\") " pod="openshift-dns/dns-default-dt9dw" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.138586 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phh78\" (UniqueName: \"kubernetes.io/projected/40767ced-66c6-41a6-8e24-46e1bfb30c48-kube-api-access-phh78\") pod \"multus-admission-controller-857f4d67dd-5hpvd\" (UID: \"40767ced-66c6-41a6-8e24-46e1bfb30c48\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5hpvd" Dec 01 14:36:23 crc kubenswrapper[4810]: W1201 14:36:23.163204 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2540546_e6b2_421d_8680_327f9cdeca7d.slice/crio-6d7cac077d0e094e96d22249023329e7d1ed2b86490c31ac6c659e1a3b72a97b WatchSource:0}: Error finding container 6d7cac077d0e094e96d22249023329e7d1ed2b86490c31ac6c659e1a3b72a97b: Status 404 returned error can't find the container with id 6d7cac077d0e094e96d22249023329e7d1ed2b86490c31ac6c659e1a3b72a97b Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.164426 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52pls\" (UniqueName: \"kubernetes.io/projected/62d665ff-bba5-4ed3-b0f2-471be90260b7-kube-api-access-52pls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jnrsh\" (UID: \"62d665ff-bba5-4ed3-b0f2-471be90260b7\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jnrsh" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.188750 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h54rq" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.195280 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-tlqrt" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.197319 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ndhl\" (UniqueName: \"kubernetes.io/projected/4dea83f8-89e3-4a33-a4f0-fbb0619ab1df-kube-api-access-7ndhl\") pod \"service-ca-operator-777779d784-xtf78\" (UID: \"4dea83f8-89e3-4a33-a4f0-fbb0619ab1df\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xtf78" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.206082 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l45n4\" (UniqueName: \"kubernetes.io/projected/3863c075-f600-4d7b-b452-1ffa4b5c80f0-kube-api-access-l45n4\") pod \"router-default-5444994796-hsnbk\" (UID: \"3863c075-f600-4d7b-b452-1ffa4b5c80f0\") " pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.221294 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s5rbr" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.222380 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:23 crc kubenswrapper[4810]: E1201 14:36:23.222762 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:23.722747447 +0000 UTC m=+149.486257050 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:23 crc kubenswrapper[4810]: W1201 14:36:23.229466 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebc6f3ed_4a0e_44df_94b7_e9b9f9aafc19.slice/crio-d983f772914c3cc569a5d407b84fc2c9153c676ed75010b648d162d66916d367 WatchSource:0}: Error finding container d983f772914c3cc569a5d407b84fc2c9153c676ed75010b648d162d66916d367: Status 404 returned error can't find the container with id d983f772914c3cc569a5d407b84fc2c9153c676ed75010b648d162d66916d367 Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.233857 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-5hpvd" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.238826 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.241130 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-hqhng" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.253218 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zkr9c" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.263350 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.263779 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jnrsh" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.270800 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-r2vhz" event={"ID":"7e7e3fa5-5c91-4694-95f4-67a503c3f42c","Type":"ContainerStarted","Data":"a11efdbeb10a147281a7e4b6316f4c994d39ad308da0998978dda69a95e6887b"} Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.271173 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xtf78" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.278141 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.294258 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dt9dw" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.300710 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q"] Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.308911 4810 generic.go:334] "Generic (PLEG): container finished" podID="2781f96f-4018-4ccf-adb8-04512de46450" containerID="4750bb7ea59132fe50315fc5c8f4a8fbaf41d3ff997dbe32c7993728cc40be25" exitCode=0 Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.309000 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" event={"ID":"2781f96f-4018-4ccf-adb8-04512de46450","Type":"ContainerDied","Data":"4750bb7ea59132fe50315fc5c8f4a8fbaf41d3ff997dbe32c7993728cc40be25"} Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.309029 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" event={"ID":"2781f96f-4018-4ccf-adb8-04512de46450","Type":"ContainerStarted","Data":"7ad3d24b2e970bd6b932750ed52ca589234434345de2e2526836a95e283cd161"} Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.309316 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n6q5s"] Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.312064 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4rgpg"] Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.313760 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xxkbc" event={"ID":"93b94518-da01-439a-9f46-2329f86020a4","Type":"ContainerStarted","Data":"6753c5a1718acae36ea60afb22e072911d466808a4f92db1f67b13717a44709f"} Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.315138 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.317675 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6qzb" event={"ID":"4f010e33-46e5-4080-a24f-47850bee04cd","Type":"ContainerStarted","Data":"f274f6bb62f6b7d2aa668b574bfd69b9f21016ba325a66c531449d260ae0022c"} Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.319070 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" event={"ID":"8e00f933-d775-4766-ae08-0d300d09fa4a","Type":"ContainerStarted","Data":"ce74b282bf620d9137d9354c2d8888dbafa34fbfb6c71378711fab4d4505fb18"} Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.322762 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pjt5m" event={"ID":"3e6111a1-b6ac-4a66-b8f6-efef4463b6df","Type":"ContainerStarted","Data":"077ba947e755a20ba1abe14df74e878638f0e810debe3a1382b8d8c78c6deac1"} Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.322795 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pjt5m" event={"ID":"3e6111a1-b6ac-4a66-b8f6-efef4463b6df","Type":"ContainerStarted","Data":"1b412e2a445bb46b95a80420982c07808016be1ecf942a714ce451abbfe4363a"} Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.323198 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:23 crc kubenswrapper[4810]: E1201 14:36:23.323390 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:23.823357008 +0000 UTC m=+149.586866611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.323554 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:23 crc kubenswrapper[4810]: E1201 14:36:23.324001 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:23.823991115 +0000 UTC m=+149.587500898 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.324094 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9" event={"ID":"5537fef3-e146-4eac-bc97-3edc211a2b3d","Type":"ContainerStarted","Data":"54dc0d66ffdc43cbbb4da638097d45ce54ee599055cd944e6c65f41fbb34b2b2"} Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.336137 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" event={"ID":"e9c93c51-49ca-4557-a26b-0ba162d4dbc2","Type":"ContainerStarted","Data":"373ee90aacb1d2f9d9c242b48cf4b84d90dedb67cea3df6ad832c47d82b302ea"} Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.338184 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-2mxj4" event={"ID":"ddafebe9-9403-44c2-9d32-2867e6c212ce","Type":"ContainerStarted","Data":"cc3156000fbc85ba5562b30999a037c39c5f7dd8c84dbfb5fad2bbdbd0171810"} Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.341232 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n" event={"ID":"d2540546-e6b2-421d-8680-327f9cdeca7d","Type":"ContainerStarted","Data":"6d7cac077d0e094e96d22249023329e7d1ed2b86490c31ac6c659e1a3b72a97b"} Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.424160 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:23 crc kubenswrapper[4810]: E1201 14:36:23.425267 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:23.925250343 +0000 UTC m=+149.688759946 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:23 crc kubenswrapper[4810]: W1201 14:36:23.470605 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda543c4e2_99d9_4b47_a9c6_927c8737ed49.slice/crio-b03bc539700923dbe0ac48ef2ca0fdcb688a34e0b940a8ff14ef02a74ce30f30 WatchSource:0}: Error finding container b03bc539700923dbe0ac48ef2ca0fdcb688a34e0b940a8ff14ef02a74ce30f30: Status 404 returned error can't find the container with id b03bc539700923dbe0ac48ef2ca0fdcb688a34e0b940a8ff14ef02a74ce30f30 Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.480068 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.490449 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9r9fc"] Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.499022 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-rf6tj"] Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.517767 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn"] Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.526392 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:23 crc kubenswrapper[4810]: E1201 14:36:23.526783 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:24.026770779 +0000 UTC m=+149.790280382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.627242 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:23 crc kubenswrapper[4810]: E1201 14:36:23.627381 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:24.12736056 +0000 UTC m=+149.890870163 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.627719 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:23 crc kubenswrapper[4810]: E1201 14:36:23.628016 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:24.128007567 +0000 UTC m=+149.891517170 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:23 crc kubenswrapper[4810]: W1201 14:36:23.631688 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88f6bb4e_362c_425c_9f67_81f76ed05e42.slice/crio-316774c10ead9265412a37a9d916d4462ac571a93bf59c4453ef5828e6bc5548 WatchSource:0}: Error finding container 316774c10ead9265412a37a9d916d4462ac571a93bf59c4453ef5828e6bc5548: Status 404 returned error can't find the container with id 316774c10ead9265412a37a9d916d4462ac571a93bf59c4453ef5828e6bc5548 Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.720162 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5267z"] Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.730350 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:23 crc kubenswrapper[4810]: E1201 14:36:23.730676 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:24.230660223 +0000 UTC m=+149.994169826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.746680 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb"] Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.771668 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cgzqr"] Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.813843 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-8tcbx"] Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.828579 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wvgpw"] Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.832264 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:23 crc kubenswrapper[4810]: E1201 14:36:23.832589 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:24.33257693 +0000 UTC m=+150.096086533 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.833951 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-t7pxb"] Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.933002 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:23 crc kubenswrapper[4810]: E1201 14:36:23.933229 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:24.433187351 +0000 UTC m=+150.196696954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:23 crc kubenswrapper[4810]: I1201 14:36:23.933703 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:23 crc kubenswrapper[4810]: E1201 14:36:23.934046 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:24.434030564 +0000 UTC m=+150.197540167 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.035925 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:24 crc kubenswrapper[4810]: E1201 14:36:24.036549 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:24.536533866 +0000 UTC m=+150.300043469 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:24 crc kubenswrapper[4810]: W1201 14:36:24.058832 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod266c7a73_0bce_4fad_b8c3_494b90a7d9b8.slice/crio-4a2d91fd94dde8b7099bd7531259624356ef9d77ddde3ff23338bbb4aa5e31d0 WatchSource:0}: Error finding container 4a2d91fd94dde8b7099bd7531259624356ef9d77ddde3ff23338bbb4aa5e31d0: Status 404 returned error can't find the container with id 4a2d91fd94dde8b7099bd7531259624356ef9d77ddde3ff23338bbb4aa5e31d0 Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.137555 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:24 crc kubenswrapper[4810]: E1201 14:36:24.137876 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:24.637865006 +0000 UTC m=+150.401374609 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.161576 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld"] Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.241264 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-88hcl"] Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.246777 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:24 crc kubenswrapper[4810]: E1201 14:36:24.246997 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:24.746970447 +0000 UTC m=+150.510480050 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.247463 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:24 crc kubenswrapper[4810]: E1201 14:36:24.247775 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:24.747764518 +0000 UTC m=+150.511274121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.334321 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n67g2"] Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.337905 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-tlqrt"] Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.353782 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:24 crc kubenswrapper[4810]: E1201 14:36:24.354180 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:24.854165786 +0000 UTC m=+150.617675389 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.459636 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:24 crc kubenswrapper[4810]: E1201 14:36:24.460174 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:24.960163362 +0000 UTC m=+150.723672965 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.564756 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:24 crc kubenswrapper[4810]: E1201 14:36:24.565200 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:25.065185102 +0000 UTC m=+150.828694705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.609174 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jnrsh"] Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.609231 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-hqhng"] Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.609253 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-s5rbr"] Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.614509 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7"] Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.651803 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-r2vhz" event={"ID":"7e7e3fa5-5c91-4694-95f4-67a503c3f42c","Type":"ContainerStarted","Data":"f4500ea0808e00ca71ca57cbb033c4d264b3f5c86d29918e9a348889ad5eafdd"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.652234 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-r2vhz" event={"ID":"7e7e3fa5-5c91-4694-95f4-67a503c3f42c","Type":"ContainerStarted","Data":"34e04387ef4f340f92a0472f3a8c012a0728c6d4b4298a3b6c8562da69e212d4"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.665617 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:24 crc kubenswrapper[4810]: E1201 14:36:24.667179 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:25.16716567 +0000 UTC m=+150.930675263 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.677147 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9r9fc" event={"ID":"54284225-c472-482c-93b2-2426031f0957","Type":"ContainerStarted","Data":"e7e60812bc89835b445696b54bbc20a39313213c950d29f53768035fb1edb230"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.682283 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-h54rq"] Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.684725 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xtf78"] Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.715636 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" event={"ID":"e9c93c51-49ca-4557-a26b-0ba162d4dbc2","Type":"ContainerStarted","Data":"f7f9e80257653a1377d5ae9754366811b70c4ea7dc9f400f31b92a359a29a2ad"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.716574 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.729155 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-5rxzl" event={"ID":"d08fa7f1-84f4-4909-ada5-f264d2215ba9","Type":"ContainerStarted","Data":"6636ada5441c04a3aa0e33cbdf5677528629c0b05a0ded1c4c4eaba16f770f11"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.735218 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n6q5s" event={"ID":"a543c4e2-99d9-4b47-a9c6-927c8737ed49","Type":"ContainerStarted","Data":"b03bc539700923dbe0ac48ef2ca0fdcb688a34e0b940a8ff14ef02a74ce30f30"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.745417 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-2mxj4" event={"ID":"ddafebe9-9403-44c2-9d32-2867e6c212ce","Type":"ContainerStarted","Data":"4353c68850bd7c754aa47da04ae8155dbdef9f67a24561fc290f49f129a398f3"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.745819 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-2mxj4" Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.755618 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6qzb" event={"ID":"4f010e33-46e5-4080-a24f-47850bee04cd","Type":"ContainerStarted","Data":"54b12d886a7045e89da6095d2e08409e90ab70a63ae1e6e414041838433f6a31"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.761062 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.761766 4810 patch_prober.go:28] interesting pod/console-operator-58897d9998-2mxj4 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.761812 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-2mxj4" podUID="ddafebe9-9403-44c2-9d32-2867e6c212ce" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.762988 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5267z" event={"ID":"367566c7-8e32-444f-866a-d64b98a2f596","Type":"ContainerStarted","Data":"3b55fc67cc6886e7704f4fb590b68cbfa82f2691af94ee5d2df7df602a4855d3"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.765198 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-88hcl" event={"ID":"e62e2a4f-5d0d-4ee2-af70-56dc93b8ea95","Type":"ContainerStarted","Data":"68e1dc14eafafaf5aa3f8d557fa37756c32bd0984de218fb4fbaa21a9f22317f"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.766713 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:24 crc kubenswrapper[4810]: E1201 14:36:24.768031 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:25.268011587 +0000 UTC m=+151.031521200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.788085 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wvgpw" event={"ID":"e69cc34e-b76e-4a7e-a4b5-c7e8436167e7","Type":"ContainerStarted","Data":"d6bb6f0d564c052c54e3f5a3ea6d7aff5c47d06e06f1ec663a803f6137414ba6"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.808449 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5hpvd"] Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.809704 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dt9dw"] Dec 01 14:36:24 crc kubenswrapper[4810]: W1201 14:36:24.818592 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc95ce2ca_f352_422f_ba4d_6de8dbdc14b7.slice/crio-0edd1df7fac7341e52cf3b2262a78fae13ee548e7cbc664afc7608769c1d21f3 WatchSource:0}: Error finding container 0edd1df7fac7341e52cf3b2262a78fae13ee548e7cbc664afc7608769c1d21f3: Status 404 returned error can't find the container with id 0edd1df7fac7341e52cf3b2262a78fae13ee548e7cbc664afc7608769c1d21f3 Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.818872 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh"] Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.821548 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zkr9c"] Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.831346 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pdv9r"] Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.838592 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-r2vhz" podStartSLOduration=126.838570849 podStartE2EDuration="2m6.838570849s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:24.837727816 +0000 UTC m=+150.601237439" watchObservedRunningTime="2025-12-01 14:36:24.838570849 +0000 UTC m=+150.602080452" Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.844850 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" event={"ID":"8e00f933-d775-4766-ae08-0d300d09fa4a","Type":"ContainerStarted","Data":"911c2efdccd7d0a05c67d77bc91f89c78869638c9e8e9c0f5d70a446410aac88"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.850099 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb" event={"ID":"6855b17f-5a68-45c7-a1b4-0c9cacc6ed14","Type":"ContainerStarted","Data":"418a456fd99d363f996753aad59298b5d7152e3b24f242ef07497cea1ba9834b"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.860863 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8tcbx" event={"ID":"f65c3828-a372-4660-97dd-e01d6efcb78c","Type":"ContainerStarted","Data":"9de934c9dd0cbb259cfc8bbaa915e0e65562752463c71daca07931128b663989"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.871544 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:24 crc kubenswrapper[4810]: E1201 14:36:24.872037 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:25.372021571 +0000 UTC m=+151.135531174 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.882267 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z" event={"ID":"ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19","Type":"ContainerStarted","Data":"68c08196c0b4a3c972e7a5799bd7050d5ab4c54409427a1c6525fadd9cf1196b"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.882320 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z" event={"ID":"ebc6f3ed-4a0e-44df-94b7-e9b9f9aafc19","Type":"ContainerStarted","Data":"d983f772914c3cc569a5d407b84fc2c9153c676ed75010b648d162d66916d367"} Dec 01 14:36:24 crc kubenswrapper[4810]: W1201 14:36:24.934120 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod246199f6_4473_41ee_ba76_804ee04e8aa6.slice/crio-1a1f4a5eb37eb52263f0160ef6a7bf1842cc55f8cfdc1b759ea7bbf94129595c WatchSource:0}: Error finding container 1a1f4a5eb37eb52263f0160ef6a7bf1842cc55f8cfdc1b759ea7bbf94129595c: Status 404 returned error can't find the container with id 1a1f4a5eb37eb52263f0160ef6a7bf1842cc55f8cfdc1b759ea7bbf94129595c Dec 01 14:36:24 crc kubenswrapper[4810]: W1201 14:36:24.959251 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40767ced_66c6_41a6_8e24_46e1bfb30c48.slice/crio-cd78e830e10a1b451737c1acdca2215b599092954950234a33f5ce6fb27813cf WatchSource:0}: Error finding container cd78e830e10a1b451737c1acdca2215b599092954950234a33f5ce6fb27813cf: Status 404 returned error can't find the container with id cd78e830e10a1b451737c1acdca2215b599092954950234a33f5ce6fb27813cf Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.961491 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" event={"ID":"ef1bca29-62d7-422b-890a-f93b965c0934","Type":"ContainerStarted","Data":"83813c5d6c40de9cc3baf26c9ba2b04db41e81679a455d033890fab431352808"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.961521 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.975833 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:24 crc kubenswrapper[4810]: E1201 14:36:24.978227 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:25.478205452 +0000 UTC m=+151.241715055 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.978837 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld" event={"ID":"484c6c08-536f-4374-971e-1fb4ef397b64","Type":"ContainerStarted","Data":"f70ec0e776cf2902bb7c7aeed924329467b17681c50f59716b54503e660cc7c1"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.984030 4810 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-4rgpg container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.984092 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" podUID="ef1bca29-62d7-422b-890a-f93b965c0934" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.996400 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" event={"ID":"ca08c59d-eb6c-4c1f-bf44-f161a760ae67","Type":"ContainerStarted","Data":"84a4a3351bfbfaafb13aaed6e7313179248d4768883d32685a5b24ec18ffbc40"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.996447 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" event={"ID":"ca08c59d-eb6c-4c1f-bf44-f161a760ae67","Type":"ContainerStarted","Data":"5417b4431dd53b77aa7188e3d7b50a6ce2617a247d0d5b9eda6b382a23435654"} Dec 01 14:36:24 crc kubenswrapper[4810]: I1201 14:36:24.997289 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.004581 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-hsnbk" event={"ID":"3863c075-f600-4d7b-b452-1ffa4b5c80f0","Type":"ContainerStarted","Data":"0e121421b7a2ece4c5236986ed3fd7bcbea73a4f94ce9589af1c2c9a06f2cb93"} Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.029147 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xxkbc" event={"ID":"93b94518-da01-439a-9f46-2329f86020a4","Type":"ContainerStarted","Data":"f7ec030f9c89448f56ccc9911063f61bd3f5f5f1feb9fcd82805ca60e87e4012"} Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.040973 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q" event={"ID":"1394009a-f690-4e2f-8086-87b9483fabbd","Type":"ContainerStarted","Data":"1c7b78fce806d4af283f3611b51a92f8d37bf3f482bd58896cae9c21f39f2b89"} Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.042225 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.083455 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:25 crc kubenswrapper[4810]: E1201 14:36:25.092787 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:25.592769309 +0000 UTC m=+151.356278912 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.160546 4810 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-2zf5q container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.160638 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q" podUID="1394009a-f690-4e2f-8086-87b9483fabbd" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.175954 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" event={"ID":"16c6f469-3ac7-46e3-84f3-a42187626e2b","Type":"ContainerStarted","Data":"8df2386374f24d7d91d290fe98ea911b42e921688325ba01500ceb2a041f8ca3"} Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.200039 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:25 crc kubenswrapper[4810]: E1201 14:36:25.200498 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:25.700460021 +0000 UTC m=+151.463969624 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.219436 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9" event={"ID":"5537fef3-e146-4eac-bc97-3edc211a2b3d","Type":"ContainerStarted","Data":"5770138bffe936211a5ae664a4cf7c4101ec1af9a7539c48a800039decd3228f"} Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.231937 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.232142 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n" event={"ID":"d2540546-e6b2-421d-8680-327f9cdeca7d","Type":"ContainerStarted","Data":"e012f08dd5579c7d6dec11405beab2c82ce433f964b907ff9a7629e7815fd7f1"} Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.250068 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-rf6tj" event={"ID":"88f6bb4e-362c-425c-9f67-81f76ed05e42","Type":"ContainerStarted","Data":"316774c10ead9265412a37a9d916d4462ac571a93bf59c4453ef5828e6bc5548"} Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.251125 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-rf6tj" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.259143 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" event={"ID":"0fa71e97-6be0-4300-bef2-b2dffbf30076","Type":"ContainerStarted","Data":"d402abb08a7945297d1cb224307fcbb1216733bd4da5c915507f50449ed641ec"} Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.265986 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.266688 4810 patch_prober.go:28] interesting pod/downloads-7954f5f757-rf6tj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.266760 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rf6tj" podUID="88f6bb4e-362c-425c-9f67-81f76ed05e42" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.277524 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" event={"ID":"266c7a73-0bce-4fad-b8c3-494b90a7d9b8","Type":"ContainerStarted","Data":"4a2d91fd94dde8b7099bd7531259624356ef9d77ddde3ff23338bbb4aa5e31d0"} Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.277826 4810 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cgzqr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.277851 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" podUID="0fa71e97-6be0-4300-bef2-b2dffbf30076" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.302932 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:25 crc kubenswrapper[4810]: E1201 14:36:25.304878 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:25.804863874 +0000 UTC m=+151.568373477 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.371905 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-5rxzl" podStartSLOduration=5.371890241 podStartE2EDuration="5.371890241s" podCreationTimestamp="2025-12-01 14:36:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:25.315663116 +0000 UTC m=+151.079172799" watchObservedRunningTime="2025-12-01 14:36:25.371890241 +0000 UTC m=+151.135399844" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.403893 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:25 crc kubenswrapper[4810]: E1201 14:36:25.404109 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:25.904094229 +0000 UTC m=+151.667603832 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.433077 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-2mxj4" podStartSLOduration=127.433057318 podStartE2EDuration="2m7.433057318s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:25.427369705 +0000 UTC m=+151.190879308" watchObservedRunningTime="2025-12-01 14:36:25.433057318 +0000 UTC m=+151.196566921" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.504974 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:25 crc kubenswrapper[4810]: E1201 14:36:25.505398 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:26.005387948 +0000 UTC m=+151.768897551 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.505451 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" podStartSLOduration=127.505432099 podStartE2EDuration="2m7.505432099s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:25.504174435 +0000 UTC m=+151.267684068" watchObservedRunningTime="2025-12-01 14:36:25.505432099 +0000 UTC m=+151.268941702" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.524603 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6qzb" podStartSLOduration=127.524575285 podStartE2EDuration="2m7.524575285s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:25.518642585 +0000 UTC m=+151.282152178" watchObservedRunningTime="2025-12-01 14:36:25.524575285 +0000 UTC m=+151.288084888" Dec 01 14:36:25 crc kubenswrapper[4810]: E1201 14:36:25.527018 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode62e2a4f_5d0d_4ee2_af70_56dc93b8ea95.slice/crio-b1bbb3fc83de48024818ad4ca33a23d8912b33a2bc1f76725cd48a732f784626.scope\": RecentStats: unable to find data in memory cache]" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.605265 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:25 crc kubenswrapper[4810]: E1201 14:36:25.606795 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:26.106754089 +0000 UTC m=+151.870263702 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.629771 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k6w5z" podStartSLOduration=127.629754849 podStartE2EDuration="2m7.629754849s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:25.628112635 +0000 UTC m=+151.391622238" watchObservedRunningTime="2025-12-01 14:36:25.629754849 +0000 UTC m=+151.393264452" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.673054 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" podStartSLOduration=127.673035526 podStartE2EDuration="2m7.673035526s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:25.67281831 +0000 UTC m=+151.436327913" watchObservedRunningTime="2025-12-01 14:36:25.673035526 +0000 UTC m=+151.436545129" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.708420 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q" podStartSLOduration=127.708388918 podStartE2EDuration="2m7.708388918s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:25.705807409 +0000 UTC m=+151.469317012" watchObservedRunningTime="2025-12-01 14:36:25.708388918 +0000 UTC m=+151.471898521" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.709640 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:25 crc kubenswrapper[4810]: E1201 14:36:25.709917 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:26.209904159 +0000 UTC m=+151.973413762 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.743544 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" podStartSLOduration=127.743526365 podStartE2EDuration="2m7.743526365s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:25.737519703 +0000 UTC m=+151.501029306" watchObservedRunningTime="2025-12-01 14:36:25.743526365 +0000 UTC m=+151.507035968" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.803826 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-fcdlt" podStartSLOduration=127.803806839 podStartE2EDuration="2m7.803806839s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:25.776756621 +0000 UTC m=+151.540266224" watchObservedRunningTime="2025-12-01 14:36:25.803806839 +0000 UTC m=+151.567316442" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.805523 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-rf6tj" podStartSLOduration=127.805513825 podStartE2EDuration="2m7.805513825s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:25.802977078 +0000 UTC m=+151.566486691" watchObservedRunningTime="2025-12-01 14:36:25.805513825 +0000 UTC m=+151.569023428" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.816684 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:25 crc kubenswrapper[4810]: E1201 14:36:25.816855 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:26.31681572 +0000 UTC m=+152.080325323 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.817048 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:25 crc kubenswrapper[4810]: E1201 14:36:25.817353 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:26.317336654 +0000 UTC m=+152.080846257 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.829977 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xxkbc" podStartSLOduration=127.829950364 podStartE2EDuration="2m7.829950364s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:25.824887107 +0000 UTC m=+151.588396740" watchObservedRunningTime="2025-12-01 14:36:25.829950364 +0000 UTC m=+151.593459967" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.896405 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" podStartSLOduration=127.896387704 podStartE2EDuration="2m7.896387704s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:25.889859319 +0000 UTC m=+151.653368922" watchObservedRunningTime="2025-12-01 14:36:25.896387704 +0000 UTC m=+151.659897307" Dec 01 14:36:25 crc kubenswrapper[4810]: I1201 14:36:25.918639 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:25 crc kubenswrapper[4810]: E1201 14:36:25.919005 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:26.418988443 +0000 UTC m=+152.182498046 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.020315 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:26 crc kubenswrapper[4810]: E1201 14:36:26.021392 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:26.521377013 +0000 UTC m=+152.284886616 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.122044 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:26 crc kubenswrapper[4810]: E1201 14:36:26.122523 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:26.622506838 +0000 UTC m=+152.386016441 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.223311 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:26 crc kubenswrapper[4810]: E1201 14:36:26.223672 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:26.723659243 +0000 UTC m=+152.487168846 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.335328 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:26 crc kubenswrapper[4810]: E1201 14:36:26.335627 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:26.83561228 +0000 UTC m=+152.599121883 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.344641 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5267z" event={"ID":"367566c7-8e32-444f-866a-d64b98a2f596","Type":"ContainerStarted","Data":"72f7a6f38ef7db6dbb35eccad12f68e88601ee93c0ac8942266c879e15658a3f"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.347262 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" event={"ID":"2781f96f-4018-4ccf-adb8-04512de46450","Type":"ContainerStarted","Data":"b47c2283c4801a4c150b27a68e56f7c0e00a233f6a283c2c7baddd235bee5f79"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.389998 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" event={"ID":"0fa71e97-6be0-4300-bef2-b2dffbf30076","Type":"ContainerStarted","Data":"987552bf4a9cd3bc3b0eb15466951f1c7575b07dd03b1505f1f8e6c0d30c382e"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.402335 4810 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-cgzqr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.402398 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" podUID="0fa71e97-6be0-4300-bef2-b2dffbf30076" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.420384 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q" event={"ID":"1394009a-f690-4e2f-8086-87b9483fabbd","Type":"ContainerStarted","Data":"2a03c549a868b6cbeeb4c2a84b307bed7e2129521c8984b87f5e4e853140fab5"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.439430 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:26 crc kubenswrapper[4810]: E1201 14:36:26.439768 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:26.939753946 +0000 UTC m=+152.703263559 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.440210 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2zf5q" Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.458576 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" event={"ID":"ef1bca29-62d7-422b-890a-f93b965c0934","Type":"ContainerStarted","Data":"2b6abaae006db7ef297be63c129a23773a98f62a25b7d76c02e9e200644030d6"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.474112 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.475327 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9r9fc" event={"ID":"54284225-c472-482c-93b2-2426031f0957","Type":"ContainerStarted","Data":"3026735745e1a1c6bf0c9f3ba0e7b7f3cf400ee688ab2342a5d23db197c2741c"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.515515 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wvgpw" event={"ID":"e69cc34e-b76e-4a7e-a4b5-c7e8436167e7","Type":"ContainerStarted","Data":"60ebfb01ed3308d831bd7b36b3b0182c59af3475cc49330b0385c23f2bf4a9ad"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.524309 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zkr9c" event={"ID":"985c8210-eb5f-41da-8b74-95d7a6e7e9c7","Type":"ContainerStarted","Data":"61374404cf0691cc3b7dd80b13eafb1268b5004c468e685c0748003edf68b11a"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.542108 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:26 crc kubenswrapper[4810]: E1201 14:36:26.543203 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:27.043133312 +0000 UTC m=+152.806642915 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.544227 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wvgpw" podStartSLOduration=128.544215712 podStartE2EDuration="2m8.544215712s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:26.535152407 +0000 UTC m=+152.298662040" watchObservedRunningTime="2025-12-01 14:36:26.544215712 +0000 UTC m=+152.307725315" Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.566814 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n6q5s" event={"ID":"a543c4e2-99d9-4b47-a9c6-927c8737ed49","Type":"ContainerStarted","Data":"494dfd98e0092791edd9cdf3cf0a4e3eabde48c8e2658205d2bd22ffb68626ef"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.579419 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pjt5m" event={"ID":"3e6111a1-b6ac-4a66-b8f6-efef4463b6df","Type":"ContainerStarted","Data":"666204bc0779e70eccd60acd97da700c4de1cba4ebb8b4c00221986e171b29cb"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.579833 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9r9fc" podStartSLOduration=128.57979111 podStartE2EDuration="2m8.57979111s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:26.576321847 +0000 UTC m=+152.339831450" watchObservedRunningTime="2025-12-01 14:36:26.57979111 +0000 UTC m=+152.343300723" Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.591748 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h54rq" event={"ID":"246199f6-4473-41ee-ba76-804ee04e8aa6","Type":"ContainerStarted","Data":"c66a89a2ae2a459a3aa414f81383b6b01d66e06b2710be95a9d463c941d30c06"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.591795 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h54rq" event={"ID":"246199f6-4473-41ee-ba76-804ee04e8aa6","Type":"ContainerStarted","Data":"1a1f4a5eb37eb52263f0160ef6a7bf1842cc55f8cfdc1b759ea7bbf94129595c"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.596830 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xtf78" event={"ID":"4dea83f8-89e3-4a33-a4f0-fbb0619ab1df","Type":"ContainerStarted","Data":"b32c5279da941e58c5cae8a90d1d1d2af975c85fba3a7249905ca97d8066c1b2"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.608197 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-n6q5s" podStartSLOduration=128.608176946 podStartE2EDuration="2m8.608176946s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:26.607291501 +0000 UTC m=+152.370801104" watchObservedRunningTime="2025-12-01 14:36:26.608176946 +0000 UTC m=+152.371686549" Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.643184 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n" event={"ID":"d2540546-e6b2-421d-8680-327f9cdeca7d","Type":"ContainerStarted","Data":"18be9fffc4567db6d4730e989fabce0d6560df9db80841b33efed9553ae3bcdb"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.651336 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:26 crc kubenswrapper[4810]: E1201 14:36:26.659556 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:27.159533499 +0000 UTC m=+152.923043102 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.724231 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-hsnbk" event={"ID":"3863c075-f600-4d7b-b452-1ffa4b5c80f0","Type":"ContainerStarted","Data":"5f414b6539f1636920db1222ea910ce3c85fdc022c97b8add86157972b3eb011"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.730847 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9" event={"ID":"5537fef3-e146-4eac-bc97-3edc211a2b3d","Type":"ContainerStarted","Data":"c0e65ff769e50c16b76af6abfc6aaa2200c70a4eac76d6d71f06843dcaf95e6a"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.735539 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-pjt5m" podStartSLOduration=128.735522757 podStartE2EDuration="2m8.735522757s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:26.696776583 +0000 UTC m=+152.460286186" watchObservedRunningTime="2025-12-01 14:36:26.735522757 +0000 UTC m=+152.499032360" Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.736332 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xlw6n" podStartSLOduration=128.736325598 podStartE2EDuration="2m8.736325598s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:26.733685007 +0000 UTC m=+152.497194610" watchObservedRunningTime="2025-12-01 14:36:26.736325598 +0000 UTC m=+152.499835201" Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.752641 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:26 crc kubenswrapper[4810]: E1201 14:36:26.753227 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:27.253212643 +0000 UTC m=+153.016722246 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.769812 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-5rxzl" event={"ID":"d08fa7f1-84f4-4909-ada5-f264d2215ba9","Type":"ContainerStarted","Data":"8367c72fb656acd8238797133386a802f97fbf344ad2fcb07208097abee91bde"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.781926 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-84qz9" podStartSLOduration=128.781910467 podStartE2EDuration="2m8.781910467s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:26.780954531 +0000 UTC m=+152.544464134" watchObservedRunningTime="2025-12-01 14:36:26.781910467 +0000 UTC m=+152.545420070" Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.795704 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s5rbr" event={"ID":"cd5edc05-8e37-4d29-a7e5-34fa7e39bfa4","Type":"ContainerStarted","Data":"2721b5aa9c0c4a6c5ef7be5935d673b24f1196d540aca806fd8e511aba5972b6"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.839046 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n67g2" event={"ID":"6705a032-c5d3-45be-acd4-94097a6c9ca6","Type":"ContainerStarted","Data":"f6fa7680ab3ee3519a06e3cd49cbd97a23e52542a175204fb7b37e3739bec523"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.840085 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n67g2" event={"ID":"6705a032-c5d3-45be-acd4-94097a6c9ca6","Type":"ContainerStarted","Data":"812856eae75e63c8cf13f0a788c871ce696a0da412753802853a6a0dcc2539ca"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.842223 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-hsnbk" podStartSLOduration=128.842187221 podStartE2EDuration="2m8.842187221s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:26.824216517 +0000 UTC m=+152.587726120" watchObservedRunningTime="2025-12-01 14:36:26.842187221 +0000 UTC m=+152.605696824" Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.855786 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:26 crc kubenswrapper[4810]: E1201 14:36:26.861063 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:27.361038519 +0000 UTC m=+153.124548122 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.862618 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-88hcl" event={"ID":"e62e2a4f-5d0d-4ee2-af70-56dc93b8ea95","Type":"ContainerStarted","Data":"b1bbb3fc83de48024818ad4ca33a23d8912b33a2bc1f76725cd48a732f784626"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.877801 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-n67g2" podStartSLOduration=128.87778509 podStartE2EDuration="2m8.87778509s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:26.874174094 +0000 UTC m=+152.637683697" watchObservedRunningTime="2025-12-01 14:36:26.87778509 +0000 UTC m=+152.641294683" Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.888767 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-hqhng" event={"ID":"c95ce2ca-f352-422f-ba4d-6de8dbdc14b7","Type":"ContainerStarted","Data":"0edd1df7fac7341e52cf3b2262a78fae13ee548e7cbc664afc7608769c1d21f3"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.911596 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-88hcl" podStartSLOduration=6.911579211 podStartE2EDuration="6.911579211s" podCreationTimestamp="2025-12-01 14:36:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:26.908710004 +0000 UTC m=+152.672219607" watchObservedRunningTime="2025-12-01 14:36:26.911579211 +0000 UTC m=+152.675088814" Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.963401 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.967004 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" event={"ID":"6dbca0cc-145c-41df-823a-eec2e7359ace","Type":"ContainerStarted","Data":"30dc5b5c01afba7b996c60aa009b5852ad7f6e436e44f2a2cd75a43e46124cf4"} Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.969619 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" Dec 01 14:36:26 crc kubenswrapper[4810]: E1201 14:36:26.978190 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:27.478155155 +0000 UTC m=+153.241664758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:26 crc kubenswrapper[4810]: I1201 14:36:26.995608 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dt9dw" event={"ID":"a3c66172-24da-4de2-b16d-f09a56ece4d5","Type":"ContainerStarted","Data":"a1bcfb4d5c56aa2bf428f71f8209874edb931c90bd1c545f85844fa872e21d4e"} Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.010606 4810 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-qh9j7 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" start-of-body= Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.010652 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" podUID="6dbca0cc-145c-41df-823a-eec2e7359ace" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.013167 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" event={"ID":"266c7a73-0bce-4fad-b8c3-494b90a7d9b8","Type":"ContainerStarted","Data":"841fec6860b2335e95f78e88ffce0ffb266716b78aa0e1354f40007bc8282193"} Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.026943 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" event={"ID":"17423ab4-d417-42d3-84f0-ca131f3dceeb","Type":"ContainerStarted","Data":"e823db0361cd900d83c51c98cd956f3477573c366b6709de882adab49dff7fdd"} Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.028291 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8tcbx" event={"ID":"f65c3828-a372-4660-97dd-e01d6efcb78c","Type":"ContainerStarted","Data":"3dd349b806f5cbfea3fb30a21c954f8fa69445b9882244563dba5a8d0a3d14fe"} Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.042067 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-hqhng" podStartSLOduration=129.042051968 podStartE2EDuration="2m9.042051968s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:26.976448619 +0000 UTC m=+152.739958222" watchObservedRunningTime="2025-12-01 14:36:27.042051968 +0000 UTC m=+152.805561571" Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.068637 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-tlqrt" event={"ID":"afa8aca0-8b8b-49de-ac08-746a5936fb3a","Type":"ContainerStarted","Data":"0dadd96bfd9c4e8e626f917a7c3ef83cdcc212c8173a91eb99d0fba0b6f35695"} Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.077327 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-t7pxb" podStartSLOduration=129.077313338 podStartE2EDuration="2m9.077313338s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:27.076218868 +0000 UTC m=+152.839728471" watchObservedRunningTime="2025-12-01 14:36:27.077313338 +0000 UTC m=+152.840822941" Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.079214 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" podStartSLOduration=129.079207738 podStartE2EDuration="2m9.079207738s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:27.04438947 +0000 UTC m=+152.807899073" watchObservedRunningTime="2025-12-01 14:36:27.079207738 +0000 UTC m=+152.842717341" Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.080754 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:27 crc kubenswrapper[4810]: E1201 14:36:27.083663 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:27.583650318 +0000 UTC m=+153.347159921 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.091747 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld" event={"ID":"484c6c08-536f-4374-971e-1fb4ef397b64","Type":"ContainerStarted","Data":"b3feacdd440d69f9f0233048aea0de0190a3da88ac5e9deaeb3af2daf12d95dc"} Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.094679 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-rf6tj" event={"ID":"88f6bb4e-362c-425c-9f67-81f76ed05e42","Type":"ContainerStarted","Data":"681d84a533f82380ae3bb3d38e045ffd33c66c9be385a42ee66f7b7ae7c4db9e"} Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.095849 4810 patch_prober.go:28] interesting pod/downloads-7954f5f757-rf6tj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.095954 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rf6tj" podUID="88f6bb4e-362c-425c-9f67-81f76ed05e42" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.101363 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jnrsh" event={"ID":"62d665ff-bba5-4ed3-b0f2-471be90260b7","Type":"ContainerStarted","Data":"1055ba55e39f209c31b74cdadec78d3785684fd0ec916704322587df6985b16a"} Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.123654 4810 generic.go:334] "Generic (PLEG): container finished" podID="16c6f469-3ac7-46e3-84f3-a42187626e2b" containerID="c2c50759a8ae7dc571efbcb481ee46052cd13230f538045b63a6f8b731d59a52" exitCode=0 Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.123766 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" event={"ID":"16c6f469-3ac7-46e3-84f3-a42187626e2b","Type":"ContainerDied","Data":"c2c50759a8ae7dc571efbcb481ee46052cd13230f538045b63a6f8b731d59a52"} Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.127070 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-8tcbx" podStartSLOduration=129.127057527 podStartE2EDuration="2m9.127057527s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:27.125374612 +0000 UTC m=+152.888884205" watchObservedRunningTime="2025-12-01 14:36:27.127057527 +0000 UTC m=+152.890567130" Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.137131 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh" event={"ID":"cf666da3-0465-4de2-9cea-7355839709f7","Type":"ContainerStarted","Data":"f576c661e2c58223ddef0a3de6fc20c36b032fe286bdbc888eba062561d89799"} Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.137870 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh" Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.164632 4810 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-9x2jh container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.164686 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh" podUID="cf666da3-0465-4de2-9cea-7355839709f7" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.176833 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5hpvd" event={"ID":"40767ced-66c6-41a6-8e24-46e1bfb30c48","Type":"ContainerStarted","Data":"cd78e830e10a1b451737c1acdca2215b599092954950234a33f5ce6fb27813cf"} Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.182076 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:27 crc kubenswrapper[4810]: E1201 14:36:27.182584 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:27.682567714 +0000 UTC m=+153.446077317 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.182626 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:27 crc kubenswrapper[4810]: E1201 14:36:27.183860 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:27.683852038 +0000 UTC m=+153.447361631 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.221166 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jnrsh" podStartSLOduration=129.221151973 podStartE2EDuration="2m9.221151973s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:27.219417826 +0000 UTC m=+152.982927429" watchObservedRunningTime="2025-12-01 14:36:27.221151973 +0000 UTC m=+152.984661576" Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.236153 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb" event={"ID":"6855b17f-5a68-45c7-a1b4-0c9cacc6ed14","Type":"ContainerStarted","Data":"83efa22ad27db7c978df79c9355ae75abaf73cf0e5c3aab7bb4dc678b71105b5"} Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.273171 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld" podStartSLOduration=129.273154135 podStartE2EDuration="2m9.273154135s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:27.271605962 +0000 UTC m=+153.035115555" watchObservedRunningTime="2025-12-01 14:36:27.273154135 +0000 UTC m=+153.036663738" Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.284057 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:27 crc kubenswrapper[4810]: E1201 14:36:27.284358 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:27.784343407 +0000 UTC m=+153.547853010 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.307953 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-2mxj4" Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.384385 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh" podStartSLOduration=129.384367071 podStartE2EDuration="2m9.384367071s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:27.328395273 +0000 UTC m=+153.091904876" watchObservedRunningTime="2025-12-01 14:36:27.384367071 +0000 UTC m=+153.147876674" Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.386529 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:27 crc kubenswrapper[4810]: E1201 14:36:27.394182 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:27.894168976 +0000 UTC m=+153.657678569 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.437679 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb" podStartSLOduration=129.437662308 podStartE2EDuration="2m9.437662308s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:27.386979452 +0000 UTC m=+153.150489055" watchObservedRunningTime="2025-12-01 14:36:27.437662308 +0000 UTC m=+153.201171911" Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.485908 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.487226 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:27 crc kubenswrapper[4810]: E1201 14:36:27.487513 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:27.987494231 +0000 UTC m=+153.751003834 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.493623 4810 patch_prober.go:28] interesting pod/router-default-5444994796-hsnbk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 14:36:27 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Dec 01 14:36:27 crc kubenswrapper[4810]: [+]process-running ok Dec 01 14:36:27 crc kubenswrapper[4810]: healthz check failed Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.493674 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hsnbk" podUID="3863c075-f600-4d7b-b452-1ffa4b5c80f0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.588515 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:27 crc kubenswrapper[4810]: E1201 14:36:27.588813 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:28.088801091 +0000 UTC m=+153.852310694 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.689705 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:27 crc kubenswrapper[4810]: E1201 14:36:27.690012 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:28.189998407 +0000 UTC m=+153.953508010 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.790962 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:27 crc kubenswrapper[4810]: E1201 14:36:27.791338 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:28.291323498 +0000 UTC m=+154.054833091 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.892359 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:27 crc kubenswrapper[4810]: E1201 14:36:27.892637 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:28.392619628 +0000 UTC m=+154.156129231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.893289 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:27 crc kubenswrapper[4810]: E1201 14:36:27.893636 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:28.393628275 +0000 UTC m=+154.157137878 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:27 crc kubenswrapper[4810]: I1201 14:36:27.998000 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:27 crc kubenswrapper[4810]: E1201 14:36:27.998286 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:28.498269354 +0000 UTC m=+154.261778957 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.100131 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:28 crc kubenswrapper[4810]: E1201 14:36:28.100426 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:28.600410977 +0000 UTC m=+154.363920580 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.200877 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:28 crc kubenswrapper[4810]: E1201 14:36:28.201076 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:28.701046379 +0000 UTC m=+154.464555982 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.202211 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:28 crc kubenswrapper[4810]: E1201 14:36:28.202545 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:28.702536969 +0000 UTC m=+154.466046572 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.275762 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xtf78" event={"ID":"4dea83f8-89e3-4a33-a4f0-fbb0619ab1df","Type":"ContainerStarted","Data":"d6f26b925c7a118617ccc6775263a218a16e484899761cc14e6040606d021a65"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.293858 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s5rbr" event={"ID":"cd5edc05-8e37-4d29-a7e5-34fa7e39bfa4","Type":"ContainerStarted","Data":"b885b656982e1acc518caa405c0e2db03df1f0e22312450944181b539ff2fed5"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.293917 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s5rbr" event={"ID":"cd5edc05-8e37-4d29-a7e5-34fa7e39bfa4","Type":"ContainerStarted","Data":"c2db0a584d2a36071b0be07a42c1b3563e45ddc9850d7401ec5da20261fb057f"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.302999 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:28 crc kubenswrapper[4810]: E1201 14:36:28.303309 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:28.803286954 +0000 UTC m=+154.566796557 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.311359 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5hpvd" event={"ID":"40767ced-66c6-41a6-8e24-46e1bfb30c48","Type":"ContainerStarted","Data":"5ba0e5ea1cb9ef7283aff844dc8b1db92b5309b383577eadf6518c2a7ac9d87d"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.311407 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5hpvd" event={"ID":"40767ced-66c6-41a6-8e24-46e1bfb30c48","Type":"ContainerStarted","Data":"9133b9d1244e9dece345515f5bfa6e76b897b45d99a5099df25812fc34978ec9"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.344295 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xtf78" podStartSLOduration=130.344280889 podStartE2EDuration="2m10.344280889s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:28.30423708 +0000 UTC m=+154.067746683" watchObservedRunningTime="2025-12-01 14:36:28.344280889 +0000 UTC m=+154.107790492" Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.345216 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s5rbr" podStartSLOduration=130.345209993 podStartE2EDuration="2m10.345209993s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:28.343864867 +0000 UTC m=+154.107374470" watchObservedRunningTime="2025-12-01 14:36:28.345209993 +0000 UTC m=+154.108719596" Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.354905 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-l56lb" event={"ID":"6855b17f-5a68-45c7-a1b4-0c9cacc6ed14","Type":"ContainerStarted","Data":"9fff1173841f778cce2bcf303e8ec458126627bb238809e7c0aa1c5600ebb89f"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.366888 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" event={"ID":"16c6f469-3ac7-46e3-84f3-a42187626e2b","Type":"ContainerStarted","Data":"e348564aaa6400240653f78b9ea158dbd8de198850443772a8eb0c7650a9d214"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.373425 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-5hpvd" podStartSLOduration=130.373399873 podStartE2EDuration="2m10.373399873s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:28.371315277 +0000 UTC m=+154.134824880" watchObservedRunningTime="2025-12-01 14:36:28.373399873 +0000 UTC m=+154.136909476" Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.386902 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zkr9c" event={"ID":"985c8210-eb5f-41da-8b74-95d7a6e7e9c7","Type":"ContainerStarted","Data":"b400f249a2b0c4e2899edd44529c72093f52cf2dff4b1473c5c94f940101b2c8"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.387118 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zkr9c" Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.387140 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zkr9c" event={"ID":"985c8210-eb5f-41da-8b74-95d7a6e7e9c7","Type":"ContainerStarted","Data":"8dffd3c08855276eaaa06dfc73618b9bf1c546513ec30c3399d6b505694abb4e"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.404456 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:28 crc kubenswrapper[4810]: E1201 14:36:28.404847 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:28.904833741 +0000 UTC m=+154.668343344 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.410234 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-hqhng" event={"ID":"c95ce2ca-f352-422f-ba4d-6de8dbdc14b7","Type":"ContainerStarted","Data":"fe93dda5178dba833518325c2eb954c62c0b0130cf7ff374666bd46b17482f20"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.419653 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" podStartSLOduration=130.419571647 podStartE2EDuration="2m10.419571647s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:28.418146189 +0000 UTC m=+154.181655802" watchObservedRunningTime="2025-12-01 14:36:28.419571647 +0000 UTC m=+154.183081250" Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.424759 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jnrsh" event={"ID":"62d665ff-bba5-4ed3-b0f2-471be90260b7","Type":"ContainerStarted","Data":"4940e891885365d6a99d4aca82f1052821d75d8cc2d18c5a53ce4e3a3b828c10"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.438503 4810 generic.go:334] "Generic (PLEG): container finished" podID="246199f6-4473-41ee-ba76-804ee04e8aa6" containerID="c66a89a2ae2a459a3aa414f81383b6b01d66e06b2710be95a9d463c941d30c06" exitCode=0 Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.438592 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h54rq" event={"ID":"246199f6-4473-41ee-ba76-804ee04e8aa6","Type":"ContainerDied","Data":"c66a89a2ae2a459a3aa414f81383b6b01d66e06b2710be95a9d463c941d30c06"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.456049 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" event={"ID":"2781f96f-4018-4ccf-adb8-04512de46450","Type":"ContainerStarted","Data":"a1fcfe841762e9d1079667836ab1594c21a5a34842a963ff6191f0d8a4acbd3a"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.456342 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zkr9c" podStartSLOduration=130.456331069 podStartE2EDuration="2m10.456331069s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:28.45603959 +0000 UTC m=+154.219549193" watchObservedRunningTime="2025-12-01 14:36:28.456331069 +0000 UTC m=+154.219840682" Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.466097 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dt9dw" event={"ID":"a3c66172-24da-4de2-b16d-f09a56ece4d5","Type":"ContainerStarted","Data":"155916e122dc075fd53ed1b9f688294199fc79e2eb9ecb0fcce1a210f0c67e91"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.466149 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dt9dw" event={"ID":"a3c66172-24da-4de2-b16d-f09a56ece4d5","Type":"ContainerStarted","Data":"1112f580db4ed6310dd9ccc683573179a79a06717c6c286d817cab5f8968bba0"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.466403 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-dt9dw" Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.468293 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh" event={"ID":"cf666da3-0465-4de2-9cea-7355839709f7","Type":"ContainerStarted","Data":"a5b3e8cfdc6fee853b83ffd320f8bed2acea58c471cb63f5456311c6e01f0b49"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.471055 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5267z" event={"ID":"367566c7-8e32-444f-866a-d64b98a2f596","Type":"ContainerStarted","Data":"b124876dcc6bc704dc2462996c66a548c9ffc801c2421ef714860726eee41373"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.472312 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" event={"ID":"17423ab4-d417-42d3-84f0-ca131f3dceeb","Type":"ContainerStarted","Data":"98cabe242f8a4060a2ef8725015887724cc076d11a48cf9644c0168f06411ef3"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.484064 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-tlqrt" event={"ID":"afa8aca0-8b8b-49de-ac08-746a5936fb3a","Type":"ContainerStarted","Data":"547a4fe435270a04133fd0a5c1e8ca752330e0cc7a104399d0186fb224318cbf"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.484111 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-tlqrt" event={"ID":"afa8aca0-8b8b-49de-ac08-746a5936fb3a","Type":"ContainerStarted","Data":"1c007c7e65240ef55ea01b7b5c17134db041c582494ef1874d076ebb851c4619"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.504796 4810 patch_prober.go:28] interesting pod/downloads-7954f5f757-rf6tj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.504844 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rf6tj" podUID="88f6bb4e-362c-425c-9f67-81f76ed05e42" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.505453 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.505558 4810 patch_prober.go:28] interesting pod/router-default-5444994796-hsnbk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 14:36:28 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Dec 01 14:36:28 crc kubenswrapper[4810]: [+]process-running ok Dec 01 14:36:28 crc kubenswrapper[4810]: healthz check failed Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.505659 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hsnbk" podUID="3863c075-f600-4d7b-b452-1ffa4b5c80f0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 14:36:28 crc kubenswrapper[4810]: E1201 14:36:28.507820 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:29.007805125 +0000 UTC m=+154.771314728 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.553164 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9x2jh" Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.553206 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" event={"ID":"6dbca0cc-145c-41df-823a-eec2e7359ace","Type":"ContainerStarted","Data":"66b0a378771bf28169d974e0c98bd0ce99c59369ee473e79bda5584f1c9c29b7"} Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.553243 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.592159 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-5267z" podStartSLOduration=130.592140688 podStartE2EDuration="2m10.592140688s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:28.563102816 +0000 UTC m=+154.326612419" watchObservedRunningTime="2025-12-01 14:36:28.592140688 +0000 UTC m=+154.355650291" Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.610657 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:28 crc kubenswrapper[4810]: E1201 14:36:28.636130 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:29.136099343 +0000 UTC m=+154.899608946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.667665 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-dt9dw" podStartSLOduration=8.667631733 podStartE2EDuration="8.667631733s" podCreationTimestamp="2025-12-01 14:36:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:28.592713433 +0000 UTC m=+154.356223036" watchObservedRunningTime="2025-12-01 14:36:28.667631733 +0000 UTC m=+154.431141336" Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.667895 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" podStartSLOduration=130.66788884 podStartE2EDuration="2m10.66788884s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:28.641586791 +0000 UTC m=+154.405096424" watchObservedRunningTime="2025-12-01 14:36:28.66788884 +0000 UTC m=+154.431398443" Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.712083 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:28 crc kubenswrapper[4810]: E1201 14:36:28.713221 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:29.2132046 +0000 UTC m=+154.976714203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.780404 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qh9j7" Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.788935 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-tlqrt" podStartSLOduration=130.788920371 podStartE2EDuration="2m10.788920371s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:28.787816591 +0000 UTC m=+154.551326194" watchObservedRunningTime="2025-12-01 14:36:28.788920371 +0000 UTC m=+154.552429974" Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.814368 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:28 crc kubenswrapper[4810]: E1201 14:36:28.814701 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:29.314689576 +0000 UTC m=+155.078199179 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:28 crc kubenswrapper[4810]: I1201 14:36:28.918986 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:28 crc kubenswrapper[4810]: E1201 14:36:28.919414 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:29.419395456 +0000 UTC m=+155.182905069 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.021181 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:29 crc kubenswrapper[4810]: E1201 14:36:29.021936 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:29.52192288 +0000 UTC m=+155.285432483 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.053758 4810 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.123123 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:29 crc kubenswrapper[4810]: E1201 14:36:29.123410 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:29.623362103 +0000 UTC m=+155.386871706 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.224650 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:29 crc kubenswrapper[4810]: E1201 14:36:29.225117 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:29.725097245 +0000 UTC m=+155.488606858 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.325931 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:29 crc kubenswrapper[4810]: E1201 14:36:29.326079 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:29.826050285 +0000 UTC m=+155.589559898 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.326145 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:29 crc kubenswrapper[4810]: E1201 14:36:29.326427 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:29.826416005 +0000 UTC m=+155.589925608 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.427718 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:29 crc kubenswrapper[4810]: E1201 14:36:29.427924 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:29.92788869 +0000 UTC m=+155.691398283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.428779 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:29 crc kubenswrapper[4810]: E1201 14:36:29.429121 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:29.929109562 +0000 UTC m=+155.692619155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.484143 4810 patch_prober.go:28] interesting pod/router-default-5444994796-hsnbk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 14:36:29 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Dec 01 14:36:29 crc kubenswrapper[4810]: [+]process-running ok Dec 01 14:36:29 crc kubenswrapper[4810]: healthz check failed Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.484205 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hsnbk" podUID="3863c075-f600-4d7b-b452-1ffa4b5c80f0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.502530 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h54rq" event={"ID":"246199f6-4473-41ee-ba76-804ee04e8aa6","Type":"ContainerStarted","Data":"28882164555a8addf19559397f3e7007c8f69e41cc4d91890cb0d18902dceca7"} Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.502623 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h54rq" Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.506363 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" event={"ID":"17423ab4-d417-42d3-84f0-ca131f3dceeb","Type":"ContainerStarted","Data":"0f663be8a13b4f61b5f913973e00aff8534ae215527785f93a18cf4628e2a37f"} Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.506408 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" event={"ID":"17423ab4-d417-42d3-84f0-ca131f3dceeb","Type":"ContainerStarted","Data":"dc55542dd7e4638bcad8a895d4b5b768904cf00ec4f372ae44646292933c9a35"} Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.530569 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:29 crc kubenswrapper[4810]: E1201 14:36:29.530743 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:30.03070796 +0000 UTC m=+155.794217563 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.530796 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:29 crc kubenswrapper[4810]: E1201 14:36:29.531173 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:30.031162072 +0000 UTC m=+155.794671745 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.631496 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:29 crc kubenswrapper[4810]: E1201 14:36:29.631763 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 14:36:30.131715092 +0000 UTC m=+155.895224695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.635616 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:29 crc kubenswrapper[4810]: E1201 14:36:29.638803 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 14:36:30.138779112 +0000 UTC m=+155.902288925 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qvm7d" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.642879 4810 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-01T14:36:29.053789578Z","Handler":null,"Name":""} Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.680154 4810 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.680215 4810 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.738269 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.768176 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.839419 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.857692 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.857754 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.860798 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h54rq" podStartSLOduration=131.860781845 podStartE2EDuration="2m11.860781845s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:29.573210935 +0000 UTC m=+155.336720538" watchObservedRunningTime="2025-12-01 14:36:29.860781845 +0000 UTC m=+155.624291458" Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.862010 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pxt6s"] Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.863517 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxt6s" Dec 01 14:36:29 crc kubenswrapper[4810]: W1201 14:36:29.865830 4810 reflector.go:561] object-"openshift-marketplace"/"community-operators-dockercfg-dmngl": failed to list *v1.Secret: secrets "community-operators-dockercfg-dmngl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Dec 01 14:36:29 crc kubenswrapper[4810]: E1201 14:36:29.865873 4810 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"community-operators-dockercfg-dmngl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"community-operators-dockercfg-dmngl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.884276 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pxt6s"] Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.941038 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5-catalog-content\") pod \"community-operators-pxt6s\" (UID: \"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5\") " pod="openshift-marketplace/community-operators-pxt6s" Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.941286 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5-utilities\") pod \"community-operators-pxt6s\" (UID: \"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5\") " pod="openshift-marketplace/community-operators-pxt6s" Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.941314 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfwql\" (UniqueName: \"kubernetes.io/projected/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5-kube-api-access-hfwql\") pod \"community-operators-pxt6s\" (UID: \"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5\") " pod="openshift-marketplace/community-operators-pxt6s" Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.959443 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qvm7d\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.964080 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fm8lj"] Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.965337 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fm8lj" Dec 01 14:36:29 crc kubenswrapper[4810]: I1201 14:36:29.966813 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.019835 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fm8lj"] Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.042119 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca896d69-0651-48e1-a001-0b3c57e1ece4-catalog-content\") pod \"certified-operators-fm8lj\" (UID: \"ca896d69-0651-48e1-a001-0b3c57e1ece4\") " pod="openshift-marketplace/certified-operators-fm8lj" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.042187 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca896d69-0651-48e1-a001-0b3c57e1ece4-utilities\") pod \"certified-operators-fm8lj\" (UID: \"ca896d69-0651-48e1-a001-0b3c57e1ece4\") " pod="openshift-marketplace/certified-operators-fm8lj" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.042217 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5-catalog-content\") pod \"community-operators-pxt6s\" (UID: \"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5\") " pod="openshift-marketplace/community-operators-pxt6s" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.042241 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5-utilities\") pod \"community-operators-pxt6s\" (UID: \"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5\") " pod="openshift-marketplace/community-operators-pxt6s" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.042269 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zrjs\" (UniqueName: \"kubernetes.io/projected/ca896d69-0651-48e1-a001-0b3c57e1ece4-kube-api-access-5zrjs\") pod \"certified-operators-fm8lj\" (UID: \"ca896d69-0651-48e1-a001-0b3c57e1ece4\") " pod="openshift-marketplace/certified-operators-fm8lj" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.042296 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfwql\" (UniqueName: \"kubernetes.io/projected/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5-kube-api-access-hfwql\") pod \"community-operators-pxt6s\" (UID: \"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5\") " pod="openshift-marketplace/community-operators-pxt6s" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.042964 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5-catalog-content\") pod \"community-operators-pxt6s\" (UID: \"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5\") " pod="openshift-marketplace/community-operators-pxt6s" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.042973 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5-utilities\") pod \"community-operators-pxt6s\" (UID: \"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5\") " pod="openshift-marketplace/community-operators-pxt6s" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.074589 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfwql\" (UniqueName: \"kubernetes.io/projected/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5-kube-api-access-hfwql\") pod \"community-operators-pxt6s\" (UID: \"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5\") " pod="openshift-marketplace/community-operators-pxt6s" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.144150 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca896d69-0651-48e1-a001-0b3c57e1ece4-catalog-content\") pod \"certified-operators-fm8lj\" (UID: \"ca896d69-0651-48e1-a001-0b3c57e1ece4\") " pod="openshift-marketplace/certified-operators-fm8lj" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.144219 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca896d69-0651-48e1-a001-0b3c57e1ece4-utilities\") pod \"certified-operators-fm8lj\" (UID: \"ca896d69-0651-48e1-a001-0b3c57e1ece4\") " pod="openshift-marketplace/certified-operators-fm8lj" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.144247 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zrjs\" (UniqueName: \"kubernetes.io/projected/ca896d69-0651-48e1-a001-0b3c57e1ece4-kube-api-access-5zrjs\") pod \"certified-operators-fm8lj\" (UID: \"ca896d69-0651-48e1-a001-0b3c57e1ece4\") " pod="openshift-marketplace/certified-operators-fm8lj" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.145003 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca896d69-0651-48e1-a001-0b3c57e1ece4-utilities\") pod \"certified-operators-fm8lj\" (UID: \"ca896d69-0651-48e1-a001-0b3c57e1ece4\") " pod="openshift-marketplace/certified-operators-fm8lj" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.145443 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca896d69-0651-48e1-a001-0b3c57e1ece4-catalog-content\") pod \"certified-operators-fm8lj\" (UID: \"ca896d69-0651-48e1-a001-0b3c57e1ece4\") " pod="openshift-marketplace/certified-operators-fm8lj" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.160679 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mxxf2"] Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.161847 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mxxf2" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.164759 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zrjs\" (UniqueName: \"kubernetes.io/projected/ca896d69-0651-48e1-a001-0b3c57e1ece4-kube-api-access-5zrjs\") pod \"certified-operators-fm8lj\" (UID: \"ca896d69-0651-48e1-a001-0b3c57e1ece4\") " pod="openshift-marketplace/certified-operators-fm8lj" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.174954 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mxxf2"] Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.230248 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.245200 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzx49\" (UniqueName: \"kubernetes.io/projected/48b25ecd-efe4-4363-975d-a17cb60d10a3-kube-api-access-gzx49\") pod \"community-operators-mxxf2\" (UID: \"48b25ecd-efe4-4363-975d-a17cb60d10a3\") " pod="openshift-marketplace/community-operators-mxxf2" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.245260 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48b25ecd-efe4-4363-975d-a17cb60d10a3-catalog-content\") pod \"community-operators-mxxf2\" (UID: \"48b25ecd-efe4-4363-975d-a17cb60d10a3\") " pod="openshift-marketplace/community-operators-mxxf2" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.245296 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48b25ecd-efe4-4363-975d-a17cb60d10a3-utilities\") pod \"community-operators-mxxf2\" (UID: \"48b25ecd-efe4-4363-975d-a17cb60d10a3\") " pod="openshift-marketplace/community-operators-mxxf2" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.282110 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fm8lj" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.346868 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzx49\" (UniqueName: \"kubernetes.io/projected/48b25ecd-efe4-4363-975d-a17cb60d10a3-kube-api-access-gzx49\") pod \"community-operators-mxxf2\" (UID: \"48b25ecd-efe4-4363-975d-a17cb60d10a3\") " pod="openshift-marketplace/community-operators-mxxf2" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.346922 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48b25ecd-efe4-4363-975d-a17cb60d10a3-catalog-content\") pod \"community-operators-mxxf2\" (UID: \"48b25ecd-efe4-4363-975d-a17cb60d10a3\") " pod="openshift-marketplace/community-operators-mxxf2" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.346957 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48b25ecd-efe4-4363-975d-a17cb60d10a3-utilities\") pod \"community-operators-mxxf2\" (UID: \"48b25ecd-efe4-4363-975d-a17cb60d10a3\") " pod="openshift-marketplace/community-operators-mxxf2" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.347463 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48b25ecd-efe4-4363-975d-a17cb60d10a3-utilities\") pod \"community-operators-mxxf2\" (UID: \"48b25ecd-efe4-4363-975d-a17cb60d10a3\") " pod="openshift-marketplace/community-operators-mxxf2" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.347435 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48b25ecd-efe4-4363-975d-a17cb60d10a3-catalog-content\") pod \"community-operators-mxxf2\" (UID: \"48b25ecd-efe4-4363-975d-a17cb60d10a3\") " pod="openshift-marketplace/community-operators-mxxf2" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.361877 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w85bt"] Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.363033 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w85bt" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.368263 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzx49\" (UniqueName: \"kubernetes.io/projected/48b25ecd-efe4-4363-975d-a17cb60d10a3-kube-api-access-gzx49\") pod \"community-operators-mxxf2\" (UID: \"48b25ecd-efe4-4363-975d-a17cb60d10a3\") " pod="openshift-marketplace/community-operators-mxxf2" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.409684 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w85bt"] Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.448332 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0eb57a13-0325-432a-a641-44903e611452-utilities\") pod \"certified-operators-w85bt\" (UID: \"0eb57a13-0325-432a-a641-44903e611452\") " pod="openshift-marketplace/certified-operators-w85bt" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.448406 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0eb57a13-0325-432a-a641-44903e611452-catalog-content\") pod \"certified-operators-w85bt\" (UID: \"0eb57a13-0325-432a-a641-44903e611452\") " pod="openshift-marketplace/certified-operators-w85bt" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.448530 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cmnm\" (UniqueName: \"kubernetes.io/projected/0eb57a13-0325-432a-a641-44903e611452-kube-api-access-4cmnm\") pod \"certified-operators-w85bt\" (UID: \"0eb57a13-0325-432a-a641-44903e611452\") " pod="openshift-marketplace/certified-operators-w85bt" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.484118 4810 patch_prober.go:28] interesting pod/router-default-5444994796-hsnbk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 14:36:30 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Dec 01 14:36:30 crc kubenswrapper[4810]: [+]process-running ok Dec 01 14:36:30 crc kubenswrapper[4810]: healthz check failed Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.484196 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hsnbk" podUID="3863c075-f600-4d7b-b452-1ffa4b5c80f0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.497970 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.512608 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" event={"ID":"17423ab4-d417-42d3-84f0-ca131f3dceeb","Type":"ContainerStarted","Data":"e534835f084b073ed6b16e659116b2cd6148aba630a28bee900c9f9f90bca3d8"} Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.514374 4810 generic.go:334] "Generic (PLEG): container finished" podID="484c6c08-536f-4374-971e-1fb4ef397b64" containerID="b3feacdd440d69f9f0233048aea0de0190a3da88ac5e9deaeb3af2daf12d95dc" exitCode=0 Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.514575 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld" event={"ID":"484c6c08-536f-4374-971e-1fb4ef397b64","Type":"ContainerDied","Data":"b3feacdd440d69f9f0233048aea0de0190a3da88ac5e9deaeb3af2daf12d95dc"} Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.534803 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-pdv9r" podStartSLOduration=10.534784988 podStartE2EDuration="10.534784988s" podCreationTimestamp="2025-12-01 14:36:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:30.533450762 +0000 UTC m=+156.296960365" watchObservedRunningTime="2025-12-01 14:36:30.534784988 +0000 UTC m=+156.298294591" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.549886 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0eb57a13-0325-432a-a641-44903e611452-catalog-content\") pod \"certified-operators-w85bt\" (UID: \"0eb57a13-0325-432a-a641-44903e611452\") " pod="openshift-marketplace/certified-operators-w85bt" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.549963 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cmnm\" (UniqueName: \"kubernetes.io/projected/0eb57a13-0325-432a-a641-44903e611452-kube-api-access-4cmnm\") pod \"certified-operators-w85bt\" (UID: \"0eb57a13-0325-432a-a641-44903e611452\") " pod="openshift-marketplace/certified-operators-w85bt" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.550013 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0eb57a13-0325-432a-a641-44903e611452-utilities\") pod \"certified-operators-w85bt\" (UID: \"0eb57a13-0325-432a-a641-44903e611452\") " pod="openshift-marketplace/certified-operators-w85bt" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.550683 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0eb57a13-0325-432a-a641-44903e611452-catalog-content\") pod \"certified-operators-w85bt\" (UID: \"0eb57a13-0325-432a-a641-44903e611452\") " pod="openshift-marketplace/certified-operators-w85bt" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.550761 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0eb57a13-0325-432a-a641-44903e611452-utilities\") pod \"certified-operators-w85bt\" (UID: \"0eb57a13-0325-432a-a641-44903e611452\") " pod="openshift-marketplace/certified-operators-w85bt" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.567932 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cmnm\" (UniqueName: \"kubernetes.io/projected/0eb57a13-0325-432a-a641-44903e611452-kube-api-access-4cmnm\") pod \"certified-operators-w85bt\" (UID: \"0eb57a13-0325-432a-a641-44903e611452\") " pod="openshift-marketplace/certified-operators-w85bt" Dec 01 14:36:30 crc kubenswrapper[4810]: I1201 14:36:30.691107 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w85bt" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.072304 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w85bt"] Dec 01 14:36:31 crc kubenswrapper[4810]: W1201 14:36:31.080574 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0eb57a13_0325_432a_a641_44903e611452.slice/crio-8c85a26918dfa33be67ec28f7eb3b0e6044fcb41f03a58ac6911b2c70278beba WatchSource:0}: Error finding container 8c85a26918dfa33be67ec28f7eb3b0e6044fcb41f03a58ac6911b2c70278beba: Status 404 returned error can't find the container with id 8c85a26918dfa33be67ec28f7eb3b0e6044fcb41f03a58ac6911b2c70278beba Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.121869 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.122116 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fm8lj"] Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.123931 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mxxf2" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.130145 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxt6s" Dec 01 14:36:31 crc kubenswrapper[4810]: W1201 14:36:31.138463 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca896d69_0651_48e1_a001_0b3c57e1ece4.slice/crio-a2ced1a9adde5165eb4fa97545fb79e841108b61cb643460e2c77fb256b9e779 WatchSource:0}: Error finding container a2ced1a9adde5165eb4fa97545fb79e841108b61cb643460e2c77fb256b9e779: Status 404 returned error can't find the container with id a2ced1a9adde5165eb4fa97545fb79e841108b61cb643460e2c77fb256b9e779 Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.184177 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qvm7d"] Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.407947 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.408729 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.412023 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.412273 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.418815 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.465080 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9a2da148-1ff1-4dae-aec0-c3b976b97987-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9a2da148-1ff1-4dae-aec0-c3b976b97987\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.465172 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9a2da148-1ff1-4dae-aec0-c3b976b97987-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9a2da148-1ff1-4dae-aec0-c3b976b97987\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.484094 4810 patch_prober.go:28] interesting pod/router-default-5444994796-hsnbk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 14:36:31 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Dec 01 14:36:31 crc kubenswrapper[4810]: [+]process-running ok Dec 01 14:36:31 crc kubenswrapper[4810]: healthz check failed Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.484144 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hsnbk" podUID="3863c075-f600-4d7b-b452-1ffa4b5c80f0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.522563 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" event={"ID":"5a67f6b6-1881-4807-b719-40ca184bd5dc","Type":"ContainerStarted","Data":"62e6f9c0e7d9389b8561ad5d1913e7d3cf03ebc018a89667997eff66b6b4ce90"} Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.523950 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fm8lj" event={"ID":"ca896d69-0651-48e1-a001-0b3c57e1ece4","Type":"ContainerStarted","Data":"a2ced1a9adde5165eb4fa97545fb79e841108b61cb643460e2c77fb256b9e779"} Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.525099 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w85bt" event={"ID":"0eb57a13-0325-432a-a641-44903e611452","Type":"ContainerStarted","Data":"8c85a26918dfa33be67ec28f7eb3b0e6044fcb41f03a58ac6911b2c70278beba"} Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.569561 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9a2da148-1ff1-4dae-aec0-c3b976b97987-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9a2da148-1ff1-4dae-aec0-c3b976b97987\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.569857 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9a2da148-1ff1-4dae-aec0-c3b976b97987-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9a2da148-1ff1-4dae-aec0-c3b976b97987\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.569990 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9a2da148-1ff1-4dae-aec0-c3b976b97987-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9a2da148-1ff1-4dae-aec0-c3b976b97987\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.586142 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pxt6s"] Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.588565 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mxxf2"] Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.597297 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9a2da148-1ff1-4dae-aec0-c3b976b97987-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9a2da148-1ff1-4dae-aec0-c3b976b97987\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 14:36:31 crc kubenswrapper[4810]: W1201 14:36:31.676130 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb18b0a4d_1ac2_4ae4_a60f_73db7ec71cb5.slice/crio-0f26490f9d7312629415a15a28fd8e4b4b4800c600bc73694c59fe0deff097b9 WatchSource:0}: Error finding container 0f26490f9d7312629415a15a28fd8e4b4b4800c600bc73694c59fe0deff097b9: Status 404 returned error can't find the container with id 0f26490f9d7312629415a15a28fd8e4b4b4800c600bc73694c59fe0deff097b9 Dec 01 14:36:31 crc kubenswrapper[4810]: W1201 14:36:31.678189 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48b25ecd_efe4_4363_975d_a17cb60d10a3.slice/crio-65b0d40e94b979c9f1db101c98618c7c3bababccb87ceabb8b8afc90ca5cce4d WatchSource:0}: Error finding container 65b0d40e94b979c9f1db101c98618c7c3bababccb87ceabb8b8afc90ca5cce4d: Status 404 returned error can't find the container with id 65b0d40e94b979c9f1db101c98618c7c3bababccb87ceabb8b8afc90ca5cce4d Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.730728 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.764931 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pcjqf"] Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.766312 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pcjqf" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.770044 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.776853 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.777063 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pcjqf"] Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.888413 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/484c6c08-536f-4374-971e-1fb4ef397b64-config-volume\") pod \"484c6c08-536f-4374-971e-1fb4ef397b64\" (UID: \"484c6c08-536f-4374-971e-1fb4ef397b64\") " Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.889119 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/484c6c08-536f-4374-971e-1fb4ef397b64-secret-volume\") pod \"484c6c08-536f-4374-971e-1fb4ef397b64\" (UID: \"484c6c08-536f-4374-971e-1fb4ef397b64\") " Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.889158 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2qvw\" (UniqueName: \"kubernetes.io/projected/484c6c08-536f-4374-971e-1fb4ef397b64-kube-api-access-m2qvw\") pod \"484c6c08-536f-4374-971e-1fb4ef397b64\" (UID: \"484c6c08-536f-4374-971e-1fb4ef397b64\") " Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.889423 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00d0bac3-a43f-469e-9335-df4013f7432e-catalog-content\") pod \"redhat-marketplace-pcjqf\" (UID: \"00d0bac3-a43f-469e-9335-df4013f7432e\") " pod="openshift-marketplace/redhat-marketplace-pcjqf" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.889458 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhhzd\" (UniqueName: \"kubernetes.io/projected/00d0bac3-a43f-469e-9335-df4013f7432e-kube-api-access-vhhzd\") pod \"redhat-marketplace-pcjqf\" (UID: \"00d0bac3-a43f-469e-9335-df4013f7432e\") " pod="openshift-marketplace/redhat-marketplace-pcjqf" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.889505 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00d0bac3-a43f-469e-9335-df4013f7432e-utilities\") pod \"redhat-marketplace-pcjqf\" (UID: \"00d0bac3-a43f-469e-9335-df4013f7432e\") " pod="openshift-marketplace/redhat-marketplace-pcjqf" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.893954 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/484c6c08-536f-4374-971e-1fb4ef397b64-config-volume" (OuterVolumeSpecName: "config-volume") pod "484c6c08-536f-4374-971e-1fb4ef397b64" (UID: "484c6c08-536f-4374-971e-1fb4ef397b64"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.900105 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/484c6c08-536f-4374-971e-1fb4ef397b64-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "484c6c08-536f-4374-971e-1fb4ef397b64" (UID: "484c6c08-536f-4374-971e-1fb4ef397b64"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.900254 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/484c6c08-536f-4374-971e-1fb4ef397b64-kube-api-access-m2qvw" (OuterVolumeSpecName: "kube-api-access-m2qvw") pod "484c6c08-536f-4374-971e-1fb4ef397b64" (UID: "484c6c08-536f-4374-971e-1fb4ef397b64"). InnerVolumeSpecName "kube-api-access-m2qvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.990849 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00d0bac3-a43f-469e-9335-df4013f7432e-catalog-content\") pod \"redhat-marketplace-pcjqf\" (UID: \"00d0bac3-a43f-469e-9335-df4013f7432e\") " pod="openshift-marketplace/redhat-marketplace-pcjqf" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.990952 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhhzd\" (UniqueName: \"kubernetes.io/projected/00d0bac3-a43f-469e-9335-df4013f7432e-kube-api-access-vhhzd\") pod \"redhat-marketplace-pcjqf\" (UID: \"00d0bac3-a43f-469e-9335-df4013f7432e\") " pod="openshift-marketplace/redhat-marketplace-pcjqf" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.991126 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00d0bac3-a43f-469e-9335-df4013f7432e-utilities\") pod \"redhat-marketplace-pcjqf\" (UID: \"00d0bac3-a43f-469e-9335-df4013f7432e\") " pod="openshift-marketplace/redhat-marketplace-pcjqf" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.991428 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/484c6c08-536f-4374-971e-1fb4ef397b64-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.991481 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2qvw\" (UniqueName: \"kubernetes.io/projected/484c6c08-536f-4374-971e-1fb4ef397b64-kube-api-access-m2qvw\") on node \"crc\" DevicePath \"\"" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.991497 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/484c6c08-536f-4374-971e-1fb4ef397b64-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.991437 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00d0bac3-a43f-469e-9335-df4013f7432e-catalog-content\") pod \"redhat-marketplace-pcjqf\" (UID: \"00d0bac3-a43f-469e-9335-df4013f7432e\") " pod="openshift-marketplace/redhat-marketplace-pcjqf" Dec 01 14:36:31 crc kubenswrapper[4810]: I1201 14:36:31.991698 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00d0bac3-a43f-469e-9335-df4013f7432e-utilities\") pod \"redhat-marketplace-pcjqf\" (UID: \"00d0bac3-a43f-469e-9335-df4013f7432e\") " pod="openshift-marketplace/redhat-marketplace-pcjqf" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.009585 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhhzd\" (UniqueName: \"kubernetes.io/projected/00d0bac3-a43f-469e-9335-df4013f7432e-kube-api-access-vhhzd\") pod \"redhat-marketplace-pcjqf\" (UID: \"00d0bac3-a43f-469e-9335-df4013f7432e\") " pod="openshift-marketplace/redhat-marketplace-pcjqf" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.111021 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pcjqf" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.161040 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jfmhr"] Dec 01 14:36:32 crc kubenswrapper[4810]: E1201 14:36:32.161253 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="484c6c08-536f-4374-971e-1fb4ef397b64" containerName="collect-profiles" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.161265 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="484c6c08-536f-4374-971e-1fb4ef397b64" containerName="collect-profiles" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.161342 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="484c6c08-536f-4374-971e-1fb4ef397b64" containerName="collect-profiles" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.162059 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jfmhr" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.169525 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.169581 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.171318 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jfmhr"] Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.176886 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.197341 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h54rq" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.203562 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 14:36:32 crc kubenswrapper[4810]: W1201 14:36:32.225686 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod9a2da148_1ff1_4dae_aec0_c3b976b97987.slice/crio-cbe1f949f64e62034792cbb2ab20d8cd45b44413eab96b3e64801a1d6793d1cb WatchSource:0}: Error finding container cbe1f949f64e62034792cbb2ab20d8cd45b44413eab96b3e64801a1d6793d1cb: Status 404 returned error can't find the container with id cbe1f949f64e62034792cbb2ab20d8cd45b44413eab96b3e64801a1d6793d1cb Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.296084 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/044c93cf-65a5-455b-af09-63e80ee90adc-utilities\") pod \"redhat-marketplace-jfmhr\" (UID: \"044c93cf-65a5-455b-af09-63e80ee90adc\") " pod="openshift-marketplace/redhat-marketplace-jfmhr" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.296715 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/044c93cf-65a5-455b-af09-63e80ee90adc-catalog-content\") pod \"redhat-marketplace-jfmhr\" (UID: \"044c93cf-65a5-455b-af09-63e80ee90adc\") " pod="openshift-marketplace/redhat-marketplace-jfmhr" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.296811 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8gd9\" (UniqueName: \"kubernetes.io/projected/044c93cf-65a5-455b-af09-63e80ee90adc-kube-api-access-m8gd9\") pod \"redhat-marketplace-jfmhr\" (UID: \"044c93cf-65a5-455b-af09-63e80ee90adc\") " pod="openshift-marketplace/redhat-marketplace-jfmhr" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.361295 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pcjqf"] Dec 01 14:36:32 crc kubenswrapper[4810]: W1201 14:36:32.367705 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00d0bac3_a43f_469e_9335_df4013f7432e.slice/crio-1757fb4e02abffa4095d30a9e9184015f1e40c344455b9f9f9b7946de8edce94 WatchSource:0}: Error finding container 1757fb4e02abffa4095d30a9e9184015f1e40c344455b9f9f9b7946de8edce94: Status 404 returned error can't find the container with id 1757fb4e02abffa4095d30a9e9184015f1e40c344455b9f9f9b7946de8edce94 Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.397876 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/044c93cf-65a5-455b-af09-63e80ee90adc-catalog-content\") pod \"redhat-marketplace-jfmhr\" (UID: \"044c93cf-65a5-455b-af09-63e80ee90adc\") " pod="openshift-marketplace/redhat-marketplace-jfmhr" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.397935 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8gd9\" (UniqueName: \"kubernetes.io/projected/044c93cf-65a5-455b-af09-63e80ee90adc-kube-api-access-m8gd9\") pod \"redhat-marketplace-jfmhr\" (UID: \"044c93cf-65a5-455b-af09-63e80ee90adc\") " pod="openshift-marketplace/redhat-marketplace-jfmhr" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.397956 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/044c93cf-65a5-455b-af09-63e80ee90adc-utilities\") pod \"redhat-marketplace-jfmhr\" (UID: \"044c93cf-65a5-455b-af09-63e80ee90adc\") " pod="openshift-marketplace/redhat-marketplace-jfmhr" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.398399 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/044c93cf-65a5-455b-af09-63e80ee90adc-utilities\") pod \"redhat-marketplace-jfmhr\" (UID: \"044c93cf-65a5-455b-af09-63e80ee90adc\") " pod="openshift-marketplace/redhat-marketplace-jfmhr" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.398666 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/044c93cf-65a5-455b-af09-63e80ee90adc-catalog-content\") pod \"redhat-marketplace-jfmhr\" (UID: \"044c93cf-65a5-455b-af09-63e80ee90adc\") " pod="openshift-marketplace/redhat-marketplace-jfmhr" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.419695 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8gd9\" (UniqueName: \"kubernetes.io/projected/044c93cf-65a5-455b-af09-63e80ee90adc-kube-api-access-m8gd9\") pod \"redhat-marketplace-jfmhr\" (UID: \"044c93cf-65a5-455b-af09-63e80ee90adc\") " pod="openshift-marketplace/redhat-marketplace-jfmhr" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.483999 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jfmhr" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.484054 4810 patch_prober.go:28] interesting pod/router-default-5444994796-hsnbk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 14:36:32 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Dec 01 14:36:32 crc kubenswrapper[4810]: [+]process-running ok Dec 01 14:36:32 crc kubenswrapper[4810]: healthz check failed Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.484099 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hsnbk" podUID="3863c075-f600-4d7b-b452-1ffa4b5c80f0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.535063 4810 generic.go:334] "Generic (PLEG): container finished" podID="ca896d69-0651-48e1-a001-0b3c57e1ece4" containerID="6342a41e762ac01621caf4778763386ce53ae43548907b267168386be609ba01" exitCode=0 Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.535131 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fm8lj" event={"ID":"ca896d69-0651-48e1-a001-0b3c57e1ece4","Type":"ContainerDied","Data":"6342a41e762ac01621caf4778763386ce53ae43548907b267168386be609ba01"} Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.536317 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcjqf" event={"ID":"00d0bac3-a43f-469e-9335-df4013f7432e","Type":"ContainerStarted","Data":"1757fb4e02abffa4095d30a9e9184015f1e40c344455b9f9f9b7946de8edce94"} Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.537679 4810 generic.go:334] "Generic (PLEG): container finished" podID="b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5" containerID="3a1d49990f7824e2a3f3cea80706c3b13951d9e30d24449a230a69cd92d6d598" exitCode=0 Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.538726 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxt6s" event={"ID":"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5","Type":"ContainerDied","Data":"3a1d49990f7824e2a3f3cea80706c3b13951d9e30d24449a230a69cd92d6d598"} Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.538820 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxt6s" event={"ID":"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5","Type":"ContainerStarted","Data":"0f26490f9d7312629415a15a28fd8e4b4b4800c600bc73694c59fe0deff097b9"} Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.540495 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.547727 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld" event={"ID":"484c6c08-536f-4374-971e-1fb4ef397b64","Type":"ContainerDied","Data":"f70ec0e776cf2902bb7c7aeed924329467b17681c50f59716b54503e660cc7c1"} Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.547769 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f70ec0e776cf2902bb7c7aeed924329467b17681c50f59716b54503e660cc7c1" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.547830 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.559010 4810 generic.go:334] "Generic (PLEG): container finished" podID="0eb57a13-0325-432a-a641-44903e611452" containerID="2865f4bc975be1b299eb1d90c2f2c414046274442d695c9cbf1d0ae7c2c55bce" exitCode=0 Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.559741 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w85bt" event={"ID":"0eb57a13-0325-432a-a641-44903e611452","Type":"ContainerDied","Data":"2865f4bc975be1b299eb1d90c2f2c414046274442d695c9cbf1d0ae7c2c55bce"} Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.562962 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.566845 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.587935 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.624059 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9a2da148-1ff1-4dae-aec0-c3b976b97987","Type":"ContainerStarted","Data":"cbe1f949f64e62034792cbb2ab20d8cd45b44413eab96b3e64801a1d6793d1cb"} Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.651225 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" event={"ID":"5a67f6b6-1881-4807-b719-40ca184bd5dc","Type":"ContainerStarted","Data":"9b13f77523af72c9131cb10ef312970aba06063f1cefd652c3ef0ec5676afbda"} Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.651513 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.656836 4810 generic.go:334] "Generic (PLEG): container finished" podID="48b25ecd-efe4-4363-975d-a17cb60d10a3" containerID="666bfc240c437d01adb6528b64d139a4a2158c5e4d116636984aab3415beb93c" exitCode=0 Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.657461 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mxxf2" event={"ID":"48b25ecd-efe4-4363-975d-a17cb60d10a3","Type":"ContainerDied","Data":"666bfc240c437d01adb6528b64d139a4a2158c5e4d116636984aab3415beb93c"} Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.657586 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mxxf2" event={"ID":"48b25ecd-efe4-4363-975d-a17cb60d10a3","Type":"ContainerStarted","Data":"65b0d40e94b979c9f1db101c98618c7c3bababccb87ceabb8b8afc90ca5cce4d"} Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.666180 4810 patch_prober.go:28] interesting pod/downloads-7954f5f757-rf6tj container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.666223 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-rf6tj" podUID="88f6bb4e-362c-425c-9f67-81f76ed05e42" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.666301 4810 patch_prober.go:28] interesting pod/downloads-7954f5f757-rf6tj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.666324 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-rf6tj" podUID="88f6bb4e-362c-425c-9f67-81f76ed05e42" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.667729 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-pnv2c" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.676288 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" podStartSLOduration=134.676270506 podStartE2EDuration="2m14.676270506s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:32.672661879 +0000 UTC m=+158.436171482" watchObservedRunningTime="2025-12-01 14:36:32.676270506 +0000 UTC m=+158.439780109" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.835871 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jfmhr"] Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.842031 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.842376 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:32 crc kubenswrapper[4810]: W1201 14:36:32.842939 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod044c93cf_65a5_455b_af09_63e80ee90adc.slice/crio-f9cf065d5741a41448de2e8fbef83d4f974ab885e6b7efcab06e3ac398bbc30f WatchSource:0}: Error finding container f9cf065d5741a41448de2e8fbef83d4f974ab885e6b7efcab06e3ac398bbc30f: Status 404 returned error can't find the container with id f9cf065d5741a41448de2e8fbef83d4f974ab885e6b7efcab06e3ac398bbc30f Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.844759 4810 patch_prober.go:28] interesting pod/console-f9d7485db-8tcbx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.34:8443/health\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.844918 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-8tcbx" podUID="f65c3828-a372-4660-97dd-e01d6efcb78c" containerName="console" probeResult="failure" output="Get \"https://10.217.0.34:8443/health\": dial tcp 10.217.0.34:8443: connect: connection refused" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.964781 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xnb48"] Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.966719 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xnb48" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.970018 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.973629 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.973682 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:36:32 crc kubenswrapper[4810]: I1201 14:36:32.979538 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xnb48"] Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.010159 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5g7h\" (UniqueName: \"kubernetes.io/projected/93bf2297-d7e7-48f2-8a96-6c33d5b3754b-kube-api-access-q5g7h\") pod \"redhat-operators-xnb48\" (UID: \"93bf2297-d7e7-48f2-8a96-6c33d5b3754b\") " pod="openshift-marketplace/redhat-operators-xnb48" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.010221 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93bf2297-d7e7-48f2-8a96-6c33d5b3754b-utilities\") pod \"redhat-operators-xnb48\" (UID: \"93bf2297-d7e7-48f2-8a96-6c33d5b3754b\") " pod="openshift-marketplace/redhat-operators-xnb48" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.010299 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93bf2297-d7e7-48f2-8a96-6c33d5b3754b-catalog-content\") pod \"redhat-operators-xnb48\" (UID: \"93bf2297-d7e7-48f2-8a96-6c33d5b3754b\") " pod="openshift-marketplace/redhat-operators-xnb48" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.111297 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5g7h\" (UniqueName: \"kubernetes.io/projected/93bf2297-d7e7-48f2-8a96-6c33d5b3754b-kube-api-access-q5g7h\") pod \"redhat-operators-xnb48\" (UID: \"93bf2297-d7e7-48f2-8a96-6c33d5b3754b\") " pod="openshift-marketplace/redhat-operators-xnb48" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.111363 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93bf2297-d7e7-48f2-8a96-6c33d5b3754b-utilities\") pod \"redhat-operators-xnb48\" (UID: \"93bf2297-d7e7-48f2-8a96-6c33d5b3754b\") " pod="openshift-marketplace/redhat-operators-xnb48" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.111400 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93bf2297-d7e7-48f2-8a96-6c33d5b3754b-catalog-content\") pod \"redhat-operators-xnb48\" (UID: \"93bf2297-d7e7-48f2-8a96-6c33d5b3754b\") " pod="openshift-marketplace/redhat-operators-xnb48" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.111881 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93bf2297-d7e7-48f2-8a96-6c33d5b3754b-catalog-content\") pod \"redhat-operators-xnb48\" (UID: \"93bf2297-d7e7-48f2-8a96-6c33d5b3754b\") " pod="openshift-marketplace/redhat-operators-xnb48" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.112059 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93bf2297-d7e7-48f2-8a96-6c33d5b3754b-utilities\") pod \"redhat-operators-xnb48\" (UID: \"93bf2297-d7e7-48f2-8a96-6c33d5b3754b\") " pod="openshift-marketplace/redhat-operators-xnb48" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.137462 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5g7h\" (UniqueName: \"kubernetes.io/projected/93bf2297-d7e7-48f2-8a96-6c33d5b3754b-kube-api-access-q5g7h\") pod \"redhat-operators-xnb48\" (UID: \"93bf2297-d7e7-48f2-8a96-6c33d5b3754b\") " pod="openshift-marketplace/redhat-operators-xnb48" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.358956 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xnb48" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.362281 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5xbfx"] Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.363579 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5xbfx" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.370047 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5xbfx"] Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.416005 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efdc70a7-17ce-46e8-b0d0-b3f917370e8b-utilities\") pod \"redhat-operators-5xbfx\" (UID: \"efdc70a7-17ce-46e8-b0d0-b3f917370e8b\") " pod="openshift-marketplace/redhat-operators-5xbfx" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.416286 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85t58\" (UniqueName: \"kubernetes.io/projected/efdc70a7-17ce-46e8-b0d0-b3f917370e8b-kube-api-access-85t58\") pod \"redhat-operators-5xbfx\" (UID: \"efdc70a7-17ce-46e8-b0d0-b3f917370e8b\") " pod="openshift-marketplace/redhat-operators-5xbfx" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.416441 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efdc70a7-17ce-46e8-b0d0-b3f917370e8b-catalog-content\") pod \"redhat-operators-5xbfx\" (UID: \"efdc70a7-17ce-46e8-b0d0-b3f917370e8b\") " pod="openshift-marketplace/redhat-operators-5xbfx" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.480828 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.484624 4810 patch_prober.go:28] interesting pod/router-default-5444994796-hsnbk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 14:36:33 crc kubenswrapper[4810]: [-]has-synced failed: reason withheld Dec 01 14:36:33 crc kubenswrapper[4810]: [+]process-running ok Dec 01 14:36:33 crc kubenswrapper[4810]: healthz check failed Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.484881 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hsnbk" podUID="3863c075-f600-4d7b-b452-1ffa4b5c80f0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.517784 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efdc70a7-17ce-46e8-b0d0-b3f917370e8b-catalog-content\") pod \"redhat-operators-5xbfx\" (UID: \"efdc70a7-17ce-46e8-b0d0-b3f917370e8b\") " pod="openshift-marketplace/redhat-operators-5xbfx" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.517943 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efdc70a7-17ce-46e8-b0d0-b3f917370e8b-utilities\") pod \"redhat-operators-5xbfx\" (UID: \"efdc70a7-17ce-46e8-b0d0-b3f917370e8b\") " pod="openshift-marketplace/redhat-operators-5xbfx" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.517966 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85t58\" (UniqueName: \"kubernetes.io/projected/efdc70a7-17ce-46e8-b0d0-b3f917370e8b-kube-api-access-85t58\") pod \"redhat-operators-5xbfx\" (UID: \"efdc70a7-17ce-46e8-b0d0-b3f917370e8b\") " pod="openshift-marketplace/redhat-operators-5xbfx" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.519532 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efdc70a7-17ce-46e8-b0d0-b3f917370e8b-catalog-content\") pod \"redhat-operators-5xbfx\" (UID: \"efdc70a7-17ce-46e8-b0d0-b3f917370e8b\") " pod="openshift-marketplace/redhat-operators-5xbfx" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.519558 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efdc70a7-17ce-46e8-b0d0-b3f917370e8b-utilities\") pod \"redhat-operators-5xbfx\" (UID: \"efdc70a7-17ce-46e8-b0d0-b3f917370e8b\") " pod="openshift-marketplace/redhat-operators-5xbfx" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.533123 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xnb48"] Dec 01 14:36:33 crc kubenswrapper[4810]: W1201 14:36:33.542934 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93bf2297_d7e7_48f2_8a96_6c33d5b3754b.slice/crio-35b2f9a2a3b39f45955f3c06072ca5b454f0a2be191a9f35a141f9035689a05b WatchSource:0}: Error finding container 35b2f9a2a3b39f45955f3c06072ca5b454f0a2be191a9f35a141f9035689a05b: Status 404 returned error can't find the container with id 35b2f9a2a3b39f45955f3c06072ca5b454f0a2be191a9f35a141f9035689a05b Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.545067 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85t58\" (UniqueName: \"kubernetes.io/projected/efdc70a7-17ce-46e8-b0d0-b3f917370e8b-kube-api-access-85t58\") pod \"redhat-operators-5xbfx\" (UID: \"efdc70a7-17ce-46e8-b0d0-b3f917370e8b\") " pod="openshift-marketplace/redhat-operators-5xbfx" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.664061 4810 generic.go:334] "Generic (PLEG): container finished" podID="00d0bac3-a43f-469e-9335-df4013f7432e" containerID="8a300e31e64fde615c50cfaa13a7675fd48dbd82ed783bf4c06ac2644aff1857" exitCode=0 Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.664124 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcjqf" event={"ID":"00d0bac3-a43f-469e-9335-df4013f7432e","Type":"ContainerDied","Data":"8a300e31e64fde615c50cfaa13a7675fd48dbd82ed783bf4c06ac2644aff1857"} Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.667729 4810 generic.go:334] "Generic (PLEG): container finished" podID="044c93cf-65a5-455b-af09-63e80ee90adc" containerID="61680d97b0ea45f2ec6a4556a79cfa33352b48beb901da6878b4d3774b0cf3fa" exitCode=0 Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.667870 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jfmhr" event={"ID":"044c93cf-65a5-455b-af09-63e80ee90adc","Type":"ContainerDied","Data":"61680d97b0ea45f2ec6a4556a79cfa33352b48beb901da6878b4d3774b0cf3fa"} Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.667893 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jfmhr" event={"ID":"044c93cf-65a5-455b-af09-63e80ee90adc","Type":"ContainerStarted","Data":"f9cf065d5741a41448de2e8fbef83d4f974ab885e6b7efcab06e3ac398bbc30f"} Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.670734 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xnb48" event={"ID":"93bf2297-d7e7-48f2-8a96-6c33d5b3754b","Type":"ContainerStarted","Data":"35b2f9a2a3b39f45955f3c06072ca5b454f0a2be191a9f35a141f9035689a05b"} Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.672136 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9a2da148-1ff1-4dae-aec0-c3b976b97987","Type":"ContainerStarted","Data":"8637762165cf2f36eeb9beb31cca01d7b95b96cf98e1e39aa3004a7e4416b80d"} Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.683791 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rvtjn" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.695513 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.695495461 podStartE2EDuration="2.695495461s" podCreationTimestamp="2025-12-01 14:36:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:33.693535958 +0000 UTC m=+159.457045601" watchObservedRunningTime="2025-12-01 14:36:33.695495461 +0000 UTC m=+159.459005064" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.708446 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5xbfx" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.882272 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.882970 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.888625 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.888896 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.891537 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.928389 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c7c266a-f21d-4768-92b8-09eb968405cd-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7c7c266a-f21d-4768-92b8-09eb968405cd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 14:36:33 crc kubenswrapper[4810]: I1201 14:36:33.928847 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c7c266a-f21d-4768-92b8-09eb968405cd-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7c7c266a-f21d-4768-92b8-09eb968405cd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 14:36:34 crc kubenswrapper[4810]: I1201 14:36:34.029963 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c7c266a-f21d-4768-92b8-09eb968405cd-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7c7c266a-f21d-4768-92b8-09eb968405cd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 14:36:34 crc kubenswrapper[4810]: I1201 14:36:34.030295 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c7c266a-f21d-4768-92b8-09eb968405cd-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7c7c266a-f21d-4768-92b8-09eb968405cd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 14:36:34 crc kubenswrapper[4810]: I1201 14:36:34.030101 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c7c266a-f21d-4768-92b8-09eb968405cd-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7c7c266a-f21d-4768-92b8-09eb968405cd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 14:36:34 crc kubenswrapper[4810]: I1201 14:36:34.046387 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c7c266a-f21d-4768-92b8-09eb968405cd-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7c7c266a-f21d-4768-92b8-09eb968405cd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 14:36:34 crc kubenswrapper[4810]: I1201 14:36:34.202418 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 14:36:34 crc kubenswrapper[4810]: I1201 14:36:34.515208 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:34 crc kubenswrapper[4810]: I1201 14:36:34.526595 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-hsnbk" Dec 01 14:36:34 crc kubenswrapper[4810]: I1201 14:36:34.561240 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5xbfx"] Dec 01 14:36:34 crc kubenswrapper[4810]: I1201 14:36:34.615505 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 14:36:34 crc kubenswrapper[4810]: W1201 14:36:34.645696 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7c7c266a_f21d_4768_92b8_09eb968405cd.slice/crio-5b2e196823a9c28234fc127bcec12828c2b95d46cd99ae99d03710bdf4a8458c WatchSource:0}: Error finding container 5b2e196823a9c28234fc127bcec12828c2b95d46cd99ae99d03710bdf4a8458c: Status 404 returned error can't find the container with id 5b2e196823a9c28234fc127bcec12828c2b95d46cd99ae99d03710bdf4a8458c Dec 01 14:36:34 crc kubenswrapper[4810]: I1201 14:36:34.683803 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5xbfx" event={"ID":"efdc70a7-17ce-46e8-b0d0-b3f917370e8b","Type":"ContainerStarted","Data":"35f43ae5ffc69861107929092cf339c31cf6b17eb803ecd95aef518046156bce"} Dec 01 14:36:34 crc kubenswrapper[4810]: I1201 14:36:34.692797 4810 generic.go:334] "Generic (PLEG): container finished" podID="93bf2297-d7e7-48f2-8a96-6c33d5b3754b" containerID="40544c213be6f99ca606c89d40f4a275a5aa86846696ceb5330ccb225c03c328" exitCode=0 Dec 01 14:36:34 crc kubenswrapper[4810]: I1201 14:36:34.692870 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xnb48" event={"ID":"93bf2297-d7e7-48f2-8a96-6c33d5b3754b","Type":"ContainerDied","Data":"40544c213be6f99ca606c89d40f4a275a5aa86846696ceb5330ccb225c03c328"} Dec 01 14:36:34 crc kubenswrapper[4810]: I1201 14:36:34.698426 4810 generic.go:334] "Generic (PLEG): container finished" podID="9a2da148-1ff1-4dae-aec0-c3b976b97987" containerID="8637762165cf2f36eeb9beb31cca01d7b95b96cf98e1e39aa3004a7e4416b80d" exitCode=0 Dec 01 14:36:34 crc kubenswrapper[4810]: I1201 14:36:34.698502 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9a2da148-1ff1-4dae-aec0-c3b976b97987","Type":"ContainerDied","Data":"8637762165cf2f36eeb9beb31cca01d7b95b96cf98e1e39aa3004a7e4416b80d"} Dec 01 14:36:34 crc kubenswrapper[4810]: I1201 14:36:34.719901 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7c7c266a-f21d-4768-92b8-09eb968405cd","Type":"ContainerStarted","Data":"5b2e196823a9c28234fc127bcec12828c2b95d46cd99ae99d03710bdf4a8458c"} Dec 01 14:36:35 crc kubenswrapper[4810]: I1201 14:36:35.735530 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7c7c266a-f21d-4768-92b8-09eb968405cd","Type":"ContainerStarted","Data":"337081cf1a375287013afe35c58cb17004b0369b52e34ff85ad02f92ac7865a4"} Dec 01 14:36:35 crc kubenswrapper[4810]: I1201 14:36:35.779660 4810 generic.go:334] "Generic (PLEG): container finished" podID="efdc70a7-17ce-46e8-b0d0-b3f917370e8b" containerID="63c9af014bb4942859b29dda99c1dc2e2af9236e18b1859c097d9ffad6fb993a" exitCode=0 Dec 01 14:36:35 crc kubenswrapper[4810]: I1201 14:36:35.782242 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5xbfx" event={"ID":"efdc70a7-17ce-46e8-b0d0-b3f917370e8b","Type":"ContainerDied","Data":"63c9af014bb4942859b29dda99c1dc2e2af9236e18b1859c097d9ffad6fb993a"} Dec 01 14:36:35 crc kubenswrapper[4810]: I1201 14:36:35.785389 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.785358198 podStartE2EDuration="2.785358198s" podCreationTimestamp="2025-12-01 14:36:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:35.758077944 +0000 UTC m=+161.521587547" watchObservedRunningTime="2025-12-01 14:36:35.785358198 +0000 UTC m=+161.548867841" Dec 01 14:36:36 crc kubenswrapper[4810]: I1201 14:36:36.196290 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 14:36:36 crc kubenswrapper[4810]: I1201 14:36:36.309775 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9a2da148-1ff1-4dae-aec0-c3b976b97987-kubelet-dir\") pod \"9a2da148-1ff1-4dae-aec0-c3b976b97987\" (UID: \"9a2da148-1ff1-4dae-aec0-c3b976b97987\") " Dec 01 14:36:36 crc kubenswrapper[4810]: I1201 14:36:36.309917 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9a2da148-1ff1-4dae-aec0-c3b976b97987-kube-api-access\") pod \"9a2da148-1ff1-4dae-aec0-c3b976b97987\" (UID: \"9a2da148-1ff1-4dae-aec0-c3b976b97987\") " Dec 01 14:36:36 crc kubenswrapper[4810]: I1201 14:36:36.309945 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a2da148-1ff1-4dae-aec0-c3b976b97987-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9a2da148-1ff1-4dae-aec0-c3b976b97987" (UID: "9a2da148-1ff1-4dae-aec0-c3b976b97987"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:36:36 crc kubenswrapper[4810]: I1201 14:36:36.310171 4810 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9a2da148-1ff1-4dae-aec0-c3b976b97987-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 14:36:36 crc kubenswrapper[4810]: I1201 14:36:36.315570 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a2da148-1ff1-4dae-aec0-c3b976b97987-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9a2da148-1ff1-4dae-aec0-c3b976b97987" (UID: "9a2da148-1ff1-4dae-aec0-c3b976b97987"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:36:36 crc kubenswrapper[4810]: I1201 14:36:36.410962 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9a2da148-1ff1-4dae-aec0-c3b976b97987-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 14:36:36 crc kubenswrapper[4810]: I1201 14:36:36.795667 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9a2da148-1ff1-4dae-aec0-c3b976b97987","Type":"ContainerDied","Data":"cbe1f949f64e62034792cbb2ab20d8cd45b44413eab96b3e64801a1d6793d1cb"} Dec 01 14:36:36 crc kubenswrapper[4810]: I1201 14:36:36.795705 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbe1f949f64e62034792cbb2ab20d8cd45b44413eab96b3e64801a1d6793d1cb" Dec 01 14:36:36 crc kubenswrapper[4810]: I1201 14:36:36.795770 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 14:36:36 crc kubenswrapper[4810]: I1201 14:36:36.800090 4810 generic.go:334] "Generic (PLEG): container finished" podID="7c7c266a-f21d-4768-92b8-09eb968405cd" containerID="337081cf1a375287013afe35c58cb17004b0369b52e34ff85ad02f92ac7865a4" exitCode=0 Dec 01 14:36:36 crc kubenswrapper[4810]: I1201 14:36:36.800131 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7c7c266a-f21d-4768-92b8-09eb968405cd","Type":"ContainerDied","Data":"337081cf1a375287013afe35c58cb17004b0369b52e34ff85ad02f92ac7865a4"} Dec 01 14:36:37 crc kubenswrapper[4810]: I1201 14:36:37.651727 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:36:38 crc kubenswrapper[4810]: I1201 14:36:38.328340 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-dt9dw" Dec 01 14:36:38 crc kubenswrapper[4810]: I1201 14:36:38.527249 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 14:36:38 crc kubenswrapper[4810]: I1201 14:36:38.718718 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c7c266a-f21d-4768-92b8-09eb968405cd-kubelet-dir\") pod \"7c7c266a-f21d-4768-92b8-09eb968405cd\" (UID: \"7c7c266a-f21d-4768-92b8-09eb968405cd\") " Dec 01 14:36:38 crc kubenswrapper[4810]: I1201 14:36:38.718910 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c7c266a-f21d-4768-92b8-09eb968405cd-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7c7c266a-f21d-4768-92b8-09eb968405cd" (UID: "7c7c266a-f21d-4768-92b8-09eb968405cd"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:36:38 crc kubenswrapper[4810]: I1201 14:36:38.718923 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c7c266a-f21d-4768-92b8-09eb968405cd-kube-api-access\") pod \"7c7c266a-f21d-4768-92b8-09eb968405cd\" (UID: \"7c7c266a-f21d-4768-92b8-09eb968405cd\") " Dec 01 14:36:38 crc kubenswrapper[4810]: I1201 14:36:38.719368 4810 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c7c266a-f21d-4768-92b8-09eb968405cd-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 14:36:38 crc kubenswrapper[4810]: I1201 14:36:38.731395 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c7c266a-f21d-4768-92b8-09eb968405cd-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7c7c266a-f21d-4768-92b8-09eb968405cd" (UID: "7c7c266a-f21d-4768-92b8-09eb968405cd"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:36:38 crc kubenswrapper[4810]: I1201 14:36:38.835816 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c7c266a-f21d-4768-92b8-09eb968405cd-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 14:36:38 crc kubenswrapper[4810]: I1201 14:36:38.872632 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7c7c266a-f21d-4768-92b8-09eb968405cd","Type":"ContainerDied","Data":"5b2e196823a9c28234fc127bcec12828c2b95d46cd99ae99d03710bdf4a8458c"} Dec 01 14:36:38 crc kubenswrapper[4810]: I1201 14:36:38.872667 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b2e196823a9c28234fc127bcec12828c2b95d46cd99ae99d03710bdf4a8458c" Dec 01 14:36:38 crc kubenswrapper[4810]: I1201 14:36:38.872725 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 14:36:40 crc kubenswrapper[4810]: I1201 14:36:40.894508 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs\") pod \"network-metrics-daemon-jlwwg\" (UID: \"d379bcb2-a35d-470f-894c-8629d5023df8\") " pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:36:40 crc kubenswrapper[4810]: I1201 14:36:40.908362 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d379bcb2-a35d-470f-894c-8629d5023df8-metrics-certs\") pod \"network-metrics-daemon-jlwwg\" (UID: \"d379bcb2-a35d-470f-894c-8629d5023df8\") " pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:36:41 crc kubenswrapper[4810]: I1201 14:36:41.145372 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jlwwg" Dec 01 14:36:41 crc kubenswrapper[4810]: I1201 14:36:41.475837 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-jlwwg"] Dec 01 14:36:41 crc kubenswrapper[4810]: I1201 14:36:41.906771 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jlwwg" event={"ID":"d379bcb2-a35d-470f-894c-8629d5023df8","Type":"ContainerStarted","Data":"88b92166744155f0df552ebaf42b6934eb82bf9d548958602c146ae007c15d55"} Dec 01 14:36:42 crc kubenswrapper[4810]: I1201 14:36:42.697104 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-rf6tj" Dec 01 14:36:42 crc kubenswrapper[4810]: I1201 14:36:42.928704 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jlwwg" event={"ID":"d379bcb2-a35d-470f-894c-8629d5023df8","Type":"ContainerStarted","Data":"0a05a8df18822b6b6686e37154ac6f4b52848da417a8846cd4aaf32063d9d70b"} Dec 01 14:36:42 crc kubenswrapper[4810]: I1201 14:36:42.946403 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:42 crc kubenswrapper[4810]: I1201 14:36:42.950348 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:36:43 crc kubenswrapper[4810]: I1201 14:36:43.939633 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jlwwg" event={"ID":"d379bcb2-a35d-470f-894c-8629d5023df8","Type":"ContainerStarted","Data":"e5158cabd8c71d214ccc59d6581c38f8b8da07f563b624e955d1f6252d3b1236"} Dec 01 14:36:43 crc kubenswrapper[4810]: I1201 14:36:43.960158 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-jlwwg" podStartSLOduration=145.960136298 podStartE2EDuration="2m25.960136298s" podCreationTimestamp="2025-12-01 14:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:36:43.960091477 +0000 UTC m=+169.723601100" watchObservedRunningTime="2025-12-01 14:36:43.960136298 +0000 UTC m=+169.723645901" Dec 01 14:36:50 crc kubenswrapper[4810]: I1201 14:36:50.246057 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:37:01 crc kubenswrapper[4810]: I1201 14:37:01.381453 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 14:37:02 crc kubenswrapper[4810]: I1201 14:37:02.972117 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:37:02 crc kubenswrapper[4810]: I1201 14:37:02.972459 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:37:03 crc kubenswrapper[4810]: I1201 14:37:03.261498 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zkr9c" Dec 01 14:37:08 crc kubenswrapper[4810]: I1201 14:37:08.093789 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 14:37:08 crc kubenswrapper[4810]: E1201 14:37:08.094683 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c7c266a-f21d-4768-92b8-09eb968405cd" containerName="pruner" Dec 01 14:37:08 crc kubenswrapper[4810]: I1201 14:37:08.094699 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c7c266a-f21d-4768-92b8-09eb968405cd" containerName="pruner" Dec 01 14:37:08 crc kubenswrapper[4810]: E1201 14:37:08.094712 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a2da148-1ff1-4dae-aec0-c3b976b97987" containerName="pruner" Dec 01 14:37:08 crc kubenswrapper[4810]: I1201 14:37:08.094720 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a2da148-1ff1-4dae-aec0-c3b976b97987" containerName="pruner" Dec 01 14:37:08 crc kubenswrapper[4810]: I1201 14:37:08.094872 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a2da148-1ff1-4dae-aec0-c3b976b97987" containerName="pruner" Dec 01 14:37:08 crc kubenswrapper[4810]: I1201 14:37:08.094887 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c7c266a-f21d-4768-92b8-09eb968405cd" containerName="pruner" Dec 01 14:37:08 crc kubenswrapper[4810]: I1201 14:37:08.095417 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 14:37:08 crc kubenswrapper[4810]: I1201 14:37:08.103651 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 14:37:08 crc kubenswrapper[4810]: I1201 14:37:08.103752 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 14:37:08 crc kubenswrapper[4810]: I1201 14:37:08.106612 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 14:37:08 crc kubenswrapper[4810]: I1201 14:37:08.207051 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 14:37:08 crc kubenswrapper[4810]: I1201 14:37:08.207322 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 14:37:08 crc kubenswrapper[4810]: I1201 14:37:08.308213 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 14:37:08 crc kubenswrapper[4810]: I1201 14:37:08.308267 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 14:37:08 crc kubenswrapper[4810]: I1201 14:37:08.308393 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 14:37:08 crc kubenswrapper[4810]: I1201 14:37:08.344387 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 14:37:08 crc kubenswrapper[4810]: I1201 14:37:08.432397 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 14:37:08 crc kubenswrapper[4810]: E1201 14:37:08.731317 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 01 14:37:08 crc kubenswrapper[4810]: E1201 14:37:08.731768 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gzx49,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-mxxf2_openshift-marketplace(48b25ecd-efe4-4363-975d-a17cb60d10a3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 14:37:08 crc kubenswrapper[4810]: E1201 14:37:08.733198 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-mxxf2" podUID="48b25ecd-efe4-4363-975d-a17cb60d10a3" Dec 01 14:37:13 crc kubenswrapper[4810]: I1201 14:37:13.082237 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 14:37:13 crc kubenswrapper[4810]: I1201 14:37:13.086876 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 14:37:13 crc kubenswrapper[4810]: I1201 14:37:13.088115 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 14:37:13 crc kubenswrapper[4810]: I1201 14:37:13.277323 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/24394371-0f74-48a7-97c3-315489b78fe9-kube-api-access\") pod \"installer-9-crc\" (UID: \"24394371-0f74-48a7-97c3-315489b78fe9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 14:37:13 crc kubenswrapper[4810]: I1201 14:37:13.277405 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/24394371-0f74-48a7-97c3-315489b78fe9-var-lock\") pod \"installer-9-crc\" (UID: \"24394371-0f74-48a7-97c3-315489b78fe9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 14:37:13 crc kubenswrapper[4810]: I1201 14:37:13.277504 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/24394371-0f74-48a7-97c3-315489b78fe9-kubelet-dir\") pod \"installer-9-crc\" (UID: \"24394371-0f74-48a7-97c3-315489b78fe9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 14:37:13 crc kubenswrapper[4810]: I1201 14:37:13.378653 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/24394371-0f74-48a7-97c3-315489b78fe9-kubelet-dir\") pod \"installer-9-crc\" (UID: \"24394371-0f74-48a7-97c3-315489b78fe9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 14:37:13 crc kubenswrapper[4810]: I1201 14:37:13.378759 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/24394371-0f74-48a7-97c3-315489b78fe9-kube-api-access\") pod \"installer-9-crc\" (UID: \"24394371-0f74-48a7-97c3-315489b78fe9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 14:37:13 crc kubenswrapper[4810]: I1201 14:37:13.378783 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/24394371-0f74-48a7-97c3-315489b78fe9-kubelet-dir\") pod \"installer-9-crc\" (UID: \"24394371-0f74-48a7-97c3-315489b78fe9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 14:37:13 crc kubenswrapper[4810]: I1201 14:37:13.378825 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/24394371-0f74-48a7-97c3-315489b78fe9-var-lock\") pod \"installer-9-crc\" (UID: \"24394371-0f74-48a7-97c3-315489b78fe9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 14:37:13 crc kubenswrapper[4810]: I1201 14:37:13.378794 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/24394371-0f74-48a7-97c3-315489b78fe9-var-lock\") pod \"installer-9-crc\" (UID: \"24394371-0f74-48a7-97c3-315489b78fe9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 14:37:13 crc kubenswrapper[4810]: I1201 14:37:13.401766 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/24394371-0f74-48a7-97c3-315489b78fe9-kube-api-access\") pod \"installer-9-crc\" (UID: \"24394371-0f74-48a7-97c3-315489b78fe9\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 14:37:13 crc kubenswrapper[4810]: I1201 14:37:13.404646 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 14:37:13 crc kubenswrapper[4810]: E1201 14:37:13.613407 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-mxxf2" podUID="48b25ecd-efe4-4363-975d-a17cb60d10a3" Dec 01 14:37:13 crc kubenswrapper[4810]: E1201 14:37:13.689230 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 01 14:37:13 crc kubenswrapper[4810]: E1201 14:37:13.689737 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q5g7h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-xnb48_openshift-marketplace(93bf2297-d7e7-48f2-8a96-6c33d5b3754b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 14:37:13 crc kubenswrapper[4810]: E1201 14:37:13.690921 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-xnb48" podUID="93bf2297-d7e7-48f2-8a96-6c33d5b3754b" Dec 01 14:37:14 crc kubenswrapper[4810]: E1201 14:37:14.984973 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-xnb48" podUID="93bf2297-d7e7-48f2-8a96-6c33d5b3754b" Dec 01 14:37:15 crc kubenswrapper[4810]: E1201 14:37:15.061700 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 01 14:37:15 crc kubenswrapper[4810]: E1201 14:37:15.061881 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5zrjs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-fm8lj_openshift-marketplace(ca896d69-0651-48e1-a001-0b3c57e1ece4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 14:37:15 crc kubenswrapper[4810]: E1201 14:37:15.063063 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-fm8lj" podUID="ca896d69-0651-48e1-a001-0b3c57e1ece4" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.054868 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-fm8lj" podUID="ca896d69-0651-48e1-a001-0b3c57e1ece4" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.145876 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.146292 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vhhzd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-pcjqf_openshift-marketplace(00d0bac3-a43f-469e-9335-df4013f7432e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.149836 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-pcjqf" podUID="00d0bac3-a43f-469e-9335-df4013f7432e" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.167220 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.167435 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hfwql,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-pxt6s_openshift-marketplace(b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.168719 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-pxt6s" podUID="b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.172552 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.172678 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-85t58,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-5xbfx_openshift-marketplace(efdc70a7-17ce-46e8-b0d0-b3f917370e8b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.174158 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-5xbfx" podUID="efdc70a7-17ce-46e8-b0d0-b3f917370e8b" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.204442 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.204613 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m8gd9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-jfmhr_openshift-marketplace(044c93cf-65a5-455b-af09-63e80ee90adc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.205359 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.205551 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4cmnm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-w85bt_openshift-marketplace(0eb57a13-0325-432a-a641-44903e611452): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.206464 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-jfmhr" podUID="044c93cf-65a5-455b-af09-63e80ee90adc" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.208190 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-w85bt" podUID="0eb57a13-0325-432a-a641-44903e611452" Dec 01 14:37:16 crc kubenswrapper[4810]: I1201 14:37:16.297492 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 14:37:16 crc kubenswrapper[4810]: W1201 14:37:16.305261 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda97aeb7e_87e4_4f40_8b28_d1a4da0d6ebc.slice/crio-e2a326868405faf0dd65e3c448619df7f96ea09fd1dc7e71b9df5af8dac08060 WatchSource:0}: Error finding container e2a326868405faf0dd65e3c448619df7f96ea09fd1dc7e71b9df5af8dac08060: Status 404 returned error can't find the container with id e2a326868405faf0dd65e3c448619df7f96ea09fd1dc7e71b9df5af8dac08060 Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.309911 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-pxt6s" podUID="b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.310003 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-w85bt" podUID="0eb57a13-0325-432a-a641-44903e611452" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.310100 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-pcjqf" podUID="00d0bac3-a43f-469e-9335-df4013f7432e" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.310162 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-5xbfx" podUID="efdc70a7-17ce-46e8-b0d0-b3f917370e8b" Dec 01 14:37:16 crc kubenswrapper[4810]: E1201 14:37:16.310447 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-jfmhr" podUID="044c93cf-65a5-455b-af09-63e80ee90adc" Dec 01 14:37:16 crc kubenswrapper[4810]: I1201 14:37:16.505002 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 14:37:17 crc kubenswrapper[4810]: I1201 14:37:17.314376 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"24394371-0f74-48a7-97c3-315489b78fe9","Type":"ContainerStarted","Data":"292f5aff760680f0d08040f0f159285e7e43a0cfcc8e392091d540866f3f1e0d"} Dec 01 14:37:17 crc kubenswrapper[4810]: I1201 14:37:17.315849 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc","Type":"ContainerStarted","Data":"bdf70bce0783653d7774214975c7c946b6b39828fffd790341e112c2b5448471"} Dec 01 14:37:17 crc kubenswrapper[4810]: I1201 14:37:17.315901 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc","Type":"ContainerStarted","Data":"e2a326868405faf0dd65e3c448619df7f96ea09fd1dc7e71b9df5af8dac08060"} Dec 01 14:37:18 crc kubenswrapper[4810]: I1201 14:37:18.321762 4810 generic.go:334] "Generic (PLEG): container finished" podID="a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc" containerID="bdf70bce0783653d7774214975c7c946b6b39828fffd790341e112c2b5448471" exitCode=0 Dec 01 14:37:18 crc kubenswrapper[4810]: I1201 14:37:18.321806 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc","Type":"ContainerDied","Data":"bdf70bce0783653d7774214975c7c946b6b39828fffd790341e112c2b5448471"} Dec 01 14:37:18 crc kubenswrapper[4810]: I1201 14:37:18.323541 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"24394371-0f74-48a7-97c3-315489b78fe9","Type":"ContainerStarted","Data":"c52e52fae75a809aa7e157420966e9451d0cca4470f9d9c742e0bce07e8a0f80"} Dec 01 14:37:18 crc kubenswrapper[4810]: I1201 14:37:18.351959 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=5.35193755 podStartE2EDuration="5.35193755s" podCreationTimestamp="2025-12-01 14:37:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:37:18.350823799 +0000 UTC m=+204.114333412" watchObservedRunningTime="2025-12-01 14:37:18.35193755 +0000 UTC m=+204.115447163" Dec 01 14:37:19 crc kubenswrapper[4810]: I1201 14:37:19.566710 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 14:37:19 crc kubenswrapper[4810]: I1201 14:37:19.660060 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc-kubelet-dir\") pod \"a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc\" (UID: \"a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc\") " Dec 01 14:37:19 crc kubenswrapper[4810]: I1201 14:37:19.660189 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc-kube-api-access\") pod \"a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc\" (UID: \"a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc\") " Dec 01 14:37:19 crc kubenswrapper[4810]: I1201 14:37:19.660319 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc" (UID: "a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:37:19 crc kubenswrapper[4810]: I1201 14:37:19.660598 4810 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:19 crc kubenswrapper[4810]: I1201 14:37:19.666391 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc" (UID: "a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:37:19 crc kubenswrapper[4810]: I1201 14:37:19.761613 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:20 crc kubenswrapper[4810]: I1201 14:37:20.334079 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc","Type":"ContainerDied","Data":"e2a326868405faf0dd65e3c448619df7f96ea09fd1dc7e71b9df5af8dac08060"} Dec 01 14:37:20 crc kubenswrapper[4810]: I1201 14:37:20.334124 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 14:37:20 crc kubenswrapper[4810]: I1201 14:37:20.334128 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2a326868405faf0dd65e3c448619df7f96ea09fd1dc7e71b9df5af8dac08060" Dec 01 14:37:28 crc kubenswrapper[4810]: I1201 14:37:28.375489 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xnb48" event={"ID":"93bf2297-d7e7-48f2-8a96-6c33d5b3754b","Type":"ContainerStarted","Data":"1256a0ea89cf35a3062cba66bd62af423c3acbaa80f3c7d8dc0852c087d53d8b"} Dec 01 14:37:28 crc kubenswrapper[4810]: I1201 14:37:28.378393 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mxxf2" event={"ID":"48b25ecd-efe4-4363-975d-a17cb60d10a3","Type":"ContainerStarted","Data":"a691c0d8ab99cbae03cdc0b8bfac7484b24ac7c1f5c85108c3b36ca1bf420673"} Dec 01 14:37:29 crc kubenswrapper[4810]: I1201 14:37:29.385656 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxt6s" event={"ID":"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5","Type":"ContainerStarted","Data":"0e21734ba2f73e2a661ec8a6491b122b5b97369fa45aef4adc74e825260f6b08"} Dec 01 14:37:29 crc kubenswrapper[4810]: I1201 14:37:29.388241 4810 generic.go:334] "Generic (PLEG): container finished" podID="93bf2297-d7e7-48f2-8a96-6c33d5b3754b" containerID="1256a0ea89cf35a3062cba66bd62af423c3acbaa80f3c7d8dc0852c087d53d8b" exitCode=0 Dec 01 14:37:29 crc kubenswrapper[4810]: I1201 14:37:29.388314 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xnb48" event={"ID":"93bf2297-d7e7-48f2-8a96-6c33d5b3754b","Type":"ContainerDied","Data":"1256a0ea89cf35a3062cba66bd62af423c3acbaa80f3c7d8dc0852c087d53d8b"} Dec 01 14:37:29 crc kubenswrapper[4810]: I1201 14:37:29.390292 4810 generic.go:334] "Generic (PLEG): container finished" podID="48b25ecd-efe4-4363-975d-a17cb60d10a3" containerID="a691c0d8ab99cbae03cdc0b8bfac7484b24ac7c1f5c85108c3b36ca1bf420673" exitCode=0 Dec 01 14:37:29 crc kubenswrapper[4810]: I1201 14:37:29.390353 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mxxf2" event={"ID":"48b25ecd-efe4-4363-975d-a17cb60d10a3","Type":"ContainerDied","Data":"a691c0d8ab99cbae03cdc0b8bfac7484b24ac7c1f5c85108c3b36ca1bf420673"} Dec 01 14:37:29 crc kubenswrapper[4810]: I1201 14:37:29.393286 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcjqf" event={"ID":"00d0bac3-a43f-469e-9335-df4013f7432e","Type":"ContainerStarted","Data":"767535a7f3e5bc715874a85d9edbd906dbf73d5d72716b57163bfa7a38854b46"} Dec 01 14:37:30 crc kubenswrapper[4810]: I1201 14:37:30.401083 4810 generic.go:334] "Generic (PLEG): container finished" podID="b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5" containerID="0e21734ba2f73e2a661ec8a6491b122b5b97369fa45aef4adc74e825260f6b08" exitCode=0 Dec 01 14:37:30 crc kubenswrapper[4810]: I1201 14:37:30.401680 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxt6s" event={"ID":"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5","Type":"ContainerDied","Data":"0e21734ba2f73e2a661ec8a6491b122b5b97369fa45aef4adc74e825260f6b08"} Dec 01 14:37:30 crc kubenswrapper[4810]: I1201 14:37:30.408912 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5xbfx" event={"ID":"efdc70a7-17ce-46e8-b0d0-b3f917370e8b","Type":"ContainerStarted","Data":"217a70cbe343fc89bceae6f32057c9d9a56c10a20558c2e2a3edc2c4efc80a5a"} Dec 01 14:37:30 crc kubenswrapper[4810]: I1201 14:37:30.412174 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xnb48" event={"ID":"93bf2297-d7e7-48f2-8a96-6c33d5b3754b","Type":"ContainerStarted","Data":"43180ba72c6bf26818cd7afe73674086f3015314ec2ac6c796a829b955537ab7"} Dec 01 14:37:30 crc kubenswrapper[4810]: I1201 14:37:30.413392 4810 generic.go:334] "Generic (PLEG): container finished" podID="0eb57a13-0325-432a-a641-44903e611452" containerID="70a724a2884c0b68f1390815d55ce8c3004d537e72fb052a1b7a5461346186c9" exitCode=0 Dec 01 14:37:30 crc kubenswrapper[4810]: I1201 14:37:30.413455 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w85bt" event={"ID":"0eb57a13-0325-432a-a641-44903e611452","Type":"ContainerDied","Data":"70a724a2884c0b68f1390815d55ce8c3004d537e72fb052a1b7a5461346186c9"} Dec 01 14:37:30 crc kubenswrapper[4810]: I1201 14:37:30.417666 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mxxf2" event={"ID":"48b25ecd-efe4-4363-975d-a17cb60d10a3","Type":"ContainerStarted","Data":"623a2e95a731a1a2a898812b1b748399cc63804daafc1db4853af735fb6b8aae"} Dec 01 14:37:30 crc kubenswrapper[4810]: I1201 14:37:30.420359 4810 generic.go:334] "Generic (PLEG): container finished" podID="00d0bac3-a43f-469e-9335-df4013f7432e" containerID="767535a7f3e5bc715874a85d9edbd906dbf73d5d72716b57163bfa7a38854b46" exitCode=0 Dec 01 14:37:30 crc kubenswrapper[4810]: I1201 14:37:30.420396 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcjqf" event={"ID":"00d0bac3-a43f-469e-9335-df4013f7432e","Type":"ContainerDied","Data":"767535a7f3e5bc715874a85d9edbd906dbf73d5d72716b57163bfa7a38854b46"} Dec 01 14:37:30 crc kubenswrapper[4810]: I1201 14:37:30.441947 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xnb48" podStartSLOduration=3.133237875 podStartE2EDuration="58.441928413s" podCreationTimestamp="2025-12-01 14:36:32 +0000 UTC" firstStartedPulling="2025-12-01 14:36:34.717935564 +0000 UTC m=+160.481445167" lastFinishedPulling="2025-12-01 14:37:30.026626102 +0000 UTC m=+215.790135705" observedRunningTime="2025-12-01 14:37:30.439754395 +0000 UTC m=+216.203263998" watchObservedRunningTime="2025-12-01 14:37:30.441928413 +0000 UTC m=+216.205438016" Dec 01 14:37:30 crc kubenswrapper[4810]: I1201 14:37:30.501929 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mxxf2" podStartSLOduration=3.2088045960000002 podStartE2EDuration="1m0.501909629s" podCreationTimestamp="2025-12-01 14:36:30 +0000 UTC" firstStartedPulling="2025-12-01 14:36:32.665691041 +0000 UTC m=+158.429200644" lastFinishedPulling="2025-12-01 14:37:29.958796054 +0000 UTC m=+215.722305677" observedRunningTime="2025-12-01 14:37:30.501344934 +0000 UTC m=+216.264854557" watchObservedRunningTime="2025-12-01 14:37:30.501909629 +0000 UTC m=+216.265419232" Dec 01 14:37:31 crc kubenswrapper[4810]: I1201 14:37:31.124576 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mxxf2" Dec 01 14:37:31 crc kubenswrapper[4810]: I1201 14:37:31.124626 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mxxf2" Dec 01 14:37:31 crc kubenswrapper[4810]: I1201 14:37:31.431934 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w85bt" event={"ID":"0eb57a13-0325-432a-a641-44903e611452","Type":"ContainerStarted","Data":"613396d7f6c01de3482298d0081376ec8d0d5ae02acee86eb5e19ca7c18a5c46"} Dec 01 14:37:31 crc kubenswrapper[4810]: I1201 14:37:31.434121 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fm8lj" event={"ID":"ca896d69-0651-48e1-a001-0b3c57e1ece4","Type":"ContainerStarted","Data":"699937a2d5d16c74f98d394717c8910b78d31ccedcc1f4fe5a7b4e8c8f009814"} Dec 01 14:37:31 crc kubenswrapper[4810]: I1201 14:37:31.437360 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcjqf" event={"ID":"00d0bac3-a43f-469e-9335-df4013f7432e","Type":"ContainerStarted","Data":"0fd88b0ee7b773525f6289e968cc51d666962280d0245c888069038085bd82c8"} Dec 01 14:37:31 crc kubenswrapper[4810]: I1201 14:37:31.440030 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jfmhr" event={"ID":"044c93cf-65a5-455b-af09-63e80ee90adc","Type":"ContainerStarted","Data":"3d5bb36d069af72390e575961f96e7873c26fb756a8f7139de2ad0c33cd6b77d"} Dec 01 14:37:31 crc kubenswrapper[4810]: I1201 14:37:31.442065 4810 generic.go:334] "Generic (PLEG): container finished" podID="efdc70a7-17ce-46e8-b0d0-b3f917370e8b" containerID="217a70cbe343fc89bceae6f32057c9d9a56c10a20558c2e2a3edc2c4efc80a5a" exitCode=0 Dec 01 14:37:31 crc kubenswrapper[4810]: I1201 14:37:31.442126 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5xbfx" event={"ID":"efdc70a7-17ce-46e8-b0d0-b3f917370e8b","Type":"ContainerDied","Data":"217a70cbe343fc89bceae6f32057c9d9a56c10a20558c2e2a3edc2c4efc80a5a"} Dec 01 14:37:31 crc kubenswrapper[4810]: I1201 14:37:31.445582 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxt6s" event={"ID":"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5","Type":"ContainerStarted","Data":"e94d75582a7295b57e07c5299519521fd2e53fd9970afd7c17f397f524ed0fd2"} Dec 01 14:37:31 crc kubenswrapper[4810]: I1201 14:37:31.455283 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w85bt" podStartSLOduration=2.995035436 podStartE2EDuration="1m1.45526219s" podCreationTimestamp="2025-12-01 14:36:30 +0000 UTC" firstStartedPulling="2025-12-01 14:36:32.574850893 +0000 UTC m=+158.338360496" lastFinishedPulling="2025-12-01 14:37:31.035077647 +0000 UTC m=+216.798587250" observedRunningTime="2025-12-01 14:37:31.450241295 +0000 UTC m=+217.213750918" watchObservedRunningTime="2025-12-01 14:37:31.45526219 +0000 UTC m=+217.218771793" Dec 01 14:37:31 crc kubenswrapper[4810]: I1201 14:37:31.468595 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pcjqf" podStartSLOduration=3.243602172 podStartE2EDuration="1m0.468580369s" podCreationTimestamp="2025-12-01 14:36:31 +0000 UTC" firstStartedPulling="2025-12-01 14:36:33.66576349 +0000 UTC m=+159.429273093" lastFinishedPulling="2025-12-01 14:37:30.890741687 +0000 UTC m=+216.654251290" observedRunningTime="2025-12-01 14:37:31.466952975 +0000 UTC m=+217.230462608" watchObservedRunningTime="2025-12-01 14:37:31.468580369 +0000 UTC m=+217.232089972" Dec 01 14:37:31 crc kubenswrapper[4810]: I1201 14:37:31.489865 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pxt6s" podStartSLOduration=3.9176568400000003 podStartE2EDuration="1m2.489847132s" podCreationTimestamp="2025-12-01 14:36:29 +0000 UTC" firstStartedPulling="2025-12-01 14:36:32.540155797 +0000 UTC m=+158.303665400" lastFinishedPulling="2025-12-01 14:37:31.112346089 +0000 UTC m=+216.875855692" observedRunningTime="2025-12-01 14:37:31.486840751 +0000 UTC m=+217.250350354" watchObservedRunningTime="2025-12-01 14:37:31.489847132 +0000 UTC m=+217.253356735" Dec 01 14:37:32 crc kubenswrapper[4810]: I1201 14:37:32.111205 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pcjqf" Dec 01 14:37:32 crc kubenswrapper[4810]: I1201 14:37:32.111668 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pcjqf" Dec 01 14:37:32 crc kubenswrapper[4810]: I1201 14:37:32.182670 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-mxxf2" podUID="48b25ecd-efe4-4363-975d-a17cb60d10a3" containerName="registry-server" probeResult="failure" output=< Dec 01 14:37:32 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Dec 01 14:37:32 crc kubenswrapper[4810]: > Dec 01 14:37:32 crc kubenswrapper[4810]: I1201 14:37:32.452300 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5xbfx" event={"ID":"efdc70a7-17ce-46e8-b0d0-b3f917370e8b","Type":"ContainerStarted","Data":"45068d930fed3e04fe3f205ac9b35aec6a5f21e556548e9a3e67c623a37a64c7"} Dec 01 14:37:32 crc kubenswrapper[4810]: I1201 14:37:32.455062 4810 generic.go:334] "Generic (PLEG): container finished" podID="ca896d69-0651-48e1-a001-0b3c57e1ece4" containerID="699937a2d5d16c74f98d394717c8910b78d31ccedcc1f4fe5a7b4e8c8f009814" exitCode=0 Dec 01 14:37:32 crc kubenswrapper[4810]: I1201 14:37:32.455125 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fm8lj" event={"ID":"ca896d69-0651-48e1-a001-0b3c57e1ece4","Type":"ContainerDied","Data":"699937a2d5d16c74f98d394717c8910b78d31ccedcc1f4fe5a7b4e8c8f009814"} Dec 01 14:37:32 crc kubenswrapper[4810]: I1201 14:37:32.458960 4810 generic.go:334] "Generic (PLEG): container finished" podID="044c93cf-65a5-455b-af09-63e80ee90adc" containerID="3d5bb36d069af72390e575961f96e7873c26fb756a8f7139de2ad0c33cd6b77d" exitCode=0 Dec 01 14:37:32 crc kubenswrapper[4810]: I1201 14:37:32.459395 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jfmhr" event={"ID":"044c93cf-65a5-455b-af09-63e80ee90adc","Type":"ContainerDied","Data":"3d5bb36d069af72390e575961f96e7873c26fb756a8f7139de2ad0c33cd6b77d"} Dec 01 14:37:32 crc kubenswrapper[4810]: I1201 14:37:32.474209 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5xbfx" podStartSLOduration=3.259618288 podStartE2EDuration="59.474187627s" podCreationTimestamp="2025-12-01 14:36:33 +0000 UTC" firstStartedPulling="2025-12-01 14:36:35.793344564 +0000 UTC m=+161.556854167" lastFinishedPulling="2025-12-01 14:37:32.007913893 +0000 UTC m=+217.771423506" observedRunningTime="2025-12-01 14:37:32.472988915 +0000 UTC m=+218.236498518" watchObservedRunningTime="2025-12-01 14:37:32.474187627 +0000 UTC m=+218.237697230" Dec 01 14:37:32 crc kubenswrapper[4810]: I1201 14:37:32.971746 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:37:32 crc kubenswrapper[4810]: I1201 14:37:32.971807 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:37:32 crc kubenswrapper[4810]: I1201 14:37:32.971864 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:37:32 crc kubenswrapper[4810]: I1201 14:37:32.972413 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 14:37:32 crc kubenswrapper[4810]: I1201 14:37:32.972611 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8" gracePeriod=600 Dec 01 14:37:33 crc kubenswrapper[4810]: I1201 14:37:33.158663 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-pcjqf" podUID="00d0bac3-a43f-469e-9335-df4013f7432e" containerName="registry-server" probeResult="failure" output=< Dec 01 14:37:33 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Dec 01 14:37:33 crc kubenswrapper[4810]: > Dec 01 14:37:33 crc kubenswrapper[4810]: I1201 14:37:33.360188 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xnb48" Dec 01 14:37:33 crc kubenswrapper[4810]: I1201 14:37:33.360570 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xnb48" Dec 01 14:37:33 crc kubenswrapper[4810]: I1201 14:37:33.466381 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8" exitCode=0 Dec 01 14:37:33 crc kubenswrapper[4810]: I1201 14:37:33.466451 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8"} Dec 01 14:37:33 crc kubenswrapper[4810]: I1201 14:37:33.710891 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5xbfx" Dec 01 14:37:33 crc kubenswrapper[4810]: I1201 14:37:33.711337 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5xbfx" Dec 01 14:37:34 crc kubenswrapper[4810]: I1201 14:37:34.417791 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xnb48" podUID="93bf2297-d7e7-48f2-8a96-6c33d5b3754b" containerName="registry-server" probeResult="failure" output=< Dec 01 14:37:34 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Dec 01 14:37:34 crc kubenswrapper[4810]: > Dec 01 14:37:34 crc kubenswrapper[4810]: I1201 14:37:34.474406 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fm8lj" event={"ID":"ca896d69-0651-48e1-a001-0b3c57e1ece4","Type":"ContainerStarted","Data":"2d08d2f1c0b268ba9a58259fe5f93283af6c1535fbe9eaaeda089e05d13363b9"} Dec 01 14:37:34 crc kubenswrapper[4810]: I1201 14:37:34.477210 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jfmhr" event={"ID":"044c93cf-65a5-455b-af09-63e80ee90adc","Type":"ContainerStarted","Data":"3830476673bdc6d02a03cc9a5ba269092d96ed9d3dd49d519708618a6c306cd0"} Dec 01 14:37:34 crc kubenswrapper[4810]: I1201 14:37:34.478969 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"3e2bceffac8b9b74a53ba6e40ccf495e62023cf75155f8f803b43b64a708ed90"} Dec 01 14:37:34 crc kubenswrapper[4810]: I1201 14:37:34.495996 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fm8lj" podStartSLOduration=4.291532416 podStartE2EDuration="1m5.49598113s" podCreationTimestamp="2025-12-01 14:36:29 +0000 UTC" firstStartedPulling="2025-12-01 14:36:32.539841899 +0000 UTC m=+158.303351512" lastFinishedPulling="2025-12-01 14:37:33.744290623 +0000 UTC m=+219.507800226" observedRunningTime="2025-12-01 14:37:34.492618089 +0000 UTC m=+220.256127682" watchObservedRunningTime="2025-12-01 14:37:34.49598113 +0000 UTC m=+220.259490733" Dec 01 14:37:34 crc kubenswrapper[4810]: I1201 14:37:34.541195 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jfmhr" podStartSLOduration=3.238671665 podStartE2EDuration="1m2.541162497s" podCreationTimestamp="2025-12-01 14:36:32 +0000 UTC" firstStartedPulling="2025-12-01 14:36:33.669707256 +0000 UTC m=+159.433216849" lastFinishedPulling="2025-12-01 14:37:32.972198078 +0000 UTC m=+218.735707681" observedRunningTime="2025-12-01 14:37:34.537573551 +0000 UTC m=+220.301083164" watchObservedRunningTime="2025-12-01 14:37:34.541162497 +0000 UTC m=+220.304672100" Dec 01 14:37:34 crc kubenswrapper[4810]: I1201 14:37:34.754201 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5xbfx" podUID="efdc70a7-17ce-46e8-b0d0-b3f917370e8b" containerName="registry-server" probeResult="failure" output=< Dec 01 14:37:34 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Dec 01 14:37:34 crc kubenswrapper[4810]: > Dec 01 14:37:40 crc kubenswrapper[4810]: I1201 14:37:40.283533 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fm8lj" Dec 01 14:37:40 crc kubenswrapper[4810]: I1201 14:37:40.284774 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fm8lj" Dec 01 14:37:40 crc kubenswrapper[4810]: I1201 14:37:40.325563 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fm8lj" Dec 01 14:37:40 crc kubenswrapper[4810]: I1201 14:37:40.546368 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fm8lj" Dec 01 14:37:40 crc kubenswrapper[4810]: I1201 14:37:40.692163 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w85bt" Dec 01 14:37:40 crc kubenswrapper[4810]: I1201 14:37:40.692726 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w85bt" Dec 01 14:37:40 crc kubenswrapper[4810]: I1201 14:37:40.730163 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w85bt" Dec 01 14:37:41 crc kubenswrapper[4810]: I1201 14:37:41.132675 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pxt6s" Dec 01 14:37:41 crc kubenswrapper[4810]: I1201 14:37:41.132716 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pxt6s" Dec 01 14:37:41 crc kubenswrapper[4810]: I1201 14:37:41.164186 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mxxf2" Dec 01 14:37:41 crc kubenswrapper[4810]: I1201 14:37:41.174285 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pxt6s" Dec 01 14:37:41 crc kubenswrapper[4810]: I1201 14:37:41.205389 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mxxf2" Dec 01 14:37:41 crc kubenswrapper[4810]: I1201 14:37:41.553994 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pxt6s" Dec 01 14:37:41 crc kubenswrapper[4810]: I1201 14:37:41.563346 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w85bt" Dec 01 14:37:42 crc kubenswrapper[4810]: I1201 14:37:42.128461 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w85bt"] Dec 01 14:37:42 crc kubenswrapper[4810]: I1201 14:37:42.189695 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pcjqf" Dec 01 14:37:42 crc kubenswrapper[4810]: I1201 14:37:42.262044 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pcjqf" Dec 01 14:37:42 crc kubenswrapper[4810]: I1201 14:37:42.277308 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fdwcc"] Dec 01 14:37:42 crc kubenswrapper[4810]: I1201 14:37:42.484940 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jfmhr" Dec 01 14:37:42 crc kubenswrapper[4810]: I1201 14:37:42.485287 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jfmhr" Dec 01 14:37:42 crc kubenswrapper[4810]: I1201 14:37:42.523193 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jfmhr" Dec 01 14:37:43 crc kubenswrapper[4810]: I1201 14:37:43.422115 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xnb48" Dec 01 14:37:43 crc kubenswrapper[4810]: I1201 14:37:43.487907 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xnb48" Dec 01 14:37:43 crc kubenswrapper[4810]: I1201 14:37:43.526228 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-w85bt" podUID="0eb57a13-0325-432a-a641-44903e611452" containerName="registry-server" containerID="cri-o://613396d7f6c01de3482298d0081376ec8d0d5ae02acee86eb5e19ca7c18a5c46" gracePeriod=2 Dec 01 14:37:43 crc kubenswrapper[4810]: I1201 14:37:43.526714 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mxxf2"] Dec 01 14:37:43 crc kubenswrapper[4810]: I1201 14:37:43.527136 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mxxf2" podUID="48b25ecd-efe4-4363-975d-a17cb60d10a3" containerName="registry-server" containerID="cri-o://623a2e95a731a1a2a898812b1b748399cc63804daafc1db4853af735fb6b8aae" gracePeriod=2 Dec 01 14:37:43 crc kubenswrapper[4810]: I1201 14:37:43.591662 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jfmhr" Dec 01 14:37:43 crc kubenswrapper[4810]: I1201 14:37:43.750037 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5xbfx" Dec 01 14:37:43 crc kubenswrapper[4810]: I1201 14:37:43.802148 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5xbfx" Dec 01 14:37:45 crc kubenswrapper[4810]: I1201 14:37:45.924269 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jfmhr"] Dec 01 14:37:45 crc kubenswrapper[4810]: I1201 14:37:45.925309 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jfmhr" podUID="044c93cf-65a5-455b-af09-63e80ee90adc" containerName="registry-server" containerID="cri-o://3830476673bdc6d02a03cc9a5ba269092d96ed9d3dd49d519708618a6c306cd0" gracePeriod=2 Dec 01 14:37:46 crc kubenswrapper[4810]: I1201 14:37:46.923767 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5xbfx"] Dec 01 14:37:46 crc kubenswrapper[4810]: I1201 14:37:46.924091 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5xbfx" podUID="efdc70a7-17ce-46e8-b0d0-b3f917370e8b" containerName="registry-server" containerID="cri-o://45068d930fed3e04fe3f205ac9b35aec6a5f21e556548e9a3e67c623a37a64c7" gracePeriod=2 Dec 01 14:37:47 crc kubenswrapper[4810]: I1201 14:37:47.564659 4810 generic.go:334] "Generic (PLEG): container finished" podID="efdc70a7-17ce-46e8-b0d0-b3f917370e8b" containerID="45068d930fed3e04fe3f205ac9b35aec6a5f21e556548e9a3e67c623a37a64c7" exitCode=0 Dec 01 14:37:47 crc kubenswrapper[4810]: I1201 14:37:47.564729 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5xbfx" event={"ID":"efdc70a7-17ce-46e8-b0d0-b3f917370e8b","Type":"ContainerDied","Data":"45068d930fed3e04fe3f205ac9b35aec6a5f21e556548e9a3e67c623a37a64c7"} Dec 01 14:37:47 crc kubenswrapper[4810]: I1201 14:37:47.572334 4810 generic.go:334] "Generic (PLEG): container finished" podID="0eb57a13-0325-432a-a641-44903e611452" containerID="613396d7f6c01de3482298d0081376ec8d0d5ae02acee86eb5e19ca7c18a5c46" exitCode=0 Dec 01 14:37:47 crc kubenswrapper[4810]: I1201 14:37:47.572369 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w85bt" event={"ID":"0eb57a13-0325-432a-a641-44903e611452","Type":"ContainerDied","Data":"613396d7f6c01de3482298d0081376ec8d0d5ae02acee86eb5e19ca7c18a5c46"} Dec 01 14:37:47 crc kubenswrapper[4810]: I1201 14:37:47.586244 4810 generic.go:334] "Generic (PLEG): container finished" podID="48b25ecd-efe4-4363-975d-a17cb60d10a3" containerID="623a2e95a731a1a2a898812b1b748399cc63804daafc1db4853af735fb6b8aae" exitCode=0 Dec 01 14:37:47 crc kubenswrapper[4810]: I1201 14:37:47.586288 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mxxf2" event={"ID":"48b25ecd-efe4-4363-975d-a17cb60d10a3","Type":"ContainerDied","Data":"623a2e95a731a1a2a898812b1b748399cc63804daafc1db4853af735fb6b8aae"} Dec 01 14:37:47 crc kubenswrapper[4810]: I1201 14:37:47.828102 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mxxf2" Dec 01 14:37:47 crc kubenswrapper[4810]: I1201 14:37:47.919771 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jfmhr" Dec 01 14:37:47 crc kubenswrapper[4810]: I1201 14:37:47.926102 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w85bt" Dec 01 14:37:47 crc kubenswrapper[4810]: I1201 14:37:47.936712 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5xbfx" Dec 01 14:37:47 crc kubenswrapper[4810]: I1201 14:37:47.941039 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48b25ecd-efe4-4363-975d-a17cb60d10a3-utilities\") pod \"48b25ecd-efe4-4363-975d-a17cb60d10a3\" (UID: \"48b25ecd-efe4-4363-975d-a17cb60d10a3\") " Dec 01 14:37:47 crc kubenswrapper[4810]: I1201 14:37:47.941259 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48b25ecd-efe4-4363-975d-a17cb60d10a3-catalog-content\") pod \"48b25ecd-efe4-4363-975d-a17cb60d10a3\" (UID: \"48b25ecd-efe4-4363-975d-a17cb60d10a3\") " Dec 01 14:37:47 crc kubenswrapper[4810]: I1201 14:37:47.941412 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzx49\" (UniqueName: \"kubernetes.io/projected/48b25ecd-efe4-4363-975d-a17cb60d10a3-kube-api-access-gzx49\") pod \"48b25ecd-efe4-4363-975d-a17cb60d10a3\" (UID: \"48b25ecd-efe4-4363-975d-a17cb60d10a3\") " Dec 01 14:37:47 crc kubenswrapper[4810]: I1201 14:37:47.941949 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48b25ecd-efe4-4363-975d-a17cb60d10a3-utilities" (OuterVolumeSpecName: "utilities") pod "48b25ecd-efe4-4363-975d-a17cb60d10a3" (UID: "48b25ecd-efe4-4363-975d-a17cb60d10a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:37:47 crc kubenswrapper[4810]: I1201 14:37:47.949081 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48b25ecd-efe4-4363-975d-a17cb60d10a3-kube-api-access-gzx49" (OuterVolumeSpecName: "kube-api-access-gzx49") pod "48b25ecd-efe4-4363-975d-a17cb60d10a3" (UID: "48b25ecd-efe4-4363-975d-a17cb60d10a3"). InnerVolumeSpecName "kube-api-access-gzx49". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:37:47 crc kubenswrapper[4810]: I1201 14:37:47.999525 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48b25ecd-efe4-4363-975d-a17cb60d10a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48b25ecd-efe4-4363-975d-a17cb60d10a3" (UID: "48b25ecd-efe4-4363-975d-a17cb60d10a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.042593 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0eb57a13-0325-432a-a641-44903e611452-catalog-content\") pod \"0eb57a13-0325-432a-a641-44903e611452\" (UID: \"0eb57a13-0325-432a-a641-44903e611452\") " Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.042680 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85t58\" (UniqueName: \"kubernetes.io/projected/efdc70a7-17ce-46e8-b0d0-b3f917370e8b-kube-api-access-85t58\") pod \"efdc70a7-17ce-46e8-b0d0-b3f917370e8b\" (UID: \"efdc70a7-17ce-46e8-b0d0-b3f917370e8b\") " Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.042736 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efdc70a7-17ce-46e8-b0d0-b3f917370e8b-catalog-content\") pod \"efdc70a7-17ce-46e8-b0d0-b3f917370e8b\" (UID: \"efdc70a7-17ce-46e8-b0d0-b3f917370e8b\") " Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.042765 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/044c93cf-65a5-455b-af09-63e80ee90adc-catalog-content\") pod \"044c93cf-65a5-455b-af09-63e80ee90adc\" (UID: \"044c93cf-65a5-455b-af09-63e80ee90adc\") " Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.042807 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efdc70a7-17ce-46e8-b0d0-b3f917370e8b-utilities\") pod \"efdc70a7-17ce-46e8-b0d0-b3f917370e8b\" (UID: \"efdc70a7-17ce-46e8-b0d0-b3f917370e8b\") " Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.042833 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/044c93cf-65a5-455b-af09-63e80ee90adc-utilities\") pod \"044c93cf-65a5-455b-af09-63e80ee90adc\" (UID: \"044c93cf-65a5-455b-af09-63e80ee90adc\") " Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.042863 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0eb57a13-0325-432a-a641-44903e611452-utilities\") pod \"0eb57a13-0325-432a-a641-44903e611452\" (UID: \"0eb57a13-0325-432a-a641-44903e611452\") " Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.042915 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8gd9\" (UniqueName: \"kubernetes.io/projected/044c93cf-65a5-455b-af09-63e80ee90adc-kube-api-access-m8gd9\") pod \"044c93cf-65a5-455b-af09-63e80ee90adc\" (UID: \"044c93cf-65a5-455b-af09-63e80ee90adc\") " Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.042980 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cmnm\" (UniqueName: \"kubernetes.io/projected/0eb57a13-0325-432a-a641-44903e611452-kube-api-access-4cmnm\") pod \"0eb57a13-0325-432a-a641-44903e611452\" (UID: \"0eb57a13-0325-432a-a641-44903e611452\") " Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.043321 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48b25ecd-efe4-4363-975d-a17cb60d10a3-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.043338 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48b25ecd-efe4-4363-975d-a17cb60d10a3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.043352 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzx49\" (UniqueName: \"kubernetes.io/projected/48b25ecd-efe4-4363-975d-a17cb60d10a3-kube-api-access-gzx49\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.046133 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efdc70a7-17ce-46e8-b0d0-b3f917370e8b-utilities" (OuterVolumeSpecName: "utilities") pod "efdc70a7-17ce-46e8-b0d0-b3f917370e8b" (UID: "efdc70a7-17ce-46e8-b0d0-b3f917370e8b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.046890 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/044c93cf-65a5-455b-af09-63e80ee90adc-utilities" (OuterVolumeSpecName: "utilities") pod "044c93cf-65a5-455b-af09-63e80ee90adc" (UID: "044c93cf-65a5-455b-af09-63e80ee90adc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.047298 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0eb57a13-0325-432a-a641-44903e611452-utilities" (OuterVolumeSpecName: "utilities") pod "0eb57a13-0325-432a-a641-44903e611452" (UID: "0eb57a13-0325-432a-a641-44903e611452"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.049498 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0eb57a13-0325-432a-a641-44903e611452-kube-api-access-4cmnm" (OuterVolumeSpecName: "kube-api-access-4cmnm") pod "0eb57a13-0325-432a-a641-44903e611452" (UID: "0eb57a13-0325-432a-a641-44903e611452"). InnerVolumeSpecName "kube-api-access-4cmnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.049709 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/044c93cf-65a5-455b-af09-63e80ee90adc-kube-api-access-m8gd9" (OuterVolumeSpecName: "kube-api-access-m8gd9") pod "044c93cf-65a5-455b-af09-63e80ee90adc" (UID: "044c93cf-65a5-455b-af09-63e80ee90adc"). InnerVolumeSpecName "kube-api-access-m8gd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.049827 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdc70a7-17ce-46e8-b0d0-b3f917370e8b-kube-api-access-85t58" (OuterVolumeSpecName: "kube-api-access-85t58") pod "efdc70a7-17ce-46e8-b0d0-b3f917370e8b" (UID: "efdc70a7-17ce-46e8-b0d0-b3f917370e8b"). InnerVolumeSpecName "kube-api-access-85t58". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.066395 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/044c93cf-65a5-455b-af09-63e80ee90adc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "044c93cf-65a5-455b-af09-63e80ee90adc" (UID: "044c93cf-65a5-455b-af09-63e80ee90adc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.099190 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0eb57a13-0325-432a-a641-44903e611452-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0eb57a13-0325-432a-a641-44903e611452" (UID: "0eb57a13-0325-432a-a641-44903e611452"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.146276 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cmnm\" (UniqueName: \"kubernetes.io/projected/0eb57a13-0325-432a-a641-44903e611452-kube-api-access-4cmnm\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.146316 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0eb57a13-0325-432a-a641-44903e611452-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.146328 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85t58\" (UniqueName: \"kubernetes.io/projected/efdc70a7-17ce-46e8-b0d0-b3f917370e8b-kube-api-access-85t58\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.146339 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/044c93cf-65a5-455b-af09-63e80ee90adc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.146350 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/efdc70a7-17ce-46e8-b0d0-b3f917370e8b-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.146362 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/044c93cf-65a5-455b-af09-63e80ee90adc-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.146371 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0eb57a13-0325-432a-a641-44903e611452-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.146381 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8gd9\" (UniqueName: \"kubernetes.io/projected/044c93cf-65a5-455b-af09-63e80ee90adc-kube-api-access-m8gd9\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.169484 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efdc70a7-17ce-46e8-b0d0-b3f917370e8b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "efdc70a7-17ce-46e8-b0d0-b3f917370e8b" (UID: "efdc70a7-17ce-46e8-b0d0-b3f917370e8b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.247810 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/efdc70a7-17ce-46e8-b0d0-b3f917370e8b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.592905 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w85bt" event={"ID":"0eb57a13-0325-432a-a641-44903e611452","Type":"ContainerDied","Data":"8c85a26918dfa33be67ec28f7eb3b0e6044fcb41f03a58ac6911b2c70278beba"} Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.592940 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w85bt" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.592954 4810 scope.go:117] "RemoveContainer" containerID="613396d7f6c01de3482298d0081376ec8d0d5ae02acee86eb5e19ca7c18a5c46" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.599331 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mxxf2" event={"ID":"48b25ecd-efe4-4363-975d-a17cb60d10a3","Type":"ContainerDied","Data":"65b0d40e94b979c9f1db101c98618c7c3bababccb87ceabb8b8afc90ca5cce4d"} Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.599401 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mxxf2" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.602792 4810 generic.go:334] "Generic (PLEG): container finished" podID="044c93cf-65a5-455b-af09-63e80ee90adc" containerID="3830476673bdc6d02a03cc9a5ba269092d96ed9d3dd49d519708618a6c306cd0" exitCode=0 Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.603003 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jfmhr" event={"ID":"044c93cf-65a5-455b-af09-63e80ee90adc","Type":"ContainerDied","Data":"3830476673bdc6d02a03cc9a5ba269092d96ed9d3dd49d519708618a6c306cd0"} Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.603442 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jfmhr" event={"ID":"044c93cf-65a5-455b-af09-63e80ee90adc","Type":"ContainerDied","Data":"f9cf065d5741a41448de2e8fbef83d4f974ab885e6b7efcab06e3ac398bbc30f"} Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.603142 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jfmhr" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.609047 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5xbfx" event={"ID":"efdc70a7-17ce-46e8-b0d0-b3f917370e8b","Type":"ContainerDied","Data":"35f43ae5ffc69861107929092cf339c31cf6b17eb803ecd95aef518046156bce"} Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.609173 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5xbfx" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.614644 4810 scope.go:117] "RemoveContainer" containerID="70a724a2884c0b68f1390815d55ce8c3004d537e72fb052a1b7a5461346186c9" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.623746 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w85bt"] Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.630838 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-w85bt"] Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.634478 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mxxf2"] Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.639960 4810 scope.go:117] "RemoveContainer" containerID="2865f4bc975be1b299eb1d90c2f2c414046274442d695c9cbf1d0ae7c2c55bce" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.645147 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mxxf2"] Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.658145 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jfmhr"] Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.666099 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jfmhr"] Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.674390 4810 scope.go:117] "RemoveContainer" containerID="623a2e95a731a1a2a898812b1b748399cc63804daafc1db4853af735fb6b8aae" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.676457 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5xbfx"] Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.681534 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5xbfx"] Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.688465 4810 scope.go:117] "RemoveContainer" containerID="a691c0d8ab99cbae03cdc0b8bfac7484b24ac7c1f5c85108c3b36ca1bf420673" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.706629 4810 scope.go:117] "RemoveContainer" containerID="666bfc240c437d01adb6528b64d139a4a2158c5e4d116636984aab3415beb93c" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.723168 4810 scope.go:117] "RemoveContainer" containerID="3830476673bdc6d02a03cc9a5ba269092d96ed9d3dd49d519708618a6c306cd0" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.741069 4810 scope.go:117] "RemoveContainer" containerID="3d5bb36d069af72390e575961f96e7873c26fb756a8f7139de2ad0c33cd6b77d" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.753578 4810 scope.go:117] "RemoveContainer" containerID="61680d97b0ea45f2ec6a4556a79cfa33352b48beb901da6878b4d3774b0cf3fa" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.768801 4810 scope.go:117] "RemoveContainer" containerID="3830476673bdc6d02a03cc9a5ba269092d96ed9d3dd49d519708618a6c306cd0" Dec 01 14:37:48 crc kubenswrapper[4810]: E1201 14:37:48.769368 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3830476673bdc6d02a03cc9a5ba269092d96ed9d3dd49d519708618a6c306cd0\": container with ID starting with 3830476673bdc6d02a03cc9a5ba269092d96ed9d3dd49d519708618a6c306cd0 not found: ID does not exist" containerID="3830476673bdc6d02a03cc9a5ba269092d96ed9d3dd49d519708618a6c306cd0" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.769603 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3830476673bdc6d02a03cc9a5ba269092d96ed9d3dd49d519708618a6c306cd0"} err="failed to get container status \"3830476673bdc6d02a03cc9a5ba269092d96ed9d3dd49d519708618a6c306cd0\": rpc error: code = NotFound desc = could not find container \"3830476673bdc6d02a03cc9a5ba269092d96ed9d3dd49d519708618a6c306cd0\": container with ID starting with 3830476673bdc6d02a03cc9a5ba269092d96ed9d3dd49d519708618a6c306cd0 not found: ID does not exist" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.769711 4810 scope.go:117] "RemoveContainer" containerID="3d5bb36d069af72390e575961f96e7873c26fb756a8f7139de2ad0c33cd6b77d" Dec 01 14:37:48 crc kubenswrapper[4810]: E1201 14:37:48.770212 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d5bb36d069af72390e575961f96e7873c26fb756a8f7139de2ad0c33cd6b77d\": container with ID starting with 3d5bb36d069af72390e575961f96e7873c26fb756a8f7139de2ad0c33cd6b77d not found: ID does not exist" containerID="3d5bb36d069af72390e575961f96e7873c26fb756a8f7139de2ad0c33cd6b77d" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.770261 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d5bb36d069af72390e575961f96e7873c26fb756a8f7139de2ad0c33cd6b77d"} err="failed to get container status \"3d5bb36d069af72390e575961f96e7873c26fb756a8f7139de2ad0c33cd6b77d\": rpc error: code = NotFound desc = could not find container \"3d5bb36d069af72390e575961f96e7873c26fb756a8f7139de2ad0c33cd6b77d\": container with ID starting with 3d5bb36d069af72390e575961f96e7873c26fb756a8f7139de2ad0c33cd6b77d not found: ID does not exist" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.770290 4810 scope.go:117] "RemoveContainer" containerID="61680d97b0ea45f2ec6a4556a79cfa33352b48beb901da6878b4d3774b0cf3fa" Dec 01 14:37:48 crc kubenswrapper[4810]: E1201 14:37:48.771400 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61680d97b0ea45f2ec6a4556a79cfa33352b48beb901da6878b4d3774b0cf3fa\": container with ID starting with 61680d97b0ea45f2ec6a4556a79cfa33352b48beb901da6878b4d3774b0cf3fa not found: ID does not exist" containerID="61680d97b0ea45f2ec6a4556a79cfa33352b48beb901da6878b4d3774b0cf3fa" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.771443 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61680d97b0ea45f2ec6a4556a79cfa33352b48beb901da6878b4d3774b0cf3fa"} err="failed to get container status \"61680d97b0ea45f2ec6a4556a79cfa33352b48beb901da6878b4d3774b0cf3fa\": rpc error: code = NotFound desc = could not find container \"61680d97b0ea45f2ec6a4556a79cfa33352b48beb901da6878b4d3774b0cf3fa\": container with ID starting with 61680d97b0ea45f2ec6a4556a79cfa33352b48beb901da6878b4d3774b0cf3fa not found: ID does not exist" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.771490 4810 scope.go:117] "RemoveContainer" containerID="45068d930fed3e04fe3f205ac9b35aec6a5f21e556548e9a3e67c623a37a64c7" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.784838 4810 scope.go:117] "RemoveContainer" containerID="217a70cbe343fc89bceae6f32057c9d9a56c10a20558c2e2a3edc2c4efc80a5a" Dec 01 14:37:48 crc kubenswrapper[4810]: I1201 14:37:48.800936 4810 scope.go:117] "RemoveContainer" containerID="63c9af014bb4942859b29dda99c1dc2e2af9236e18b1859c097d9ffad6fb993a" Dec 01 14:37:50 crc kubenswrapper[4810]: I1201 14:37:50.500424 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="044c93cf-65a5-455b-af09-63e80ee90adc" path="/var/lib/kubelet/pods/044c93cf-65a5-455b-af09-63e80ee90adc/volumes" Dec 01 14:37:50 crc kubenswrapper[4810]: I1201 14:37:50.502441 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0eb57a13-0325-432a-a641-44903e611452" path="/var/lib/kubelet/pods/0eb57a13-0325-432a-a641-44903e611452/volumes" Dec 01 14:37:50 crc kubenswrapper[4810]: I1201 14:37:50.503665 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48b25ecd-efe4-4363-975d-a17cb60d10a3" path="/var/lib/kubelet/pods/48b25ecd-efe4-4363-975d-a17cb60d10a3/volumes" Dec 01 14:37:50 crc kubenswrapper[4810]: I1201 14:37:50.505626 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdc70a7-17ce-46e8-b0d0-b3f917370e8b" path="/var/lib/kubelet/pods/efdc70a7-17ce-46e8-b0d0-b3f917370e8b/volumes" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.252291 4810 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.252925 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="044c93cf-65a5-455b-af09-63e80ee90adc" containerName="extract-content" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.252945 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="044c93cf-65a5-455b-af09-63e80ee90adc" containerName="extract-content" Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.252963 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48b25ecd-efe4-4363-975d-a17cb60d10a3" containerName="extract-content" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.252973 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="48b25ecd-efe4-4363-975d-a17cb60d10a3" containerName="extract-content" Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.252987 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eb57a13-0325-432a-a641-44903e611452" containerName="extract-content" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.252996 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eb57a13-0325-432a-a641-44903e611452" containerName="extract-content" Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.253006 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="044c93cf-65a5-455b-af09-63e80ee90adc" containerName="registry-server" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.253015 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="044c93cf-65a5-455b-af09-63e80ee90adc" containerName="registry-server" Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.253031 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="044c93cf-65a5-455b-af09-63e80ee90adc" containerName="extract-utilities" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.253062 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="044c93cf-65a5-455b-af09-63e80ee90adc" containerName="extract-utilities" Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.253079 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efdc70a7-17ce-46e8-b0d0-b3f917370e8b" containerName="extract-utilities" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.253089 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="efdc70a7-17ce-46e8-b0d0-b3f917370e8b" containerName="extract-utilities" Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.253104 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48b25ecd-efe4-4363-975d-a17cb60d10a3" containerName="registry-server" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.253114 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="48b25ecd-efe4-4363-975d-a17cb60d10a3" containerName="registry-server" Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.253129 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efdc70a7-17ce-46e8-b0d0-b3f917370e8b" containerName="registry-server" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.253139 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="efdc70a7-17ce-46e8-b0d0-b3f917370e8b" containerName="registry-server" Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.253154 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48b25ecd-efe4-4363-975d-a17cb60d10a3" containerName="extract-utilities" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.253165 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="48b25ecd-efe4-4363-975d-a17cb60d10a3" containerName="extract-utilities" Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.253178 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eb57a13-0325-432a-a641-44903e611452" containerName="registry-server" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.253188 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eb57a13-0325-432a-a641-44903e611452" containerName="registry-server" Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.253203 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eb57a13-0325-432a-a641-44903e611452" containerName="extract-utilities" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.253214 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eb57a13-0325-432a-a641-44903e611452" containerName="extract-utilities" Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.253229 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efdc70a7-17ce-46e8-b0d0-b3f917370e8b" containerName="extract-content" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.253239 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="efdc70a7-17ce-46e8-b0d0-b3f917370e8b" containerName="extract-content" Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.253251 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc" containerName="pruner" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.253262 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc" containerName="pruner" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.253397 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0eb57a13-0325-432a-a641-44903e611452" containerName="registry-server" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.253427 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="48b25ecd-efe4-4363-975d-a17cb60d10a3" containerName="registry-server" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.253442 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="efdc70a7-17ce-46e8-b0d0-b3f917370e8b" containerName="registry-server" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.253454 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="044c93cf-65a5-455b-af09-63e80ee90adc" containerName="registry-server" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.253491 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a97aeb7e-87e4-4f40-8b28-d1a4da0d6ebc" containerName="pruner" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.253949 4810 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.254098 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.254297 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef" gracePeriod=15 Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.254356 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88" gracePeriod=15 Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.254451 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066" gracePeriod=15 Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.254445 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135" gracePeriod=15 Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.254550 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3" gracePeriod=15 Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.255236 4810 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.255385 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.255397 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.255412 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.255429 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.255447 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.255497 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.255512 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.255520 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.255531 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.255538 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.255557 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.255565 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.255744 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.255764 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.255773 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.255786 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.255795 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.255809 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 14:37:55 crc kubenswrapper[4810]: E1201 14:37:55.255924 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.255933 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.437544 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.438153 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.438193 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.438225 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.438242 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.438289 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.438305 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.438322 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.539373 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.539415 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.539445 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.539490 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.539516 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.539540 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.539541 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.539581 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.539593 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.539831 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.539556 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.539860 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.539870 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.539891 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.539889 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.539923 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.659589 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.661088 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.662171 4810 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135" exitCode=0 Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.662207 4810 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88" exitCode=0 Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.662224 4810 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066" exitCode=0 Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.662239 4810 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3" exitCode=2 Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.662269 4810 scope.go:117] "RemoveContainer" containerID="871397519f50cc33c19ed018777a7f66dafe6efd7324ea4f8062881e3c8b8a9a" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.665160 4810 generic.go:334] "Generic (PLEG): container finished" podID="24394371-0f74-48a7-97c3-315489b78fe9" containerID="c52e52fae75a809aa7e157420966e9451d0cca4470f9d9c742e0bce07e8a0f80" exitCode=0 Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.665202 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"24394371-0f74-48a7-97c3-315489b78fe9","Type":"ContainerDied","Data":"c52e52fae75a809aa7e157420966e9451d0cca4470f9d9c742e0bce07e8a0f80"} Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.666373 4810 status_manager.go:851] "Failed to get status for pod" podUID="24394371-0f74-48a7-97c3-315489b78fe9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.667042 4810 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.871612 4810 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Dec 01 14:37:55 crc kubenswrapper[4810]: I1201 14:37:55.871730 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Dec 01 14:37:56 crc kubenswrapper[4810]: I1201 14:37:56.673749 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 14:37:56 crc kubenswrapper[4810]: I1201 14:37:56.886684 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 14:37:56 crc kubenswrapper[4810]: I1201 14:37:56.887375 4810 status_manager.go:851] "Failed to get status for pod" podUID="24394371-0f74-48a7-97c3-315489b78fe9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:37:57 crc kubenswrapper[4810]: I1201 14:37:57.060769 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/24394371-0f74-48a7-97c3-315489b78fe9-kube-api-access\") pod \"24394371-0f74-48a7-97c3-315489b78fe9\" (UID: \"24394371-0f74-48a7-97c3-315489b78fe9\") " Dec 01 14:37:57 crc kubenswrapper[4810]: I1201 14:37:57.060916 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/24394371-0f74-48a7-97c3-315489b78fe9-var-lock\") pod \"24394371-0f74-48a7-97c3-315489b78fe9\" (UID: \"24394371-0f74-48a7-97c3-315489b78fe9\") " Dec 01 14:37:57 crc kubenswrapper[4810]: I1201 14:37:57.060957 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/24394371-0f74-48a7-97c3-315489b78fe9-kubelet-dir\") pod \"24394371-0f74-48a7-97c3-315489b78fe9\" (UID: \"24394371-0f74-48a7-97c3-315489b78fe9\") " Dec 01 14:37:57 crc kubenswrapper[4810]: I1201 14:37:57.060965 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24394371-0f74-48a7-97c3-315489b78fe9-var-lock" (OuterVolumeSpecName: "var-lock") pod "24394371-0f74-48a7-97c3-315489b78fe9" (UID: "24394371-0f74-48a7-97c3-315489b78fe9"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:37:57 crc kubenswrapper[4810]: I1201 14:37:57.061078 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24394371-0f74-48a7-97c3-315489b78fe9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "24394371-0f74-48a7-97c3-315489b78fe9" (UID: "24394371-0f74-48a7-97c3-315489b78fe9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:37:57 crc kubenswrapper[4810]: I1201 14:37:57.061330 4810 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/24394371-0f74-48a7-97c3-315489b78fe9-var-lock\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:57 crc kubenswrapper[4810]: I1201 14:37:57.061352 4810 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/24394371-0f74-48a7-97c3-315489b78fe9-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:57 crc kubenswrapper[4810]: I1201 14:37:57.065947 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24394371-0f74-48a7-97c3-315489b78fe9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "24394371-0f74-48a7-97c3-315489b78fe9" (UID: "24394371-0f74-48a7-97c3-315489b78fe9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:37:57 crc kubenswrapper[4810]: I1201 14:37:57.162423 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/24394371-0f74-48a7-97c3-315489b78fe9-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:57 crc kubenswrapper[4810]: E1201 14:37:57.375933 4810 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:37:57 crc kubenswrapper[4810]: E1201 14:37:57.376342 4810 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:37:57 crc kubenswrapper[4810]: E1201 14:37:57.377270 4810 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:37:57 crc kubenswrapper[4810]: E1201 14:37:57.377614 4810 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:37:57 crc kubenswrapper[4810]: E1201 14:37:57.377846 4810 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:37:57 crc kubenswrapper[4810]: I1201 14:37:57.377885 4810 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 01 14:37:57 crc kubenswrapper[4810]: E1201 14:37:57.378104 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="200ms" Dec 01 14:37:57 crc kubenswrapper[4810]: E1201 14:37:57.579095 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="400ms" Dec 01 14:37:57 crc kubenswrapper[4810]: I1201 14:37:57.693022 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"24394371-0f74-48a7-97c3-315489b78fe9","Type":"ContainerDied","Data":"292f5aff760680f0d08040f0f159285e7e43a0cfcc8e392091d540866f3f1e0d"} Dec 01 14:37:57 crc kubenswrapper[4810]: I1201 14:37:57.693065 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="292f5aff760680f0d08040f0f159285e7e43a0cfcc8e392091d540866f3f1e0d" Dec 01 14:37:57 crc kubenswrapper[4810]: I1201 14:37:57.693088 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 14:37:57 crc kubenswrapper[4810]: I1201 14:37:57.709693 4810 status_manager.go:851] "Failed to get status for pod" podUID="24394371-0f74-48a7-97c3-315489b78fe9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:37:57 crc kubenswrapper[4810]: E1201 14:37:57.980382 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="800ms" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.445439 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.446509 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.446927 4810 status_manager.go:851] "Failed to get status for pod" podUID="24394371-0f74-48a7-97c3-315489b78fe9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.447135 4810 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.582577 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.582701 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.583055 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.583100 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.583168 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.583249 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.583392 4810 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.583543 4810 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.583683 4810 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.701186 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.703576 4810 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef" exitCode=0 Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.703646 4810 scope.go:117] "RemoveContainer" containerID="66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.703677 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.704275 4810 status_manager.go:851] "Failed to get status for pod" podUID="24394371-0f74-48a7-97c3-315489b78fe9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.704459 4810 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.719053 4810 scope.go:117] "RemoveContainer" containerID="5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.719379 4810 status_manager.go:851] "Failed to get status for pod" podUID="24394371-0f74-48a7-97c3-315489b78fe9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.719666 4810 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.732695 4810 scope.go:117] "RemoveContainer" containerID="39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.743871 4810 scope.go:117] "RemoveContainer" containerID="8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.755189 4810 scope.go:117] "RemoveContainer" containerID="45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.769393 4810 scope.go:117] "RemoveContainer" containerID="9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6" Dec 01 14:37:58 crc kubenswrapper[4810]: E1201 14:37:58.781505 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="1.6s" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.787922 4810 scope.go:117] "RemoveContainer" containerID="66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135" Dec 01 14:37:58 crc kubenswrapper[4810]: E1201 14:37:58.788583 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\": container with ID starting with 66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135 not found: ID does not exist" containerID="66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.789047 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135"} err="failed to get container status \"66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\": rpc error: code = NotFound desc = could not find container \"66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135\": container with ID starting with 66b7fce00306fe55264d746db8678b0c1c6d91a13b7aa691a9acf19e1d7c7135 not found: ID does not exist" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.789136 4810 scope.go:117] "RemoveContainer" containerID="5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88" Dec 01 14:37:58 crc kubenswrapper[4810]: E1201 14:37:58.790017 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\": container with ID starting with 5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88 not found: ID does not exist" containerID="5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.790050 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88"} err="failed to get container status \"5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\": rpc error: code = NotFound desc = could not find container \"5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88\": container with ID starting with 5bd43eb744ccb7dd6c6411718bb828582d0c22dd81ea830a5e4d3c512047cc88 not found: ID does not exist" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.790075 4810 scope.go:117] "RemoveContainer" containerID="39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066" Dec 01 14:37:58 crc kubenswrapper[4810]: E1201 14:37:58.790856 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\": container with ID starting with 39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066 not found: ID does not exist" containerID="39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.790881 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066"} err="failed to get container status \"39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\": rpc error: code = NotFound desc = could not find container \"39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066\": container with ID starting with 39fdaa5baef6f3ac481618111bdc129c145740341590c4a3c2391dff99619066 not found: ID does not exist" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.790899 4810 scope.go:117] "RemoveContainer" containerID="8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3" Dec 01 14:37:58 crc kubenswrapper[4810]: E1201 14:37:58.791086 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\": container with ID starting with 8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3 not found: ID does not exist" containerID="8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.791106 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3"} err="failed to get container status \"8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\": rpc error: code = NotFound desc = could not find container \"8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3\": container with ID starting with 8abf8ed182919e88248bec7a53d4707d8f5350ebbe244a53caa3c471884576d3 not found: ID does not exist" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.791120 4810 scope.go:117] "RemoveContainer" containerID="45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef" Dec 01 14:37:58 crc kubenswrapper[4810]: E1201 14:37:58.791315 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\": container with ID starting with 45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef not found: ID does not exist" containerID="45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.791333 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef"} err="failed to get container status \"45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\": rpc error: code = NotFound desc = could not find container \"45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef\": container with ID starting with 45b974b08a216f0247f6fbd4c3ad73d9f2f1447ae50559546fd30befe41429ef not found: ID does not exist" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.791346 4810 scope.go:117] "RemoveContainer" containerID="9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6" Dec 01 14:37:58 crc kubenswrapper[4810]: E1201 14:37:58.791686 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\": container with ID starting with 9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6 not found: ID does not exist" containerID="9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6" Dec 01 14:37:58 crc kubenswrapper[4810]: I1201 14:37:58.791724 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6"} err="failed to get container status \"9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\": rpc error: code = NotFound desc = could not find container \"9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6\": container with ID starting with 9f0a18317b28e2cde314faea58f20f7b10a97b14ef9a30eb19cb79edcc1b5ee6 not found: ID does not exist" Dec 01 14:38:00 crc kubenswrapper[4810]: E1201 14:38:00.301498 4810 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.182:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:38:00 crc kubenswrapper[4810]: I1201 14:38:00.301997 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:38:00 crc kubenswrapper[4810]: E1201 14:38:00.350210 4810 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.182:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d1e3deef2f38b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 14:38:00.349840267 +0000 UTC m=+246.113349880,LastTimestamp:2025-12-01 14:38:00.349840267 +0000 UTC m=+246.113349880,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 14:38:00 crc kubenswrapper[4810]: E1201 14:38:00.382619 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="3.2s" Dec 01 14:38:00 crc kubenswrapper[4810]: I1201 14:38:00.501567 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 01 14:38:00 crc kubenswrapper[4810]: I1201 14:38:00.718028 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"206bc2a2f9fd203b471578db4666b029bef18c2fbdc71f8a660d0b1f5442b50d"} Dec 01 14:38:00 crc kubenswrapper[4810]: I1201 14:38:00.718092 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"7fff2c5535497a5fb2feae6d36122400c9fbddbe2a6bae1197596ff9d4e0bc30"} Dec 01 14:38:00 crc kubenswrapper[4810]: E1201 14:38:00.718874 4810 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.182:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:38:00 crc kubenswrapper[4810]: I1201 14:38:00.718884 4810 status_manager.go:851] "Failed to get status for pod" podUID="24394371-0f74-48a7-97c3-315489b78fe9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:03 crc kubenswrapper[4810]: E1201 14:38:03.584106 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="6.4s" Dec 01 14:38:04 crc kubenswrapper[4810]: I1201 14:38:04.497269 4810 status_manager.go:851] "Failed to get status for pod" podUID="24394371-0f74-48a7-97c3-315489b78fe9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.303031 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" podUID="ca08c59d-eb6c-4c1f-bf44-f161a760ae67" containerName="oauth-openshift" containerID="cri-o://84a4a3351bfbfaafb13aaed6e7313179248d4768883d32685a5b24ec18ffbc40" gracePeriod=15 Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.709198 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.710225 4810 status_manager.go:851] "Failed to get status for pod" podUID="24394371-0f74-48a7-97c3-315489b78fe9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.710688 4810 status_manager.go:851] "Failed to get status for pod" podUID="ca08c59d-eb6c-4c1f-bf44-f161a760ae67" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-fdwcc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.763958 4810 generic.go:334] "Generic (PLEG): container finished" podID="ca08c59d-eb6c-4c1f-bf44-f161a760ae67" containerID="84a4a3351bfbfaafb13aaed6e7313179248d4768883d32685a5b24ec18ffbc40" exitCode=0 Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.764015 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" event={"ID":"ca08c59d-eb6c-4c1f-bf44-f161a760ae67","Type":"ContainerDied","Data":"84a4a3351bfbfaafb13aaed6e7313179248d4768883d32685a5b24ec18ffbc40"} Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.764065 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" event={"ID":"ca08c59d-eb6c-4c1f-bf44-f161a760ae67","Type":"ContainerDied","Data":"5417b4431dd53b77aa7188e3d7b50a6ce2617a247d0d5b9eda6b382a23435654"} Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.764091 4810 scope.go:117] "RemoveContainer" containerID="84a4a3351bfbfaafb13aaed6e7313179248d4768883d32685a5b24ec18ffbc40" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.764087 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.764911 4810 status_manager.go:851] "Failed to get status for pod" podUID="24394371-0f74-48a7-97c3-315489b78fe9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.765183 4810 status_manager.go:851] "Failed to get status for pod" podUID="ca08c59d-eb6c-4c1f-bf44-f161a760ae67" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-fdwcc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.796042 4810 scope.go:117] "RemoveContainer" containerID="84a4a3351bfbfaafb13aaed6e7313179248d4768883d32685a5b24ec18ffbc40" Dec 01 14:38:07 crc kubenswrapper[4810]: E1201 14:38:07.796636 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84a4a3351bfbfaafb13aaed6e7313179248d4768883d32685a5b24ec18ffbc40\": container with ID starting with 84a4a3351bfbfaafb13aaed6e7313179248d4768883d32685a5b24ec18ffbc40 not found: ID does not exist" containerID="84a4a3351bfbfaafb13aaed6e7313179248d4768883d32685a5b24ec18ffbc40" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.796696 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84a4a3351bfbfaafb13aaed6e7313179248d4768883d32685a5b24ec18ffbc40"} err="failed to get container status \"84a4a3351bfbfaafb13aaed6e7313179248d4768883d32685a5b24ec18ffbc40\": rpc error: code = NotFound desc = could not find container \"84a4a3351bfbfaafb13aaed6e7313179248d4768883d32685a5b24ec18ffbc40\": container with ID starting with 84a4a3351bfbfaafb13aaed6e7313179248d4768883d32685a5b24ec18ffbc40 not found: ID does not exist" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.904870 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-session\") pod \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.904937 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-serving-cert\") pod \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.904964 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-audit-dir\") pod \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.904995 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-cliconfig\") pod \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.905029 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-audit-policies\") pod \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.905078 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-ocp-branding-template\") pod \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.905115 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-service-ca\") pod \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.905110 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "ca08c59d-eb6c-4c1f-bf44-f161a760ae67" (UID: "ca08c59d-eb6c-4c1f-bf44-f161a760ae67"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.905164 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-template-error\") pod \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.905201 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5xpn\" (UniqueName: \"kubernetes.io/projected/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-kube-api-access-j5xpn\") pod \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.905237 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-template-login\") pod \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.905272 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-trusted-ca-bundle\") pod \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.905304 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-idp-0-file-data\") pod \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.905344 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-template-provider-selection\") pod \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.905386 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-router-certs\") pod \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\" (UID: \"ca08c59d-eb6c-4c1f-bf44-f161a760ae67\") " Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.905650 4810 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.906152 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "ca08c59d-eb6c-4c1f-bf44-f161a760ae67" (UID: "ca08c59d-eb6c-4c1f-bf44-f161a760ae67"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.906309 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "ca08c59d-eb6c-4c1f-bf44-f161a760ae67" (UID: "ca08c59d-eb6c-4c1f-bf44-f161a760ae67"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.906701 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "ca08c59d-eb6c-4c1f-bf44-f161a760ae67" (UID: "ca08c59d-eb6c-4c1f-bf44-f161a760ae67"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.907547 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "ca08c59d-eb6c-4c1f-bf44-f161a760ae67" (UID: "ca08c59d-eb6c-4c1f-bf44-f161a760ae67"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.912822 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "ca08c59d-eb6c-4c1f-bf44-f161a760ae67" (UID: "ca08c59d-eb6c-4c1f-bf44-f161a760ae67"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.913119 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "ca08c59d-eb6c-4c1f-bf44-f161a760ae67" (UID: "ca08c59d-eb6c-4c1f-bf44-f161a760ae67"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.913495 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-kube-api-access-j5xpn" (OuterVolumeSpecName: "kube-api-access-j5xpn") pod "ca08c59d-eb6c-4c1f-bf44-f161a760ae67" (UID: "ca08c59d-eb6c-4c1f-bf44-f161a760ae67"). InnerVolumeSpecName "kube-api-access-j5xpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.913561 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "ca08c59d-eb6c-4c1f-bf44-f161a760ae67" (UID: "ca08c59d-eb6c-4c1f-bf44-f161a760ae67"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.914348 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "ca08c59d-eb6c-4c1f-bf44-f161a760ae67" (UID: "ca08c59d-eb6c-4c1f-bf44-f161a760ae67"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.914790 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "ca08c59d-eb6c-4c1f-bf44-f161a760ae67" (UID: "ca08c59d-eb6c-4c1f-bf44-f161a760ae67"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.915236 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "ca08c59d-eb6c-4c1f-bf44-f161a760ae67" (UID: "ca08c59d-eb6c-4c1f-bf44-f161a760ae67"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.915496 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "ca08c59d-eb6c-4c1f-bf44-f161a760ae67" (UID: "ca08c59d-eb6c-4c1f-bf44-f161a760ae67"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:38:07 crc kubenswrapper[4810]: I1201 14:38:07.917737 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "ca08c59d-eb6c-4c1f-bf44-f161a760ae67" (UID: "ca08c59d-eb6c-4c1f-bf44-f161a760ae67"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:38:08 crc kubenswrapper[4810]: I1201 14:38:08.007268 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:08 crc kubenswrapper[4810]: I1201 14:38:08.007310 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:08 crc kubenswrapper[4810]: I1201 14:38:08.007322 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:08 crc kubenswrapper[4810]: I1201 14:38:08.007334 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:08 crc kubenswrapper[4810]: I1201 14:38:08.007344 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:08 crc kubenswrapper[4810]: I1201 14:38:08.007354 4810 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:08 crc kubenswrapper[4810]: I1201 14:38:08.007364 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:08 crc kubenswrapper[4810]: I1201 14:38:08.007375 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:08 crc kubenswrapper[4810]: I1201 14:38:08.007383 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:08 crc kubenswrapper[4810]: I1201 14:38:08.007391 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5xpn\" (UniqueName: \"kubernetes.io/projected/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-kube-api-access-j5xpn\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:08 crc kubenswrapper[4810]: I1201 14:38:08.007400 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:08 crc kubenswrapper[4810]: I1201 14:38:08.007408 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:08 crc kubenswrapper[4810]: I1201 14:38:08.007416 4810 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ca08c59d-eb6c-4c1f-bf44-f161a760ae67-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:08 crc kubenswrapper[4810]: I1201 14:38:08.088123 4810 status_manager.go:851] "Failed to get status for pod" podUID="ca08c59d-eb6c-4c1f-bf44-f161a760ae67" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-fdwcc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:08 crc kubenswrapper[4810]: I1201 14:38:08.088574 4810 status_manager.go:851] "Failed to get status for pod" podUID="24394371-0f74-48a7-97c3-315489b78fe9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:09 crc kubenswrapper[4810]: I1201 14:38:09.784530 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 01 14:38:09 crc kubenswrapper[4810]: I1201 14:38:09.784904 4810 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e" exitCode=1 Dec 01 14:38:09 crc kubenswrapper[4810]: I1201 14:38:09.784940 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e"} Dec 01 14:38:09 crc kubenswrapper[4810]: I1201 14:38:09.785524 4810 scope.go:117] "RemoveContainer" containerID="7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e" Dec 01 14:38:09 crc kubenswrapper[4810]: I1201 14:38:09.786288 4810 status_manager.go:851] "Failed to get status for pod" podUID="ca08c59d-eb6c-4c1f-bf44-f161a760ae67" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-fdwcc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:09 crc kubenswrapper[4810]: I1201 14:38:09.786769 4810 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:09 crc kubenswrapper[4810]: I1201 14:38:09.787431 4810 status_manager.go:851] "Failed to get status for pod" podUID="24394371-0f74-48a7-97c3-315489b78fe9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:09 crc kubenswrapper[4810]: E1201 14:38:09.908915 4810 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.182:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d1e3deef2f38b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 14:38:00.349840267 +0000 UTC m=+246.113349880,LastTimestamp:2025-12-01 14:38:00.349840267 +0000 UTC m=+246.113349880,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 14:38:09 crc kubenswrapper[4810]: E1201 14:38:09.985994 4810 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="7s" Dec 01 14:38:10 crc kubenswrapper[4810]: I1201 14:38:10.491102 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:38:10 crc kubenswrapper[4810]: I1201 14:38:10.492138 4810 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:10 crc kubenswrapper[4810]: I1201 14:38:10.492656 4810 status_manager.go:851] "Failed to get status for pod" podUID="24394371-0f74-48a7-97c3-315489b78fe9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:10 crc kubenswrapper[4810]: I1201 14:38:10.493099 4810 status_manager.go:851] "Failed to get status for pod" podUID="ca08c59d-eb6c-4c1f-bf44-f161a760ae67" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-fdwcc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:10 crc kubenswrapper[4810]: I1201 14:38:10.504012 4810 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="41bbeb3b-624e-4b84-896f-ecd37a5cb7f8" Dec 01 14:38:10 crc kubenswrapper[4810]: I1201 14:38:10.504049 4810 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="41bbeb3b-624e-4b84-896f-ecd37a5cb7f8" Dec 01 14:38:10 crc kubenswrapper[4810]: E1201 14:38:10.504597 4810 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:38:10 crc kubenswrapper[4810]: I1201 14:38:10.505278 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:38:10 crc kubenswrapper[4810]: W1201 14:38:10.524276 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-4bfe0c57bec3d0006233f7f0869e3ad26db90f03b69d233a5f3ebaf12f022eb5 WatchSource:0}: Error finding container 4bfe0c57bec3d0006233f7f0869e3ad26db90f03b69d233a5f3ebaf12f022eb5: Status 404 returned error can't find the container with id 4bfe0c57bec3d0006233f7f0869e3ad26db90f03b69d233a5f3ebaf12f022eb5 Dec 01 14:38:10 crc kubenswrapper[4810]: I1201 14:38:10.795773 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 01 14:38:10 crc kubenswrapper[4810]: I1201 14:38:10.796316 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3657f23abef8beba44444f3b182e43dc58df375c694d65275604a15ce3724277"} Dec 01 14:38:10 crc kubenswrapper[4810]: I1201 14:38:10.798517 4810 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:10 crc kubenswrapper[4810]: I1201 14:38:10.799335 4810 status_manager.go:851] "Failed to get status for pod" podUID="24394371-0f74-48a7-97c3-315489b78fe9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:10 crc kubenswrapper[4810]: I1201 14:38:10.799927 4810 status_manager.go:851] "Failed to get status for pod" podUID="ca08c59d-eb6c-4c1f-bf44-f161a760ae67" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-fdwcc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:10 crc kubenswrapper[4810]: I1201 14:38:10.800697 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4bfe0c57bec3d0006233f7f0869e3ad26db90f03b69d233a5f3ebaf12f022eb5"} Dec 01 14:38:11 crc kubenswrapper[4810]: I1201 14:38:11.806567 4810 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="874f470ecda3dc9199f9ac7c39492e27223c0c4aa5c553171aca855a1b5c8f73" exitCode=0 Dec 01 14:38:11 crc kubenswrapper[4810]: I1201 14:38:11.806609 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"874f470ecda3dc9199f9ac7c39492e27223c0c4aa5c553171aca855a1b5c8f73"} Dec 01 14:38:11 crc kubenswrapper[4810]: I1201 14:38:11.806804 4810 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="41bbeb3b-624e-4b84-896f-ecd37a5cb7f8" Dec 01 14:38:11 crc kubenswrapper[4810]: I1201 14:38:11.806826 4810 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="41bbeb3b-624e-4b84-896f-ecd37a5cb7f8" Dec 01 14:38:11 crc kubenswrapper[4810]: E1201 14:38:11.807159 4810 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:38:11 crc kubenswrapper[4810]: I1201 14:38:11.807257 4810 status_manager.go:851] "Failed to get status for pod" podUID="24394371-0f74-48a7-97c3-315489b78fe9" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:11 crc kubenswrapper[4810]: I1201 14:38:11.807530 4810 status_manager.go:851] "Failed to get status for pod" podUID="ca08c59d-eb6c-4c1f-bf44-f161a760ae67" pod="openshift-authentication/oauth-openshift-558db77b4-fdwcc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-fdwcc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:11 crc kubenswrapper[4810]: I1201 14:38:11.808208 4810 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Dec 01 14:38:12 crc kubenswrapper[4810]: I1201 14:38:12.816034 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"dc0c51416a087e0eef7f876c95c502783e3e33f7b0031e25f52ee03387535d9a"} Dec 01 14:38:12 crc kubenswrapper[4810]: I1201 14:38:12.816358 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"85a8a5467886d6022403b529c1b97e4b495326cac1fb2ff16781d44653127172"} Dec 01 14:38:12 crc kubenswrapper[4810]: I1201 14:38:12.816367 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4f970002a864ad2d84ad1928f2f2555bbaa426278a7264c2147ad4b90bd6307e"} Dec 01 14:38:12 crc kubenswrapper[4810]: I1201 14:38:12.816376 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ffed0ba39e72785c6e17d0e75e707948138b4bd1812f268970d695059f562dbe"} Dec 01 14:38:13 crc kubenswrapper[4810]: I1201 14:38:13.529842 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:38:13 crc kubenswrapper[4810]: I1201 14:38:13.530173 4810 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 01 14:38:13 crc kubenswrapper[4810]: I1201 14:38:13.530249 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 01 14:38:13 crc kubenswrapper[4810]: I1201 14:38:13.824145 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c0e4b4423895b6d58cdb314f93fcaa3ebdcaf9fea7322d08b9ee84e396e9983d"} Dec 01 14:38:13 crc kubenswrapper[4810]: I1201 14:38:13.824318 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:38:13 crc kubenswrapper[4810]: I1201 14:38:13.824395 4810 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="41bbeb3b-624e-4b84-896f-ecd37a5cb7f8" Dec 01 14:38:13 crc kubenswrapper[4810]: I1201 14:38:13.824413 4810 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="41bbeb3b-624e-4b84-896f-ecd37a5cb7f8" Dec 01 14:38:15 crc kubenswrapper[4810]: I1201 14:38:15.506101 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:38:15 crc kubenswrapper[4810]: I1201 14:38:15.506157 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:38:15 crc kubenswrapper[4810]: I1201 14:38:15.512980 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:38:18 crc kubenswrapper[4810]: I1201 14:38:18.536517 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:38:18 crc kubenswrapper[4810]: I1201 14:38:18.833530 4810 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:38:18 crc kubenswrapper[4810]: I1201 14:38:18.849629 4810 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="41bbeb3b-624e-4b84-896f-ecd37a5cb7f8" Dec 01 14:38:18 crc kubenswrapper[4810]: I1201 14:38:18.849661 4810 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="41bbeb3b-624e-4b84-896f-ecd37a5cb7f8" Dec 01 14:38:18 crc kubenswrapper[4810]: I1201 14:38:18.853373 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:38:18 crc kubenswrapper[4810]: I1201 14:38:18.917437 4810 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="8be1a72c-bdfb-4318-bdbc-49aded88b084" Dec 01 14:38:19 crc kubenswrapper[4810]: I1201 14:38:19.855904 4810 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="41bbeb3b-624e-4b84-896f-ecd37a5cb7f8" Dec 01 14:38:19 crc kubenswrapper[4810]: I1201 14:38:19.857616 4810 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="41bbeb3b-624e-4b84-896f-ecd37a5cb7f8" Dec 01 14:38:19 crc kubenswrapper[4810]: I1201 14:38:19.858529 4810 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="8be1a72c-bdfb-4318-bdbc-49aded88b084" Dec 01 14:38:23 crc kubenswrapper[4810]: I1201 14:38:23.529569 4810 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 01 14:38:23 crc kubenswrapper[4810]: I1201 14:38:23.530729 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 01 14:38:28 crc kubenswrapper[4810]: I1201 14:38:28.612869 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 01 14:38:28 crc kubenswrapper[4810]: I1201 14:38:28.617539 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 01 14:38:28 crc kubenswrapper[4810]: I1201 14:38:28.758600 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 01 14:38:28 crc kubenswrapper[4810]: I1201 14:38:28.920688 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 01 14:38:29 crc kubenswrapper[4810]: I1201 14:38:29.395077 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 01 14:38:29 crc kubenswrapper[4810]: I1201 14:38:29.580222 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 01 14:38:29 crc kubenswrapper[4810]: I1201 14:38:29.733896 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 01 14:38:29 crc kubenswrapper[4810]: I1201 14:38:29.859438 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 01 14:38:30 crc kubenswrapper[4810]: I1201 14:38:30.195551 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 01 14:38:30 crc kubenswrapper[4810]: I1201 14:38:30.267249 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 01 14:38:30 crc kubenswrapper[4810]: I1201 14:38:30.512776 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 01 14:38:30 crc kubenswrapper[4810]: I1201 14:38:30.716810 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 01 14:38:30 crc kubenswrapper[4810]: I1201 14:38:30.771425 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 01 14:38:30 crc kubenswrapper[4810]: I1201 14:38:30.881459 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 01 14:38:31 crc kubenswrapper[4810]: I1201 14:38:31.092577 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 01 14:38:31 crc kubenswrapper[4810]: I1201 14:38:31.220206 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 01 14:38:31 crc kubenswrapper[4810]: I1201 14:38:31.295078 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 14:38:31 crc kubenswrapper[4810]: I1201 14:38:31.383877 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 01 14:38:31 crc kubenswrapper[4810]: I1201 14:38:31.493915 4810 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 01 14:38:31 crc kubenswrapper[4810]: I1201 14:38:31.814545 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 01 14:38:32 crc kubenswrapper[4810]: I1201 14:38:32.077273 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 01 14:38:32 crc kubenswrapper[4810]: I1201 14:38:32.094206 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 01 14:38:32 crc kubenswrapper[4810]: I1201 14:38:32.220035 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 01 14:38:32 crc kubenswrapper[4810]: I1201 14:38:32.246233 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 01 14:38:32 crc kubenswrapper[4810]: I1201 14:38:32.256072 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 01 14:38:32 crc kubenswrapper[4810]: I1201 14:38:32.361959 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 14:38:32 crc kubenswrapper[4810]: I1201 14:38:32.362899 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 01 14:38:32 crc kubenswrapper[4810]: I1201 14:38:32.393042 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 01 14:38:32 crc kubenswrapper[4810]: I1201 14:38:32.414846 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 01 14:38:32 crc kubenswrapper[4810]: I1201 14:38:32.422143 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 01 14:38:32 crc kubenswrapper[4810]: I1201 14:38:32.506927 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 01 14:38:32 crc kubenswrapper[4810]: I1201 14:38:32.540249 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 01 14:38:32 crc kubenswrapper[4810]: I1201 14:38:32.771350 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 01 14:38:32 crc kubenswrapper[4810]: I1201 14:38:32.782248 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 01 14:38:32 crc kubenswrapper[4810]: I1201 14:38:32.866973 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 01 14:38:32 crc kubenswrapper[4810]: I1201 14:38:32.868399 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 01 14:38:32 crc kubenswrapper[4810]: I1201 14:38:32.958892 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.035085 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.051487 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.077485 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.091518 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.123571 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.175979 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.183800 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.210120 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.225145 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.417016 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.498736 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.529900 4810 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.529988 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.530058 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.530981 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"3657f23abef8beba44444f3b182e43dc58df375c694d65275604a15ce3724277"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.531193 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://3657f23abef8beba44444f3b182e43dc58df375c694d65275604a15ce3724277" gracePeriod=30 Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.554862 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.603383 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.632798 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.751426 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.778825 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.836920 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.932619 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.934791 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.942529 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 01 14:38:33 crc kubenswrapper[4810]: I1201 14:38:33.969087 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 01 14:38:34 crc kubenswrapper[4810]: I1201 14:38:34.016188 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 01 14:38:34 crc kubenswrapper[4810]: I1201 14:38:34.055399 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 01 14:38:34 crc kubenswrapper[4810]: I1201 14:38:34.222921 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 01 14:38:34 crc kubenswrapper[4810]: I1201 14:38:34.247036 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 01 14:38:34 crc kubenswrapper[4810]: I1201 14:38:34.275169 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 01 14:38:34 crc kubenswrapper[4810]: I1201 14:38:34.377239 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 01 14:38:34 crc kubenswrapper[4810]: I1201 14:38:34.461623 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 01 14:38:34 crc kubenswrapper[4810]: I1201 14:38:34.521884 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 01 14:38:34 crc kubenswrapper[4810]: I1201 14:38:34.586203 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 01 14:38:34 crc kubenswrapper[4810]: I1201 14:38:34.662526 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 14:38:34 crc kubenswrapper[4810]: I1201 14:38:34.872301 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 01 14:38:35 crc kubenswrapper[4810]: I1201 14:38:35.054275 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 01 14:38:35 crc kubenswrapper[4810]: I1201 14:38:35.527596 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 01 14:38:35 crc kubenswrapper[4810]: I1201 14:38:35.571266 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 01 14:38:35 crc kubenswrapper[4810]: I1201 14:38:35.639880 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 01 14:38:35 crc kubenswrapper[4810]: I1201 14:38:35.790214 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 01 14:38:35 crc kubenswrapper[4810]: I1201 14:38:35.798135 4810 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 01 14:38:35 crc kubenswrapper[4810]: I1201 14:38:35.802197 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-fdwcc","openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 14:38:35 crc kubenswrapper[4810]: I1201 14:38:35.802259 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 14:38:35 crc kubenswrapper[4810]: I1201 14:38:35.806752 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 14:38:35 crc kubenswrapper[4810]: I1201 14:38:35.823295 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=17.823269108 podStartE2EDuration="17.823269108s" podCreationTimestamp="2025-12-01 14:38:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:38:35.822072755 +0000 UTC m=+281.585582388" watchObservedRunningTime="2025-12-01 14:38:35.823269108 +0000 UTC m=+281.586778711" Dec 01 14:38:35 crc kubenswrapper[4810]: I1201 14:38:35.823794 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 01 14:38:35 crc kubenswrapper[4810]: I1201 14:38:35.947343 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 01 14:38:35 crc kubenswrapper[4810]: I1201 14:38:35.995602 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.029990 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.032133 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.038287 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.039692 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.048587 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.050099 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.105685 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.151800 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.264991 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.293040 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.294490 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.345503 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.496174 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.500030 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca08c59d-eb6c-4c1f-bf44-f161a760ae67" path="/var/lib/kubelet/pods/ca08c59d-eb6c-4c1f-bf44-f161a760ae67/volumes" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.508970 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.577782 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.587802 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.588383 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.590694 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.661295 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.810972 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.811589 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.864639 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.876623 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.897882 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 01 14:38:36 crc kubenswrapper[4810]: I1201 14:38:36.934175 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 14:38:37 crc kubenswrapper[4810]: I1201 14:38:37.073199 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 01 14:38:37 crc kubenswrapper[4810]: I1201 14:38:37.082517 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 01 14:38:37 crc kubenswrapper[4810]: I1201 14:38:37.145159 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 01 14:38:37 crc kubenswrapper[4810]: I1201 14:38:37.267068 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 01 14:38:37 crc kubenswrapper[4810]: I1201 14:38:37.386696 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 01 14:38:37 crc kubenswrapper[4810]: I1201 14:38:37.434408 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 14:38:37 crc kubenswrapper[4810]: I1201 14:38:37.604453 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 01 14:38:37 crc kubenswrapper[4810]: I1201 14:38:37.622766 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 01 14:38:37 crc kubenswrapper[4810]: I1201 14:38:37.732264 4810 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 01 14:38:37 crc kubenswrapper[4810]: I1201 14:38:37.733955 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 01 14:38:37 crc kubenswrapper[4810]: I1201 14:38:37.762671 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 01 14:38:37 crc kubenswrapper[4810]: I1201 14:38:37.806864 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 01 14:38:37 crc kubenswrapper[4810]: I1201 14:38:37.838561 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 01 14:38:37 crc kubenswrapper[4810]: I1201 14:38:37.937560 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.051683 4810 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.142699 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.198257 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.229287 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.345834 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.373073 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.400721 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.437943 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.444614 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.457714 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.526070 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.528308 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.552618 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.591245 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.615825 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.655069 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.710727 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.714549 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.820884 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.835955 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.871495 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.891805 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.924193 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 01 14:38:38 crc kubenswrapper[4810]: I1201 14:38:38.954381 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.071087 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.112355 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.133942 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.233866 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.331060 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.356392 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.435222 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.440643 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.591068 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.604292 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.613364 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.618918 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.692330 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.735245 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.735449 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.768178 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.812075 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.828957 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.899713 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.902912 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.985041 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 01 14:38:39 crc kubenswrapper[4810]: I1201 14:38:39.996752 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.064116 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.099315 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.269841 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.286374 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.293351 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.386956 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.416313 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.416820 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.420763 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.421815 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.462147 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.497772 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.541382 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.593524 4810 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.655960 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.741912 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.754847 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.823809 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.881129 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 01 14:38:40 crc kubenswrapper[4810]: I1201 14:38:40.965255 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.015534 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.038911 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.085457 4810 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.105541 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.158985 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.230184 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.338117 4810 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.338361 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://206bc2a2f9fd203b471578db4666b029bef18c2fbdc71f8a660d0b1f5442b50d" gracePeriod=5 Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.347450 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.355829 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.460230 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.611887 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.621456 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.692118 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.697383 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.765734 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6c495d4986-pxq7h"] Dec 01 14:38:41 crc kubenswrapper[4810]: E1201 14:38:41.765954 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.765967 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 14:38:41 crc kubenswrapper[4810]: E1201 14:38:41.765977 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca08c59d-eb6c-4c1f-bf44-f161a760ae67" containerName="oauth-openshift" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.765986 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca08c59d-eb6c-4c1f-bf44-f161a760ae67" containerName="oauth-openshift" Dec 01 14:38:41 crc kubenswrapper[4810]: E1201 14:38:41.766009 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24394371-0f74-48a7-97c3-315489b78fe9" containerName="installer" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.766017 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="24394371-0f74-48a7-97c3-315489b78fe9" containerName="installer" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.766615 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca08c59d-eb6c-4c1f-bf44-f161a760ae67" containerName="oauth-openshift" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.766645 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.773635 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="24394371-0f74-48a7-97c3-315489b78fe9" containerName="installer" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.775240 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.782974 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.783328 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.783551 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.783781 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.785383 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.790547 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.791993 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.793266 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.793873 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.793947 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.795523 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.795577 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.809385 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.816222 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.825428 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.919893 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.944302 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-session\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.944625 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.944771 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.944935 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.945059 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-user-template-login\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.945196 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-user-template-error\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.945314 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75fb1ba2-187a-493c-95a9-daf956df018e-audit-dir\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.945424 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-router-certs\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.945550 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.945660 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmvd9\" (UniqueName: \"kubernetes.io/projected/75fb1ba2-187a-493c-95a9-daf956df018e-kube-api-access-hmvd9\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.945791 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.945905 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.946039 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-service-ca\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.946160 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/75fb1ba2-187a-493c-95a9-daf956df018e-audit-policies\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.982147 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 01 14:38:41 crc kubenswrapper[4810]: I1201 14:38:41.998272 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.047805 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-session\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.048129 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.049046 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.049168 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.049259 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-user-template-login\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.049345 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-user-template-error\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.049420 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75fb1ba2-187a-493c-95a9-daf956df018e-audit-dir\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.049512 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-router-certs\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.049594 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.049673 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmvd9\" (UniqueName: \"kubernetes.io/projected/75fb1ba2-187a-493c-95a9-daf956df018e-kube-api-access-hmvd9\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.049730 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75fb1ba2-187a-493c-95a9-daf956df018e-audit-dir\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.049755 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.049851 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.049941 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-service-ca\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.050014 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/75fb1ba2-187a-493c-95a9-daf956df018e-audit-policies\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.050146 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.050618 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.050983 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/75fb1ba2-187a-493c-95a9-daf956df018e-audit-policies\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.051196 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-service-ca\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.054013 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.054418 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.054580 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.054638 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-router-certs\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.055790 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-user-template-login\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.056335 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-session\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.059907 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-user-template-error\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.061829 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/75fb1ba2-187a-493c-95a9-daf956df018e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.068634 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmvd9\" (UniqueName: \"kubernetes.io/projected/75fb1ba2-187a-493c-95a9-daf956df018e-kube-api-access-hmvd9\") pod \"oauth-openshift-6c495d4986-pxq7h\" (UID: \"75fb1ba2-187a-493c-95a9-daf956df018e\") " pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.075508 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.102689 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.123210 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.132736 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.179538 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.440621 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.450387 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.519282 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.536316 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.539088 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.682195 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 01 14:38:42 crc kubenswrapper[4810]: I1201 14:38:42.684403 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 01 14:38:43 crc kubenswrapper[4810]: I1201 14:38:43.152310 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 01 14:38:43 crc kubenswrapper[4810]: I1201 14:38:43.174859 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 01 14:38:43 crc kubenswrapper[4810]: I1201 14:38:43.175678 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 01 14:38:43 crc kubenswrapper[4810]: I1201 14:38:43.448543 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 14:38:44 crc kubenswrapper[4810]: I1201 14:38:44.203933 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 01 14:38:44 crc kubenswrapper[4810]: I1201 14:38:44.289454 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 01 14:38:44 crc kubenswrapper[4810]: I1201 14:38:44.295082 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 01 14:38:44 crc kubenswrapper[4810]: I1201 14:38:44.384144 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 01 14:38:44 crc kubenswrapper[4810]: I1201 14:38:44.434808 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 01 14:38:44 crc kubenswrapper[4810]: I1201 14:38:44.557112 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 01 14:38:45 crc kubenswrapper[4810]: I1201 14:38:45.228600 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 01 14:38:46 crc kubenswrapper[4810]: I1201 14:38:46.908717 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 01 14:38:46 crc kubenswrapper[4810]: I1201 14:38:46.908793 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:38:46 crc kubenswrapper[4810]: I1201 14:38:46.931702 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 14:38:46 crc kubenswrapper[4810]: I1201 14:38:46.931749 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 14:38:46 crc kubenswrapper[4810]: I1201 14:38:46.931972 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:38:46 crc kubenswrapper[4810]: I1201 14:38:46.931990 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:38:47 crc kubenswrapper[4810]: I1201 14:38:47.022307 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 01 14:38:47 crc kubenswrapper[4810]: I1201 14:38:47.022370 4810 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="206bc2a2f9fd203b471578db4666b029bef18c2fbdc71f8a660d0b1f5442b50d" exitCode=137 Dec 01 14:38:47 crc kubenswrapper[4810]: I1201 14:38:47.022435 4810 scope.go:117] "RemoveContainer" containerID="206bc2a2f9fd203b471578db4666b029bef18c2fbdc71f8a660d0b1f5442b50d" Dec 01 14:38:47 crc kubenswrapper[4810]: I1201 14:38:47.022595 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 14:38:47 crc kubenswrapper[4810]: I1201 14:38:47.032795 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 14:38:47 crc kubenswrapper[4810]: I1201 14:38:47.032900 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:38:47 crc kubenswrapper[4810]: I1201 14:38:47.032919 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 14:38:47 crc kubenswrapper[4810]: I1201 14:38:47.033020 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:38:47 crc kubenswrapper[4810]: I1201 14:38:47.033148 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 14:38:47 crc kubenswrapper[4810]: I1201 14:38:47.033929 4810 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:47 crc kubenswrapper[4810]: I1201 14:38:47.033964 4810 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:47 crc kubenswrapper[4810]: I1201 14:38:47.033976 4810 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:47 crc kubenswrapper[4810]: I1201 14:38:47.033990 4810 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:47 crc kubenswrapper[4810]: I1201 14:38:47.046303 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:38:47 crc kubenswrapper[4810]: I1201 14:38:47.051878 4810 scope.go:117] "RemoveContainer" containerID="206bc2a2f9fd203b471578db4666b029bef18c2fbdc71f8a660d0b1f5442b50d" Dec 01 14:38:47 crc kubenswrapper[4810]: E1201 14:38:47.052803 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"206bc2a2f9fd203b471578db4666b029bef18c2fbdc71f8a660d0b1f5442b50d\": container with ID starting with 206bc2a2f9fd203b471578db4666b029bef18c2fbdc71f8a660d0b1f5442b50d not found: ID does not exist" containerID="206bc2a2f9fd203b471578db4666b029bef18c2fbdc71f8a660d0b1f5442b50d" Dec 01 14:38:47 crc kubenswrapper[4810]: I1201 14:38:47.052858 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"206bc2a2f9fd203b471578db4666b029bef18c2fbdc71f8a660d0b1f5442b50d"} err="failed to get container status \"206bc2a2f9fd203b471578db4666b029bef18c2fbdc71f8a660d0b1f5442b50d\": rpc error: code = NotFound desc = could not find container \"206bc2a2f9fd203b471578db4666b029bef18c2fbdc71f8a660d0b1f5442b50d\": container with ID starting with 206bc2a2f9fd203b471578db4666b029bef18c2fbdc71f8a660d0b1f5442b50d not found: ID does not exist" Dec 01 14:38:47 crc kubenswrapper[4810]: I1201 14:38:47.134666 4810 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 14:38:48 crc kubenswrapper[4810]: I1201 14:38:48.503004 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 01 14:38:57 crc kubenswrapper[4810]: I1201 14:38:57.816849 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 01 14:38:59 crc kubenswrapper[4810]: I1201 14:38:59.919239 4810 generic.go:334] "Generic (PLEG): container finished" podID="0fa71e97-6be0-4300-bef2-b2dffbf30076" containerID="987552bf4a9cd3bc3b0eb15466951f1c7575b07dd03b1505f1f8e6c0d30c382e" exitCode=0 Dec 01 14:38:59 crc kubenswrapper[4810]: I1201 14:38:59.919391 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" event={"ID":"0fa71e97-6be0-4300-bef2-b2dffbf30076","Type":"ContainerDied","Data":"987552bf4a9cd3bc3b0eb15466951f1c7575b07dd03b1505f1f8e6c0d30c382e"} Dec 01 14:38:59 crc kubenswrapper[4810]: I1201 14:38:59.920764 4810 scope.go:117] "RemoveContainer" containerID="987552bf4a9cd3bc3b0eb15466951f1c7575b07dd03b1505f1f8e6c0d30c382e" Dec 01 14:39:00 crc kubenswrapper[4810]: I1201 14:39:00.933622 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" event={"ID":"0fa71e97-6be0-4300-bef2-b2dffbf30076","Type":"ContainerStarted","Data":"c9c96013698eda0a27436b9bb842d6d859dced3ff353666d8f3d9dbcd7b2c0c3"} Dec 01 14:39:00 crc kubenswrapper[4810]: I1201 14:39:00.934815 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" Dec 01 14:39:00 crc kubenswrapper[4810]: I1201 14:39:00.936734 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" Dec 01 14:39:01 crc kubenswrapper[4810]: I1201 14:39:01.345397 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 01 14:39:03 crc kubenswrapper[4810]: I1201 14:39:03.619741 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 01 14:39:03 crc kubenswrapper[4810]: I1201 14:39:03.918430 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 01 14:39:03 crc kubenswrapper[4810]: I1201 14:39:03.955731 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 01 14:39:03 crc kubenswrapper[4810]: I1201 14:39:03.957164 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 01 14:39:03 crc kubenswrapper[4810]: I1201 14:39:03.957209 4810 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="3657f23abef8beba44444f3b182e43dc58df375c694d65275604a15ce3724277" exitCode=137 Dec 01 14:39:03 crc kubenswrapper[4810]: I1201 14:39:03.957238 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"3657f23abef8beba44444f3b182e43dc58df375c694d65275604a15ce3724277"} Dec 01 14:39:03 crc kubenswrapper[4810]: I1201 14:39:03.957268 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"07165629c122b65b1101038a210e84c57c442cb084fb6e0d85507f73373db1d6"} Dec 01 14:39:03 crc kubenswrapper[4810]: I1201 14:39:03.957286 4810 scope.go:117] "RemoveContainer" containerID="7285ef3168d24a80bb90d52ae55055263d227e1d40cac6d8f44c9113d900277e" Dec 01 14:39:04 crc kubenswrapper[4810]: I1201 14:39:04.966802 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 01 14:39:06 crc kubenswrapper[4810]: I1201 14:39:06.164300 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 01 14:39:07 crc kubenswrapper[4810]: I1201 14:39:07.301451 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 01 14:39:07 crc kubenswrapper[4810]: I1201 14:39:07.519433 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 01 14:39:08 crc kubenswrapper[4810]: I1201 14:39:08.109776 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 01 14:39:08 crc kubenswrapper[4810]: I1201 14:39:08.536214 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:39:10 crc kubenswrapper[4810]: I1201 14:39:10.311659 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 01 14:39:13 crc kubenswrapper[4810]: I1201 14:39:13.529567 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:39:13 crc kubenswrapper[4810]: I1201 14:39:13.534491 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:39:13 crc kubenswrapper[4810]: I1201 14:39:13.921785 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 01 14:39:13 crc kubenswrapper[4810]: I1201 14:39:13.960428 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 01 14:39:14 crc kubenswrapper[4810]: I1201 14:39:14.005441 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 14:39:14 crc kubenswrapper[4810]: I1201 14:39:14.026036 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 14:39:14 crc kubenswrapper[4810]: I1201 14:39:14.474370 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 01 14:39:14 crc kubenswrapper[4810]: I1201 14:39:14.996387 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 01 14:39:16 crc kubenswrapper[4810]: I1201 14:39:16.518361 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6c495d4986-pxq7h"] Dec 01 14:39:16 crc kubenswrapper[4810]: I1201 14:39:16.905854 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6c495d4986-pxq7h"] Dec 01 14:39:17 crc kubenswrapper[4810]: I1201 14:39:17.037262 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" event={"ID":"75fb1ba2-187a-493c-95a9-daf956df018e","Type":"ContainerStarted","Data":"1e6c8b6d01712fec9a6df092709c611c617a19b8a861e701df7539b4c4fed0ee"} Dec 01 14:39:18 crc kubenswrapper[4810]: I1201 14:39:18.044607 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" event={"ID":"75fb1ba2-187a-493c-95a9-daf956df018e","Type":"ContainerStarted","Data":"28db8cc43ccc3848a96a4558e7e170c6b6214739984cec34a5e616636f5af17c"} Dec 01 14:39:18 crc kubenswrapper[4810]: I1201 14:39:18.045006 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:39:18 crc kubenswrapper[4810]: I1201 14:39:18.049935 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" Dec 01 14:39:18 crc kubenswrapper[4810]: I1201 14:39:18.082080 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6c495d4986-pxq7h" podStartSLOduration=96.08205347 podStartE2EDuration="1m36.08205347s" podCreationTimestamp="2025-12-01 14:37:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:39:18.068070078 +0000 UTC m=+323.831579731" watchObservedRunningTime="2025-12-01 14:39:18.08205347 +0000 UTC m=+323.845563083" Dec 01 14:39:19 crc kubenswrapper[4810]: I1201 14:39:19.035899 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 01 14:39:24 crc kubenswrapper[4810]: I1201 14:39:24.680701 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.025063 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4rgpg"] Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.025526 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" podUID="ef1bca29-62d7-422b-890a-f93b965c0934" containerName="controller-manager" containerID="cri-o://2b6abaae006db7ef297be63c129a23773a98f62a25b7d76c02e9e200644030d6" gracePeriod=30 Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.126967 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4"] Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.127300 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" podUID="e9c93c51-49ca-4557-a26b-0ba162d4dbc2" containerName="route-controller-manager" containerID="cri-o://f7f9e80257653a1377d5ae9754366811b70c4ea7dc9f400f31b92a359a29a2ad" gracePeriod=30 Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.422147 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.508904 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.573530 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ef1bca29-62d7-422b-890a-f93b965c0934-client-ca\") pod \"ef1bca29-62d7-422b-890a-f93b965c0934\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.573646 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef1bca29-62d7-422b-890a-f93b965c0934-config\") pod \"ef1bca29-62d7-422b-890a-f93b965c0934\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.573757 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2zhn\" (UniqueName: \"kubernetes.io/projected/ef1bca29-62d7-422b-890a-f93b965c0934-kube-api-access-x2zhn\") pod \"ef1bca29-62d7-422b-890a-f93b965c0934\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.573786 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef1bca29-62d7-422b-890a-f93b965c0934-serving-cert\") pod \"ef1bca29-62d7-422b-890a-f93b965c0934\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.573811 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ef1bca29-62d7-422b-890a-f93b965c0934-proxy-ca-bundles\") pod \"ef1bca29-62d7-422b-890a-f93b965c0934\" (UID: \"ef1bca29-62d7-422b-890a-f93b965c0934\") " Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.574853 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef1bca29-62d7-422b-890a-f93b965c0934-client-ca" (OuterVolumeSpecName: "client-ca") pod "ef1bca29-62d7-422b-890a-f93b965c0934" (UID: "ef1bca29-62d7-422b-890a-f93b965c0934"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.574861 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef1bca29-62d7-422b-890a-f93b965c0934-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ef1bca29-62d7-422b-890a-f93b965c0934" (UID: "ef1bca29-62d7-422b-890a-f93b965c0934"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.575111 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef1bca29-62d7-422b-890a-f93b965c0934-config" (OuterVolumeSpecName: "config") pod "ef1bca29-62d7-422b-890a-f93b965c0934" (UID: "ef1bca29-62d7-422b-890a-f93b965c0934"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.580359 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef1bca29-62d7-422b-890a-f93b965c0934-kube-api-access-x2zhn" (OuterVolumeSpecName: "kube-api-access-x2zhn") pod "ef1bca29-62d7-422b-890a-f93b965c0934" (UID: "ef1bca29-62d7-422b-890a-f93b965c0934"). InnerVolumeSpecName "kube-api-access-x2zhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.580738 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef1bca29-62d7-422b-890a-f93b965c0934-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ef1bca29-62d7-422b-890a-f93b965c0934" (UID: "ef1bca29-62d7-422b-890a-f93b965c0934"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.675788 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-config\") pod \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\" (UID: \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\") " Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.676070 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-serving-cert\") pod \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\" (UID: \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\") " Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.676117 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr6t8\" (UniqueName: \"kubernetes.io/projected/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-kube-api-access-sr6t8\") pod \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\" (UID: \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\") " Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.676228 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-client-ca\") pod \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\" (UID: \"e9c93c51-49ca-4557-a26b-0ba162d4dbc2\") " Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.676732 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef1bca29-62d7-422b-890a-f93b965c0934-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.676772 4810 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ef1bca29-62d7-422b-890a-f93b965c0934-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.676804 4810 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ef1bca29-62d7-422b-890a-f93b965c0934-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.676829 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef1bca29-62d7-422b-890a-f93b965c0934-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.676856 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2zhn\" (UniqueName: \"kubernetes.io/projected/ef1bca29-62d7-422b-890a-f93b965c0934-kube-api-access-x2zhn\") on node \"crc\" DevicePath \"\"" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.678762 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-client-ca" (OuterVolumeSpecName: "client-ca") pod "e9c93c51-49ca-4557-a26b-0ba162d4dbc2" (UID: "e9c93c51-49ca-4557-a26b-0ba162d4dbc2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.679866 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-config" (OuterVolumeSpecName: "config") pod "e9c93c51-49ca-4557-a26b-0ba162d4dbc2" (UID: "e9c93c51-49ca-4557-a26b-0ba162d4dbc2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.682362 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-kube-api-access-sr6t8" (OuterVolumeSpecName: "kube-api-access-sr6t8") pod "e9c93c51-49ca-4557-a26b-0ba162d4dbc2" (UID: "e9c93c51-49ca-4557-a26b-0ba162d4dbc2"). InnerVolumeSpecName "kube-api-access-sr6t8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.682819 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e9c93c51-49ca-4557-a26b-0ba162d4dbc2" (UID: "e9c93c51-49ca-4557-a26b-0ba162d4dbc2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.778759 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.778796 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr6t8\" (UniqueName: \"kubernetes.io/projected/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-kube-api-access-sr6t8\") on node \"crc\" DevicePath \"\"" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.778807 4810 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:39:25 crc kubenswrapper[4810]: I1201 14:39:25.778816 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9c93c51-49ca-4557-a26b-0ba162d4dbc2-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.101328 4810 generic.go:334] "Generic (PLEG): container finished" podID="e9c93c51-49ca-4557-a26b-0ba162d4dbc2" containerID="f7f9e80257653a1377d5ae9754366811b70c4ea7dc9f400f31b92a359a29a2ad" exitCode=0 Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.101395 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" event={"ID":"e9c93c51-49ca-4557-a26b-0ba162d4dbc2","Type":"ContainerDied","Data":"f7f9e80257653a1377d5ae9754366811b70c4ea7dc9f400f31b92a359a29a2ad"} Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.101671 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" event={"ID":"e9c93c51-49ca-4557-a26b-0ba162d4dbc2","Type":"ContainerDied","Data":"373ee90aacb1d2f9d9c242b48cf4b84d90dedb67cea3df6ad832c47d82b302ea"} Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.101691 4810 scope.go:117] "RemoveContainer" containerID="f7f9e80257653a1377d5ae9754366811b70c4ea7dc9f400f31b92a359a29a2ad" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.101419 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.108168 4810 generic.go:334] "Generic (PLEG): container finished" podID="ef1bca29-62d7-422b-890a-f93b965c0934" containerID="2b6abaae006db7ef297be63c129a23773a98f62a25b7d76c02e9e200644030d6" exitCode=0 Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.108219 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" event={"ID":"ef1bca29-62d7-422b-890a-f93b965c0934","Type":"ContainerDied","Data":"2b6abaae006db7ef297be63c129a23773a98f62a25b7d76c02e9e200644030d6"} Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.108255 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" event={"ID":"ef1bca29-62d7-422b-890a-f93b965c0934","Type":"ContainerDied","Data":"83813c5d6c40de9cc3baf26c9ba2b04db41e81679a455d033890fab431352808"} Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.108321 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4rgpg" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.125446 4810 scope.go:117] "RemoveContainer" containerID="f7f9e80257653a1377d5ae9754366811b70c4ea7dc9f400f31b92a359a29a2ad" Dec 01 14:39:26 crc kubenswrapper[4810]: E1201 14:39:26.128436 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7f9e80257653a1377d5ae9754366811b70c4ea7dc9f400f31b92a359a29a2ad\": container with ID starting with f7f9e80257653a1377d5ae9754366811b70c4ea7dc9f400f31b92a359a29a2ad not found: ID does not exist" containerID="f7f9e80257653a1377d5ae9754366811b70c4ea7dc9f400f31b92a359a29a2ad" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.128483 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7f9e80257653a1377d5ae9754366811b70c4ea7dc9f400f31b92a359a29a2ad"} err="failed to get container status \"f7f9e80257653a1377d5ae9754366811b70c4ea7dc9f400f31b92a359a29a2ad\": rpc error: code = NotFound desc = could not find container \"f7f9e80257653a1377d5ae9754366811b70c4ea7dc9f400f31b92a359a29a2ad\": container with ID starting with f7f9e80257653a1377d5ae9754366811b70c4ea7dc9f400f31b92a359a29a2ad not found: ID does not exist" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.128504 4810 scope.go:117] "RemoveContainer" containerID="2b6abaae006db7ef297be63c129a23773a98f62a25b7d76c02e9e200644030d6" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.132141 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4"] Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.135432 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gn5f4"] Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.146775 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4rgpg"] Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.148427 4810 scope.go:117] "RemoveContainer" containerID="2b6abaae006db7ef297be63c129a23773a98f62a25b7d76c02e9e200644030d6" Dec 01 14:39:26 crc kubenswrapper[4810]: E1201 14:39:26.149143 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b6abaae006db7ef297be63c129a23773a98f62a25b7d76c02e9e200644030d6\": container with ID starting with 2b6abaae006db7ef297be63c129a23773a98f62a25b7d76c02e9e200644030d6 not found: ID does not exist" containerID="2b6abaae006db7ef297be63c129a23773a98f62a25b7d76c02e9e200644030d6" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.149182 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b6abaae006db7ef297be63c129a23773a98f62a25b7d76c02e9e200644030d6"} err="failed to get container status \"2b6abaae006db7ef297be63c129a23773a98f62a25b7d76c02e9e200644030d6\": rpc error: code = NotFound desc = could not find container \"2b6abaae006db7ef297be63c129a23773a98f62a25b7d76c02e9e200644030d6\": container with ID starting with 2b6abaae006db7ef297be63c129a23773a98f62a25b7d76c02e9e200644030d6 not found: ID does not exist" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.150940 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4rgpg"] Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.499136 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9c93c51-49ca-4557-a26b-0ba162d4dbc2" path="/var/lib/kubelet/pods/e9c93c51-49ca-4557-a26b-0ba162d4dbc2/volumes" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.499682 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef1bca29-62d7-422b-890a-f93b965c0934" path="/var/lib/kubelet/pods/ef1bca29-62d7-422b-890a-f93b965c0934/volumes" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.667999 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-589f9f9656-s8r98"] Dec 01 14:39:26 crc kubenswrapper[4810]: E1201 14:39:26.668289 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef1bca29-62d7-422b-890a-f93b965c0934" containerName="controller-manager" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.668317 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef1bca29-62d7-422b-890a-f93b965c0934" containerName="controller-manager" Dec 01 14:39:26 crc kubenswrapper[4810]: E1201 14:39:26.668334 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9c93c51-49ca-4557-a26b-0ba162d4dbc2" containerName="route-controller-manager" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.668342 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9c93c51-49ca-4557-a26b-0ba162d4dbc2" containerName="route-controller-manager" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.668435 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9c93c51-49ca-4557-a26b-0ba162d4dbc2" containerName="route-controller-manager" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.668453 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef1bca29-62d7-422b-890a-f93b965c0934" containerName="controller-manager" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.668900 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.674059 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.674674 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.675214 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.675539 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.676577 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.676743 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.678384 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j"] Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.682283 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.691778 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.693783 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.694083 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.694157 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.694174 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.695704 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.700029 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-589f9f9656-s8r98"] Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.700992 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.709515 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j"] Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.791751 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3e26cbeb-782a-4b96-ada3-edb4534c254c-client-ca\") pod \"route-controller-manager-7c6b77d5f7-knd4j\" (UID: \"3e26cbeb-782a-4b96-ada3-edb4534c254c\") " pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.791839 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c62b420-8bd8-4dda-a43c-7b940514e381-proxy-ca-bundles\") pod \"controller-manager-589f9f9656-s8r98\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.791867 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncqxb\" (UniqueName: \"kubernetes.io/projected/9c62b420-8bd8-4dda-a43c-7b940514e381-kube-api-access-ncqxb\") pod \"controller-manager-589f9f9656-s8r98\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.791899 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qs67h\" (UniqueName: \"kubernetes.io/projected/3e26cbeb-782a-4b96-ada3-edb4534c254c-kube-api-access-qs67h\") pod \"route-controller-manager-7c6b77d5f7-knd4j\" (UID: \"3e26cbeb-782a-4b96-ada3-edb4534c254c\") " pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.791940 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c62b420-8bd8-4dda-a43c-7b940514e381-client-ca\") pod \"controller-manager-589f9f9656-s8r98\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.791970 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e26cbeb-782a-4b96-ada3-edb4534c254c-serving-cert\") pod \"route-controller-manager-7c6b77d5f7-knd4j\" (UID: \"3e26cbeb-782a-4b96-ada3-edb4534c254c\") " pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.791991 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c62b420-8bd8-4dda-a43c-7b940514e381-config\") pod \"controller-manager-589f9f9656-s8r98\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.792025 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e26cbeb-782a-4b96-ada3-edb4534c254c-config\") pod \"route-controller-manager-7c6b77d5f7-knd4j\" (UID: \"3e26cbeb-782a-4b96-ada3-edb4534c254c\") " pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.792076 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c62b420-8bd8-4dda-a43c-7b940514e381-serving-cert\") pod \"controller-manager-589f9f9656-s8r98\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.893215 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3e26cbeb-782a-4b96-ada3-edb4534c254c-client-ca\") pod \"route-controller-manager-7c6b77d5f7-knd4j\" (UID: \"3e26cbeb-782a-4b96-ada3-edb4534c254c\") " pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.893354 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c62b420-8bd8-4dda-a43c-7b940514e381-proxy-ca-bundles\") pod \"controller-manager-589f9f9656-s8r98\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.893380 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncqxb\" (UniqueName: \"kubernetes.io/projected/9c62b420-8bd8-4dda-a43c-7b940514e381-kube-api-access-ncqxb\") pod \"controller-manager-589f9f9656-s8r98\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.893426 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qs67h\" (UniqueName: \"kubernetes.io/projected/3e26cbeb-782a-4b96-ada3-edb4534c254c-kube-api-access-qs67h\") pod \"route-controller-manager-7c6b77d5f7-knd4j\" (UID: \"3e26cbeb-782a-4b96-ada3-edb4534c254c\") " pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.893460 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c62b420-8bd8-4dda-a43c-7b940514e381-client-ca\") pod \"controller-manager-589f9f9656-s8r98\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.893516 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e26cbeb-782a-4b96-ada3-edb4534c254c-serving-cert\") pod \"route-controller-manager-7c6b77d5f7-knd4j\" (UID: \"3e26cbeb-782a-4b96-ada3-edb4534c254c\") " pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.893535 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c62b420-8bd8-4dda-a43c-7b940514e381-config\") pod \"controller-manager-589f9f9656-s8r98\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.893585 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e26cbeb-782a-4b96-ada3-edb4534c254c-config\") pod \"route-controller-manager-7c6b77d5f7-knd4j\" (UID: \"3e26cbeb-782a-4b96-ada3-edb4534c254c\") " pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.893609 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c62b420-8bd8-4dda-a43c-7b940514e381-serving-cert\") pod \"controller-manager-589f9f9656-s8r98\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.894374 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3e26cbeb-782a-4b96-ada3-edb4534c254c-client-ca\") pod \"route-controller-manager-7c6b77d5f7-knd4j\" (UID: \"3e26cbeb-782a-4b96-ada3-edb4534c254c\") " pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.894574 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c62b420-8bd8-4dda-a43c-7b940514e381-proxy-ca-bundles\") pod \"controller-manager-589f9f9656-s8r98\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.895398 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c62b420-8bd8-4dda-a43c-7b940514e381-config\") pod \"controller-manager-589f9f9656-s8r98\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.896037 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c62b420-8bd8-4dda-a43c-7b940514e381-client-ca\") pod \"controller-manager-589f9f9656-s8r98\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.896654 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e26cbeb-782a-4b96-ada3-edb4534c254c-config\") pod \"route-controller-manager-7c6b77d5f7-knd4j\" (UID: \"3e26cbeb-782a-4b96-ada3-edb4534c254c\") " pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.899444 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e26cbeb-782a-4b96-ada3-edb4534c254c-serving-cert\") pod \"route-controller-manager-7c6b77d5f7-knd4j\" (UID: \"3e26cbeb-782a-4b96-ada3-edb4534c254c\") " pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.899710 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c62b420-8bd8-4dda-a43c-7b940514e381-serving-cert\") pod \"controller-manager-589f9f9656-s8r98\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.913108 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncqxb\" (UniqueName: \"kubernetes.io/projected/9c62b420-8bd8-4dda-a43c-7b940514e381-kube-api-access-ncqxb\") pod \"controller-manager-589f9f9656-s8r98\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.917567 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qs67h\" (UniqueName: \"kubernetes.io/projected/3e26cbeb-782a-4b96-ada3-edb4534c254c-kube-api-access-qs67h\") pod \"route-controller-manager-7c6b77d5f7-knd4j\" (UID: \"3e26cbeb-782a-4b96-ada3-edb4534c254c\") " pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" Dec 01 14:39:26 crc kubenswrapper[4810]: I1201 14:39:26.997427 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:27 crc kubenswrapper[4810]: I1201 14:39:27.011320 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" Dec 01 14:39:27 crc kubenswrapper[4810]: I1201 14:39:27.326366 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j"] Dec 01 14:39:27 crc kubenswrapper[4810]: I1201 14:39:27.365582 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-589f9f9656-s8r98"] Dec 01 14:39:27 crc kubenswrapper[4810]: W1201 14:39:27.378601 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c62b420_8bd8_4dda_a43c_7b940514e381.slice/crio-1abf7f46d8a96a5fd9f1b95d323ee1853c30a3f58a5864fdd4e56e81a9c02498 WatchSource:0}: Error finding container 1abf7f46d8a96a5fd9f1b95d323ee1853c30a3f58a5864fdd4e56e81a9c02498: Status 404 returned error can't find the container with id 1abf7f46d8a96a5fd9f1b95d323ee1853c30a3f58a5864fdd4e56e81a9c02498 Dec 01 14:39:28 crc kubenswrapper[4810]: I1201 14:39:28.124390 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" event={"ID":"3e26cbeb-782a-4b96-ada3-edb4534c254c","Type":"ContainerStarted","Data":"ebb154db900be4067f2fdd22ee5a32ba4cc3ed726f610f3eb7d543c013694b3d"} Dec 01 14:39:28 crc kubenswrapper[4810]: I1201 14:39:28.125034 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" Dec 01 14:39:28 crc kubenswrapper[4810]: I1201 14:39:28.125062 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" event={"ID":"3e26cbeb-782a-4b96-ada3-edb4534c254c","Type":"ContainerStarted","Data":"362ca02e0a5dc16b24c3b3ccda4853abffe4f46dfd396d612243112ad6cdb190"} Dec 01 14:39:28 crc kubenswrapper[4810]: I1201 14:39:28.127299 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" event={"ID":"9c62b420-8bd8-4dda-a43c-7b940514e381","Type":"ContainerStarted","Data":"3e14d97c1e8f32697118b7b3ed4c298bd121da33fc88d39b2874f9b4a64d6bd4"} Dec 01 14:39:28 crc kubenswrapper[4810]: I1201 14:39:28.127369 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" event={"ID":"9c62b420-8bd8-4dda-a43c-7b940514e381","Type":"ContainerStarted","Data":"1abf7f46d8a96a5fd9f1b95d323ee1853c30a3f58a5864fdd4e56e81a9c02498"} Dec 01 14:39:28 crc kubenswrapper[4810]: I1201 14:39:28.127615 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:28 crc kubenswrapper[4810]: I1201 14:39:28.132336 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" Dec 01 14:39:28 crc kubenswrapper[4810]: I1201 14:39:28.135102 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:39:28 crc kubenswrapper[4810]: I1201 14:39:28.145601 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7c6b77d5f7-knd4j" podStartSLOduration=3.145576211 podStartE2EDuration="3.145576211s" podCreationTimestamp="2025-12-01 14:39:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:39:28.140635313 +0000 UTC m=+333.904144916" watchObservedRunningTime="2025-12-01 14:39:28.145576211 +0000 UTC m=+333.909085814" Dec 01 14:39:28 crc kubenswrapper[4810]: I1201 14:39:28.156995 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" podStartSLOduration=3.156976419 podStartE2EDuration="3.156976419s" podCreationTimestamp="2025-12-01 14:39:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:39:28.155624592 +0000 UTC m=+333.919134195" watchObservedRunningTime="2025-12-01 14:39:28.156976419 +0000 UTC m=+333.920486012" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.434773 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-q8psn"] Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.436455 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.446620 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-q8psn"] Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.546663 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5b96578-cba7-4ff8-85d1-2647a39ffa31-trusted-ca\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.546727 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5b96578-cba7-4ff8-85d1-2647a39ffa31-bound-sa-token\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.546803 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.546835 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d5b96578-cba7-4ff8-85d1-2647a39ffa31-registry-tls\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.546852 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d5b96578-cba7-4ff8-85d1-2647a39ffa31-ca-trust-extracted\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.546879 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d5b96578-cba7-4ff8-85d1-2647a39ffa31-installation-pull-secrets\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.546896 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cllkr\" (UniqueName: \"kubernetes.io/projected/d5b96578-cba7-4ff8-85d1-2647a39ffa31-kube-api-access-cllkr\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.546938 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d5b96578-cba7-4ff8-85d1-2647a39ffa31-registry-certificates\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.570763 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.648625 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d5b96578-cba7-4ff8-85d1-2647a39ffa31-registry-certificates\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.648689 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5b96578-cba7-4ff8-85d1-2647a39ffa31-trusted-ca\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.648723 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5b96578-cba7-4ff8-85d1-2647a39ffa31-bound-sa-token\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.648758 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d5b96578-cba7-4ff8-85d1-2647a39ffa31-registry-tls\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.648772 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d5b96578-cba7-4ff8-85d1-2647a39ffa31-ca-trust-extracted\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.648800 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d5b96578-cba7-4ff8-85d1-2647a39ffa31-installation-pull-secrets\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.648815 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cllkr\" (UniqueName: \"kubernetes.io/projected/d5b96578-cba7-4ff8-85d1-2647a39ffa31-kube-api-access-cllkr\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.649642 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d5b96578-cba7-4ff8-85d1-2647a39ffa31-ca-trust-extracted\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.650322 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d5b96578-cba7-4ff8-85d1-2647a39ffa31-registry-certificates\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.650387 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d5b96578-cba7-4ff8-85d1-2647a39ffa31-trusted-ca\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.656524 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d5b96578-cba7-4ff8-85d1-2647a39ffa31-registry-tls\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.668715 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d5b96578-cba7-4ff8-85d1-2647a39ffa31-installation-pull-secrets\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.679259 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d5b96578-cba7-4ff8-85d1-2647a39ffa31-bound-sa-token\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.679510 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cllkr\" (UniqueName: \"kubernetes.io/projected/d5b96578-cba7-4ff8-85d1-2647a39ffa31-kube-api-access-cllkr\") pod \"image-registry-66df7c8f76-q8psn\" (UID: \"d5b96578-cba7-4ff8-85d1-2647a39ffa31\") " pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:58 crc kubenswrapper[4810]: I1201 14:39:58.753254 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:39:59 crc kubenswrapper[4810]: I1201 14:39:59.160354 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-q8psn"] Dec 01 14:39:59 crc kubenswrapper[4810]: W1201 14:39:59.172494 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5b96578_cba7_4ff8_85d1_2647a39ffa31.slice/crio-79112db269910ea637959f85a5c5212d75d698e37a0a165f7d6262254a60fa38 WatchSource:0}: Error finding container 79112db269910ea637959f85a5c5212d75d698e37a0a165f7d6262254a60fa38: Status 404 returned error can't find the container with id 79112db269910ea637959f85a5c5212d75d698e37a0a165f7d6262254a60fa38 Dec 01 14:39:59 crc kubenswrapper[4810]: I1201 14:39:59.326515 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" event={"ID":"d5b96578-cba7-4ff8-85d1-2647a39ffa31","Type":"ContainerStarted","Data":"9e869cdec1ca0e7b17be2a91937a5eb1067cc9de0c1da50ce8489057196bd965"} Dec 01 14:39:59 crc kubenswrapper[4810]: I1201 14:39:59.326770 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" event={"ID":"d5b96578-cba7-4ff8-85d1-2647a39ffa31","Type":"ContainerStarted","Data":"79112db269910ea637959f85a5c5212d75d698e37a0a165f7d6262254a60fa38"} Dec 01 14:39:59 crc kubenswrapper[4810]: I1201 14:39:59.326786 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:40:02 crc kubenswrapper[4810]: I1201 14:40:02.972557 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:40:02 crc kubenswrapper[4810]: I1201 14:40:02.972887 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.012514 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" podStartSLOduration=7.012484097 podStartE2EDuration="7.012484097s" podCreationTimestamp="2025-12-01 14:39:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:39:59.351727429 +0000 UTC m=+365.115237032" watchObservedRunningTime="2025-12-01 14:40:05.012484097 +0000 UTC m=+370.775993700" Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.019048 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-589f9f9656-s8r98"] Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.019738 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" podUID="9c62b420-8bd8-4dda-a43c-7b940514e381" containerName="controller-manager" containerID="cri-o://3e14d97c1e8f32697118b7b3ed4c298bd121da33fc88d39b2874f9b4a64d6bd4" gracePeriod=30 Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.357683 4810 generic.go:334] "Generic (PLEG): container finished" podID="9c62b420-8bd8-4dda-a43c-7b940514e381" containerID="3e14d97c1e8f32697118b7b3ed4c298bd121da33fc88d39b2874f9b4a64d6bd4" exitCode=0 Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.358013 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" event={"ID":"9c62b420-8bd8-4dda-a43c-7b940514e381","Type":"ContainerDied","Data":"3e14d97c1e8f32697118b7b3ed4c298bd121da33fc88d39b2874f9b4a64d6bd4"} Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.392390 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.555179 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c62b420-8bd8-4dda-a43c-7b940514e381-client-ca\") pod \"9c62b420-8bd8-4dda-a43c-7b940514e381\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.555341 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c62b420-8bd8-4dda-a43c-7b940514e381-config\") pod \"9c62b420-8bd8-4dda-a43c-7b940514e381\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.555677 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c62b420-8bd8-4dda-a43c-7b940514e381-proxy-ca-bundles\") pod \"9c62b420-8bd8-4dda-a43c-7b940514e381\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.555754 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c62b420-8bd8-4dda-a43c-7b940514e381-serving-cert\") pod \"9c62b420-8bd8-4dda-a43c-7b940514e381\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.555781 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncqxb\" (UniqueName: \"kubernetes.io/projected/9c62b420-8bd8-4dda-a43c-7b940514e381-kube-api-access-ncqxb\") pod \"9c62b420-8bd8-4dda-a43c-7b940514e381\" (UID: \"9c62b420-8bd8-4dda-a43c-7b940514e381\") " Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.556655 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c62b420-8bd8-4dda-a43c-7b940514e381-client-ca" (OuterVolumeSpecName: "client-ca") pod "9c62b420-8bd8-4dda-a43c-7b940514e381" (UID: "9c62b420-8bd8-4dda-a43c-7b940514e381"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.556709 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c62b420-8bd8-4dda-a43c-7b940514e381-config" (OuterVolumeSpecName: "config") pod "9c62b420-8bd8-4dda-a43c-7b940514e381" (UID: "9c62b420-8bd8-4dda-a43c-7b940514e381"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.556988 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c62b420-8bd8-4dda-a43c-7b940514e381-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9c62b420-8bd8-4dda-a43c-7b940514e381" (UID: "9c62b420-8bd8-4dda-a43c-7b940514e381"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.561605 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c62b420-8bd8-4dda-a43c-7b940514e381-kube-api-access-ncqxb" (OuterVolumeSpecName: "kube-api-access-ncqxb") pod "9c62b420-8bd8-4dda-a43c-7b940514e381" (UID: "9c62b420-8bd8-4dda-a43c-7b940514e381"). InnerVolumeSpecName "kube-api-access-ncqxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.561687 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c62b420-8bd8-4dda-a43c-7b940514e381-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9c62b420-8bd8-4dda-a43c-7b940514e381" (UID: "9c62b420-8bd8-4dda-a43c-7b940514e381"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.657093 4810 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9c62b420-8bd8-4dda-a43c-7b940514e381-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.657149 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c62b420-8bd8-4dda-a43c-7b940514e381-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.657166 4810 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9c62b420-8bd8-4dda-a43c-7b940514e381-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.657179 4810 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c62b420-8bd8-4dda-a43c-7b940514e381-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:05 crc kubenswrapper[4810]: I1201 14:40:05.657193 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncqxb\" (UniqueName: \"kubernetes.io/projected/9c62b420-8bd8-4dda-a43c-7b940514e381-kube-api-access-ncqxb\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.366163 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" event={"ID":"9c62b420-8bd8-4dda-a43c-7b940514e381","Type":"ContainerDied","Data":"1abf7f46d8a96a5fd9f1b95d323ee1853c30a3f58a5864fdd4e56e81a9c02498"} Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.366507 4810 scope.go:117] "RemoveContainer" containerID="3e14d97c1e8f32697118b7b3ed4c298bd121da33fc88d39b2874f9b4a64d6bd4" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.366623 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-589f9f9656-s8r98" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.403612 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-589f9f9656-s8r98"] Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.406701 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-589f9f9656-s8r98"] Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.499313 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c62b420-8bd8-4dda-a43c-7b940514e381" path="/var/lib/kubelet/pods/9c62b420-8bd8-4dda-a43c-7b940514e381/volumes" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.711157 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-64df456547-z2ngz"] Dec 01 14:40:06 crc kubenswrapper[4810]: E1201 14:40:06.711853 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c62b420-8bd8-4dda-a43c-7b940514e381" containerName="controller-manager" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.711867 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c62b420-8bd8-4dda-a43c-7b940514e381" containerName="controller-manager" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.712035 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c62b420-8bd8-4dda-a43c-7b940514e381" containerName="controller-manager" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.712595 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.715681 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.716827 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.717048 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.716985 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.717181 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.717506 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.723461 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-64df456547-z2ngz"] Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.725568 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.771021 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67c46e96-256b-45c3-aad8-2402fd866831-serving-cert\") pod \"controller-manager-64df456547-z2ngz\" (UID: \"67c46e96-256b-45c3-aad8-2402fd866831\") " pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.771224 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67c46e96-256b-45c3-aad8-2402fd866831-client-ca\") pod \"controller-manager-64df456547-z2ngz\" (UID: \"67c46e96-256b-45c3-aad8-2402fd866831\") " pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.771262 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67c46e96-256b-45c3-aad8-2402fd866831-config\") pod \"controller-manager-64df456547-z2ngz\" (UID: \"67c46e96-256b-45c3-aad8-2402fd866831\") " pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.872041 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67c46e96-256b-45c3-aad8-2402fd866831-serving-cert\") pod \"controller-manager-64df456547-z2ngz\" (UID: \"67c46e96-256b-45c3-aad8-2402fd866831\") " pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.872105 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9mbs\" (UniqueName: \"kubernetes.io/projected/67c46e96-256b-45c3-aad8-2402fd866831-kube-api-access-m9mbs\") pod \"controller-manager-64df456547-z2ngz\" (UID: \"67c46e96-256b-45c3-aad8-2402fd866831\") " pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.872152 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/67c46e96-256b-45c3-aad8-2402fd866831-proxy-ca-bundles\") pod \"controller-manager-64df456547-z2ngz\" (UID: \"67c46e96-256b-45c3-aad8-2402fd866831\") " pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.872190 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67c46e96-256b-45c3-aad8-2402fd866831-client-ca\") pod \"controller-manager-64df456547-z2ngz\" (UID: \"67c46e96-256b-45c3-aad8-2402fd866831\") " pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.872419 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67c46e96-256b-45c3-aad8-2402fd866831-config\") pod \"controller-manager-64df456547-z2ngz\" (UID: \"67c46e96-256b-45c3-aad8-2402fd866831\") " pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.873205 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67c46e96-256b-45c3-aad8-2402fd866831-client-ca\") pod \"controller-manager-64df456547-z2ngz\" (UID: \"67c46e96-256b-45c3-aad8-2402fd866831\") " pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.873574 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67c46e96-256b-45c3-aad8-2402fd866831-config\") pod \"controller-manager-64df456547-z2ngz\" (UID: \"67c46e96-256b-45c3-aad8-2402fd866831\") " pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.876459 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67c46e96-256b-45c3-aad8-2402fd866831-serving-cert\") pod \"controller-manager-64df456547-z2ngz\" (UID: \"67c46e96-256b-45c3-aad8-2402fd866831\") " pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.973623 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9mbs\" (UniqueName: \"kubernetes.io/projected/67c46e96-256b-45c3-aad8-2402fd866831-kube-api-access-m9mbs\") pod \"controller-manager-64df456547-z2ngz\" (UID: \"67c46e96-256b-45c3-aad8-2402fd866831\") " pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.973689 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/67c46e96-256b-45c3-aad8-2402fd866831-proxy-ca-bundles\") pod \"controller-manager-64df456547-z2ngz\" (UID: \"67c46e96-256b-45c3-aad8-2402fd866831\") " pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.974889 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/67c46e96-256b-45c3-aad8-2402fd866831-proxy-ca-bundles\") pod \"controller-manager-64df456547-z2ngz\" (UID: \"67c46e96-256b-45c3-aad8-2402fd866831\") " pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:06 crc kubenswrapper[4810]: I1201 14:40:06.988774 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9mbs\" (UniqueName: \"kubernetes.io/projected/67c46e96-256b-45c3-aad8-2402fd866831-kube-api-access-m9mbs\") pod \"controller-manager-64df456547-z2ngz\" (UID: \"67c46e96-256b-45c3-aad8-2402fd866831\") " pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:07 crc kubenswrapper[4810]: I1201 14:40:07.027333 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:07 crc kubenswrapper[4810]: I1201 14:40:07.232789 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-64df456547-z2ngz"] Dec 01 14:40:07 crc kubenswrapper[4810]: I1201 14:40:07.377370 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" event={"ID":"67c46e96-256b-45c3-aad8-2402fd866831","Type":"ContainerStarted","Data":"b4f56948393ad0daba638f98a5b8abef9f1eb818cca595770c63b52042a38c5b"} Dec 01 14:40:07 crc kubenswrapper[4810]: I1201 14:40:07.377430 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" event={"ID":"67c46e96-256b-45c3-aad8-2402fd866831","Type":"ContainerStarted","Data":"8d49bae2e0d0c4e959c0bd5e6e5ab1b0c684cb009d9f52a844ef6122495de22a"} Dec 01 14:40:07 crc kubenswrapper[4810]: I1201 14:40:07.377976 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:07 crc kubenswrapper[4810]: I1201 14:40:07.378862 4810 patch_prober.go:28] interesting pod/controller-manager-64df456547-z2ngz container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.61:8443/healthz\": dial tcp 10.217.0.61:8443: connect: connection refused" start-of-body= Dec 01 14:40:07 crc kubenswrapper[4810]: I1201 14:40:07.378896 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" podUID="67c46e96-256b-45c3-aad8-2402fd866831" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.61:8443/healthz\": dial tcp 10.217.0.61:8443: connect: connection refused" Dec 01 14:40:08 crc kubenswrapper[4810]: I1201 14:40:08.385601 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" Dec 01 14:40:08 crc kubenswrapper[4810]: I1201 14:40:08.403695 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-64df456547-z2ngz" podStartSLOduration=3.403673664 podStartE2EDuration="3.403673664s" podCreationTimestamp="2025-12-01 14:40:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:40:07.40158082 +0000 UTC m=+373.165090423" watchObservedRunningTime="2025-12-01 14:40:08.403673664 +0000 UTC m=+374.167183267" Dec 01 14:40:18 crc kubenswrapper[4810]: I1201 14:40:18.758160 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-q8psn" Dec 01 14:40:18 crc kubenswrapper[4810]: I1201 14:40:18.807896 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qvm7d"] Dec 01 14:40:32 crc kubenswrapper[4810]: I1201 14:40:32.972253 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:40:32 crc kubenswrapper[4810]: I1201 14:40:32.972959 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.128480 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fm8lj"] Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.131492 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fm8lj" podUID="ca896d69-0651-48e1-a001-0b3c57e1ece4" containerName="registry-server" containerID="cri-o://2d08d2f1c0b268ba9a58259fe5f93283af6c1535fbe9eaaeda089e05d13363b9" gracePeriod=30 Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.143818 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pxt6s"] Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.144031 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pxt6s" podUID="b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5" containerName="registry-server" containerID="cri-o://e94d75582a7295b57e07c5299519521fd2e53fd9970afd7c17f397f524ed0fd2" gracePeriod=30 Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.171590 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cgzqr"] Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.171835 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" podUID="0fa71e97-6be0-4300-bef2-b2dffbf30076" containerName="marketplace-operator" containerID="cri-o://c9c96013698eda0a27436b9bb842d6d859dced3ff353666d8f3d9dbcd7b2c0c3" gracePeriod=30 Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.179840 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pcjqf"] Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.180074 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pcjqf" podUID="00d0bac3-a43f-469e-9335-df4013f7432e" containerName="registry-server" containerID="cri-o://0fd88b0ee7b773525f6289e968cc51d666962280d0245c888069038085bd82c8" gracePeriod=30 Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.186785 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xnb48"] Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.187349 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xnb48" podUID="93bf2297-d7e7-48f2-8a96-6c33d5b3754b" containerName="registry-server" containerID="cri-o://43180ba72c6bf26818cd7afe73674086f3015314ec2ac6c796a829b955537ab7" gracePeriod=30 Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.189245 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8ss9b"] Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.189957 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8ss9b" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.199879 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8ss9b"] Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.292202 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8ss9b\" (UID: \"97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b\") " pod="openshift-marketplace/marketplace-operator-79b997595-8ss9b" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.292257 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8ss9b\" (UID: \"97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b\") " pod="openshift-marketplace/marketplace-operator-79b997595-8ss9b" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.292360 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xzjk\" (UniqueName: \"kubernetes.io/projected/97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b-kube-api-access-7xzjk\") pod \"marketplace-operator-79b997595-8ss9b\" (UID: \"97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b\") " pod="openshift-marketplace/marketplace-operator-79b997595-8ss9b" Dec 01 14:40:43 crc kubenswrapper[4810]: E1201 14:40:43.360077 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 43180ba72c6bf26818cd7afe73674086f3015314ec2ac6c796a829b955537ab7 is running failed: container process not found" containerID="43180ba72c6bf26818cd7afe73674086f3015314ec2ac6c796a829b955537ab7" cmd=["grpc_health_probe","-addr=:50051"] Dec 01 14:40:43 crc kubenswrapper[4810]: E1201 14:40:43.361013 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 43180ba72c6bf26818cd7afe73674086f3015314ec2ac6c796a829b955537ab7 is running failed: container process not found" containerID="43180ba72c6bf26818cd7afe73674086f3015314ec2ac6c796a829b955537ab7" cmd=["grpc_health_probe","-addr=:50051"] Dec 01 14:40:43 crc kubenswrapper[4810]: E1201 14:40:43.362911 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 43180ba72c6bf26818cd7afe73674086f3015314ec2ac6c796a829b955537ab7 is running failed: container process not found" containerID="43180ba72c6bf26818cd7afe73674086f3015314ec2ac6c796a829b955537ab7" cmd=["grpc_health_probe","-addr=:50051"] Dec 01 14:40:43 crc kubenswrapper[4810]: E1201 14:40:43.362974 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 43180ba72c6bf26818cd7afe73674086f3015314ec2ac6c796a829b955537ab7 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-xnb48" podUID="93bf2297-d7e7-48f2-8a96-6c33d5b3754b" containerName="registry-server" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.393831 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8ss9b\" (UID: \"97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b\") " pod="openshift-marketplace/marketplace-operator-79b997595-8ss9b" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.393894 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8ss9b\" (UID: \"97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b\") " pod="openshift-marketplace/marketplace-operator-79b997595-8ss9b" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.393971 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xzjk\" (UniqueName: \"kubernetes.io/projected/97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b-kube-api-access-7xzjk\") pod \"marketplace-operator-79b997595-8ss9b\" (UID: \"97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b\") " pod="openshift-marketplace/marketplace-operator-79b997595-8ss9b" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.395220 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8ss9b\" (UID: \"97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b\") " pod="openshift-marketplace/marketplace-operator-79b997595-8ss9b" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.402519 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8ss9b\" (UID: \"97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b\") " pod="openshift-marketplace/marketplace-operator-79b997595-8ss9b" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.411031 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xzjk\" (UniqueName: \"kubernetes.io/projected/97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b-kube-api-access-7xzjk\") pod \"marketplace-operator-79b997595-8ss9b\" (UID: \"97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b\") " pod="openshift-marketplace/marketplace-operator-79b997595-8ss9b" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.590681 4810 generic.go:334] "Generic (PLEG): container finished" podID="00d0bac3-a43f-469e-9335-df4013f7432e" containerID="0fd88b0ee7b773525f6289e968cc51d666962280d0245c888069038085bd82c8" exitCode=0 Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.590753 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcjqf" event={"ID":"00d0bac3-a43f-469e-9335-df4013f7432e","Type":"ContainerDied","Data":"0fd88b0ee7b773525f6289e968cc51d666962280d0245c888069038085bd82c8"} Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.592620 4810 generic.go:334] "Generic (PLEG): container finished" podID="b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5" containerID="e94d75582a7295b57e07c5299519521fd2e53fd9970afd7c17f397f524ed0fd2" exitCode=0 Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.592664 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxt6s" event={"ID":"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5","Type":"ContainerDied","Data":"e94d75582a7295b57e07c5299519521fd2e53fd9970afd7c17f397f524ed0fd2"} Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.594589 4810 generic.go:334] "Generic (PLEG): container finished" podID="93bf2297-d7e7-48f2-8a96-6c33d5b3754b" containerID="43180ba72c6bf26818cd7afe73674086f3015314ec2ac6c796a829b955537ab7" exitCode=0 Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.594632 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xnb48" event={"ID":"93bf2297-d7e7-48f2-8a96-6c33d5b3754b","Type":"ContainerDied","Data":"43180ba72c6bf26818cd7afe73674086f3015314ec2ac6c796a829b955537ab7"} Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.596626 4810 generic.go:334] "Generic (PLEG): container finished" podID="0fa71e97-6be0-4300-bef2-b2dffbf30076" containerID="c9c96013698eda0a27436b9bb842d6d859dced3ff353666d8f3d9dbcd7b2c0c3" exitCode=0 Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.596664 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" event={"ID":"0fa71e97-6be0-4300-bef2-b2dffbf30076","Type":"ContainerDied","Data":"c9c96013698eda0a27436b9bb842d6d859dced3ff353666d8f3d9dbcd7b2c0c3"} Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.596684 4810 scope.go:117] "RemoveContainer" containerID="987552bf4a9cd3bc3b0eb15466951f1c7575b07dd03b1505f1f8e6c0d30c382e" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.598775 4810 generic.go:334] "Generic (PLEG): container finished" podID="ca896d69-0651-48e1-a001-0b3c57e1ece4" containerID="2d08d2f1c0b268ba9a58259fe5f93283af6c1535fbe9eaaeda089e05d13363b9" exitCode=0 Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.598797 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fm8lj" event={"ID":"ca896d69-0651-48e1-a001-0b3c57e1ece4","Type":"ContainerDied","Data":"2d08d2f1c0b268ba9a58259fe5f93283af6c1535fbe9eaaeda089e05d13363b9"} Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.643098 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8ss9b" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.648907 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxt6s" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.698794 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfwql\" (UniqueName: \"kubernetes.io/projected/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5-kube-api-access-hfwql\") pod \"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5\" (UID: \"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5\") " Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.698849 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5-catalog-content\") pod \"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5\" (UID: \"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5\") " Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.698935 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5-utilities\") pod \"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5\" (UID: \"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5\") " Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.700830 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5-utilities" (OuterVolumeSpecName: "utilities") pod "b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5" (UID: "b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.706670 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5-kube-api-access-hfwql" (OuterVolumeSpecName: "kube-api-access-hfwql") pod "b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5" (UID: "b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5"). InnerVolumeSpecName "kube-api-access-hfwql". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.714456 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xnb48" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.750379 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.753414 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fm8lj" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.787967 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5" (UID: "b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.799794 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zrjs\" (UniqueName: \"kubernetes.io/projected/ca896d69-0651-48e1-a001-0b3c57e1ece4-kube-api-access-5zrjs\") pod \"ca896d69-0651-48e1-a001-0b3c57e1ece4\" (UID: \"ca896d69-0651-48e1-a001-0b3c57e1ece4\") " Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.799853 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93bf2297-d7e7-48f2-8a96-6c33d5b3754b-catalog-content\") pod \"93bf2297-d7e7-48f2-8a96-6c33d5b3754b\" (UID: \"93bf2297-d7e7-48f2-8a96-6c33d5b3754b\") " Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.799906 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0fa71e97-6be0-4300-bef2-b2dffbf30076-marketplace-trusted-ca\") pod \"0fa71e97-6be0-4300-bef2-b2dffbf30076\" (UID: \"0fa71e97-6be0-4300-bef2-b2dffbf30076\") " Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.799930 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca896d69-0651-48e1-a001-0b3c57e1ece4-utilities\") pod \"ca896d69-0651-48e1-a001-0b3c57e1ece4\" (UID: \"ca896d69-0651-48e1-a001-0b3c57e1ece4\") " Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.799959 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0fa71e97-6be0-4300-bef2-b2dffbf30076-marketplace-operator-metrics\") pod \"0fa71e97-6be0-4300-bef2-b2dffbf30076\" (UID: \"0fa71e97-6be0-4300-bef2-b2dffbf30076\") " Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.799986 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5g7h\" (UniqueName: \"kubernetes.io/projected/93bf2297-d7e7-48f2-8a96-6c33d5b3754b-kube-api-access-q5g7h\") pod \"93bf2297-d7e7-48f2-8a96-6c33d5b3754b\" (UID: \"93bf2297-d7e7-48f2-8a96-6c33d5b3754b\") " Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.800010 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrkzf\" (UniqueName: \"kubernetes.io/projected/0fa71e97-6be0-4300-bef2-b2dffbf30076-kube-api-access-lrkzf\") pod \"0fa71e97-6be0-4300-bef2-b2dffbf30076\" (UID: \"0fa71e97-6be0-4300-bef2-b2dffbf30076\") " Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.800043 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca896d69-0651-48e1-a001-0b3c57e1ece4-catalog-content\") pod \"ca896d69-0651-48e1-a001-0b3c57e1ece4\" (UID: \"ca896d69-0651-48e1-a001-0b3c57e1ece4\") " Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.800057 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93bf2297-d7e7-48f2-8a96-6c33d5b3754b-utilities\") pod \"93bf2297-d7e7-48f2-8a96-6c33d5b3754b\" (UID: \"93bf2297-d7e7-48f2-8a96-6c33d5b3754b\") " Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.800303 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.800316 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfwql\" (UniqueName: \"kubernetes.io/projected/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5-kube-api-access-hfwql\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.800344 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.801534 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca896d69-0651-48e1-a001-0b3c57e1ece4-utilities" (OuterVolumeSpecName: "utilities") pod "ca896d69-0651-48e1-a001-0b3c57e1ece4" (UID: "ca896d69-0651-48e1-a001-0b3c57e1ece4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.805219 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93bf2297-d7e7-48f2-8a96-6c33d5b3754b-kube-api-access-q5g7h" (OuterVolumeSpecName: "kube-api-access-q5g7h") pod "93bf2297-d7e7-48f2-8a96-6c33d5b3754b" (UID: "93bf2297-d7e7-48f2-8a96-6c33d5b3754b"). InnerVolumeSpecName "kube-api-access-q5g7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.805843 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fa71e97-6be0-4300-bef2-b2dffbf30076-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "0fa71e97-6be0-4300-bef2-b2dffbf30076" (UID: "0fa71e97-6be0-4300-bef2-b2dffbf30076"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.805925 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca896d69-0651-48e1-a001-0b3c57e1ece4-kube-api-access-5zrjs" (OuterVolumeSpecName: "kube-api-access-5zrjs") pod "ca896d69-0651-48e1-a001-0b3c57e1ece4" (UID: "ca896d69-0651-48e1-a001-0b3c57e1ece4"). InnerVolumeSpecName "kube-api-access-5zrjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.806464 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93bf2297-d7e7-48f2-8a96-6c33d5b3754b-utilities" (OuterVolumeSpecName: "utilities") pod "93bf2297-d7e7-48f2-8a96-6c33d5b3754b" (UID: "93bf2297-d7e7-48f2-8a96-6c33d5b3754b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.808023 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fa71e97-6be0-4300-bef2-b2dffbf30076-kube-api-access-lrkzf" (OuterVolumeSpecName: "kube-api-access-lrkzf") pod "0fa71e97-6be0-4300-bef2-b2dffbf30076" (UID: "0fa71e97-6be0-4300-bef2-b2dffbf30076"). InnerVolumeSpecName "kube-api-access-lrkzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.817831 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pcjqf" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.819246 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fa71e97-6be0-4300-bef2-b2dffbf30076-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "0fa71e97-6be0-4300-bef2-b2dffbf30076" (UID: "0fa71e97-6be0-4300-bef2-b2dffbf30076"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.850331 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" podUID="5a67f6b6-1881-4807-b719-40ca184bd5dc" containerName="registry" containerID="cri-o://9b13f77523af72c9131cb10ef312970aba06063f1cefd652c3ef0ec5676afbda" gracePeriod=30 Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.862048 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca896d69-0651-48e1-a001-0b3c57e1ece4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca896d69-0651-48e1-a001-0b3c57e1ece4" (UID: "ca896d69-0651-48e1-a001-0b3c57e1ece4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.901320 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00d0bac3-a43f-469e-9335-df4013f7432e-catalog-content\") pod \"00d0bac3-a43f-469e-9335-df4013f7432e\" (UID: \"00d0bac3-a43f-469e-9335-df4013f7432e\") " Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.901455 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00d0bac3-a43f-469e-9335-df4013f7432e-utilities\") pod \"00d0bac3-a43f-469e-9335-df4013f7432e\" (UID: \"00d0bac3-a43f-469e-9335-df4013f7432e\") " Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.901565 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhhzd\" (UniqueName: \"kubernetes.io/projected/00d0bac3-a43f-469e-9335-df4013f7432e-kube-api-access-vhhzd\") pod \"00d0bac3-a43f-469e-9335-df4013f7432e\" (UID: \"00d0bac3-a43f-469e-9335-df4013f7432e\") " Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.901823 4810 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0fa71e97-6be0-4300-bef2-b2dffbf30076-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.901845 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca896d69-0651-48e1-a001-0b3c57e1ece4-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.901859 4810 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0fa71e97-6be0-4300-bef2-b2dffbf30076-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.901874 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5g7h\" (UniqueName: \"kubernetes.io/projected/93bf2297-d7e7-48f2-8a96-6c33d5b3754b-kube-api-access-q5g7h\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.901887 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrkzf\" (UniqueName: \"kubernetes.io/projected/0fa71e97-6be0-4300-bef2-b2dffbf30076-kube-api-access-lrkzf\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.901899 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93bf2297-d7e7-48f2-8a96-6c33d5b3754b-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.901911 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca896d69-0651-48e1-a001-0b3c57e1ece4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.901921 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zrjs\" (UniqueName: \"kubernetes.io/projected/ca896d69-0651-48e1-a001-0b3c57e1ece4-kube-api-access-5zrjs\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.902760 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00d0bac3-a43f-469e-9335-df4013f7432e-utilities" (OuterVolumeSpecName: "utilities") pod "00d0bac3-a43f-469e-9335-df4013f7432e" (UID: "00d0bac3-a43f-469e-9335-df4013f7432e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.908748 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00d0bac3-a43f-469e-9335-df4013f7432e-kube-api-access-vhhzd" (OuterVolumeSpecName: "kube-api-access-vhhzd") pod "00d0bac3-a43f-469e-9335-df4013f7432e" (UID: "00d0bac3-a43f-469e-9335-df4013f7432e"). InnerVolumeSpecName "kube-api-access-vhhzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.921691 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00d0bac3-a43f-469e-9335-df4013f7432e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00d0bac3-a43f-469e-9335-df4013f7432e" (UID: "00d0bac3-a43f-469e-9335-df4013f7432e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:40:43 crc kubenswrapper[4810]: I1201 14:40:43.948105 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93bf2297-d7e7-48f2-8a96-6c33d5b3754b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93bf2297-d7e7-48f2-8a96-6c33d5b3754b" (UID: "93bf2297-d7e7-48f2-8a96-6c33d5b3754b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.002681 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00d0bac3-a43f-469e-9335-df4013f7432e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.002716 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93bf2297-d7e7-48f2-8a96-6c33d5b3754b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.002728 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00d0bac3-a43f-469e-9335-df4013f7432e-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.002745 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhhzd\" (UniqueName: \"kubernetes.io/projected/00d0bac3-a43f-469e-9335-df4013f7432e-kube-api-access-vhhzd\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.122544 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8ss9b"] Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.276917 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.306996 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv5zp\" (UniqueName: \"kubernetes.io/projected/5a67f6b6-1881-4807-b719-40ca184bd5dc-kube-api-access-bv5zp\") pod \"5a67f6b6-1881-4807-b719-40ca184bd5dc\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.307090 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5a67f6b6-1881-4807-b719-40ca184bd5dc-ca-trust-extracted\") pod \"5a67f6b6-1881-4807-b719-40ca184bd5dc\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.307327 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"5a67f6b6-1881-4807-b719-40ca184bd5dc\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.307368 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5a67f6b6-1881-4807-b719-40ca184bd5dc-trusted-ca\") pod \"5a67f6b6-1881-4807-b719-40ca184bd5dc\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.307386 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5a67f6b6-1881-4807-b719-40ca184bd5dc-registry-certificates\") pod \"5a67f6b6-1881-4807-b719-40ca184bd5dc\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.307422 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5a67f6b6-1881-4807-b719-40ca184bd5dc-installation-pull-secrets\") pod \"5a67f6b6-1881-4807-b719-40ca184bd5dc\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.307449 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5a67f6b6-1881-4807-b719-40ca184bd5dc-bound-sa-token\") pod \"5a67f6b6-1881-4807-b719-40ca184bd5dc\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.307528 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5a67f6b6-1881-4807-b719-40ca184bd5dc-registry-tls\") pod \"5a67f6b6-1881-4807-b719-40ca184bd5dc\" (UID: \"5a67f6b6-1881-4807-b719-40ca184bd5dc\") " Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.308734 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a67f6b6-1881-4807-b719-40ca184bd5dc-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "5a67f6b6-1881-4807-b719-40ca184bd5dc" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.309577 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a67f6b6-1881-4807-b719-40ca184bd5dc-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "5a67f6b6-1881-4807-b719-40ca184bd5dc" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.311279 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a67f6b6-1881-4807-b719-40ca184bd5dc-kube-api-access-bv5zp" (OuterVolumeSpecName: "kube-api-access-bv5zp") pod "5a67f6b6-1881-4807-b719-40ca184bd5dc" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc"). InnerVolumeSpecName "kube-api-access-bv5zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.311991 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a67f6b6-1881-4807-b719-40ca184bd5dc-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "5a67f6b6-1881-4807-b719-40ca184bd5dc" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.313460 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a67f6b6-1881-4807-b719-40ca184bd5dc-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "5a67f6b6-1881-4807-b719-40ca184bd5dc" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.318964 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a67f6b6-1881-4807-b719-40ca184bd5dc-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "5a67f6b6-1881-4807-b719-40ca184bd5dc" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.319697 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "5a67f6b6-1881-4807-b719-40ca184bd5dc" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.329420 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a67f6b6-1881-4807-b719-40ca184bd5dc-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "5a67f6b6-1881-4807-b719-40ca184bd5dc" (UID: "5a67f6b6-1881-4807-b719-40ca184bd5dc"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.408164 4810 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5a67f6b6-1881-4807-b719-40ca184bd5dc-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.408202 4810 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5a67f6b6-1881-4807-b719-40ca184bd5dc-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.408216 4810 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5a67f6b6-1881-4807-b719-40ca184bd5dc-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.408225 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv5zp\" (UniqueName: \"kubernetes.io/projected/5a67f6b6-1881-4807-b719-40ca184bd5dc-kube-api-access-bv5zp\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.408235 4810 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5a67f6b6-1881-4807-b719-40ca184bd5dc-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.408243 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5a67f6b6-1881-4807-b719-40ca184bd5dc-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.408251 4810 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5a67f6b6-1881-4807-b719-40ca184bd5dc-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.607084 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pcjqf" event={"ID":"00d0bac3-a43f-469e-9335-df4013f7432e","Type":"ContainerDied","Data":"1757fb4e02abffa4095d30a9e9184015f1e40c344455b9f9f9b7946de8edce94"} Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.607168 4810 scope.go:117] "RemoveContainer" containerID="0fd88b0ee7b773525f6289e968cc51d666962280d0245c888069038085bd82c8" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.607374 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pcjqf" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.613553 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxt6s" event={"ID":"b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5","Type":"ContainerDied","Data":"0f26490f9d7312629415a15a28fd8e4b4b4800c600bc73694c59fe0deff097b9"} Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.613576 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxt6s" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.619514 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xnb48" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.621461 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xnb48" event={"ID":"93bf2297-d7e7-48f2-8a96-6c33d5b3754b","Type":"ContainerDied","Data":"35b2f9a2a3b39f45955f3c06072ca5b454f0a2be191a9f35a141f9035689a05b"} Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.623964 4810 generic.go:334] "Generic (PLEG): container finished" podID="5a67f6b6-1881-4807-b719-40ca184bd5dc" containerID="9b13f77523af72c9131cb10ef312970aba06063f1cefd652c3ef0ec5676afbda" exitCode=0 Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.624063 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" event={"ID":"5a67f6b6-1881-4807-b719-40ca184bd5dc","Type":"ContainerDied","Data":"9b13f77523af72c9131cb10ef312970aba06063f1cefd652c3ef0ec5676afbda"} Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.624088 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" event={"ID":"5a67f6b6-1881-4807-b719-40ca184bd5dc","Type":"ContainerDied","Data":"62e6f9c0e7d9389b8561ad5d1913e7d3cf03ebc018a89667997eff66b6b4ce90"} Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.624110 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qvm7d" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.631714 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pcjqf"] Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.634701 4810 scope.go:117] "RemoveContainer" containerID="767535a7f3e5bc715874a85d9edbd906dbf73d5d72716b57163bfa7a38854b46" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.637403 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.637709 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-cgzqr" event={"ID":"0fa71e97-6be0-4300-bef2-b2dffbf30076","Type":"ContainerDied","Data":"d402abb08a7945297d1cb224307fcbb1216733bd4da5c915507f50449ed641ec"} Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.638148 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pcjqf"] Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.642271 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fm8lj" event={"ID":"ca896d69-0651-48e1-a001-0b3c57e1ece4","Type":"ContainerDied","Data":"a2ced1a9adde5165eb4fa97545fb79e841108b61cb643460e2c77fb256b9e779"} Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.642291 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fm8lj" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.652247 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8ss9b" event={"ID":"97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b","Type":"ContainerStarted","Data":"770eca56765e1d71fe4f3f4be17ad50dd802a1cfb3cb0fcec2abd2df80faefb7"} Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.652309 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8ss9b" event={"ID":"97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b","Type":"ContainerStarted","Data":"8d00549cb5f83210f80825166291ff523ce27a17a8fc12fb375e1366e0087c80"} Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.658740 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-8ss9b" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.658789 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pxt6s"] Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.658813 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pxt6s"] Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.661918 4810 scope.go:117] "RemoveContainer" containerID="8a300e31e64fde615c50cfaa13a7675fd48dbd82ed783bf4c06ac2644aff1857" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.662854 4810 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8ss9b container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.62:8080/healthz\": dial tcp 10.217.0.62:8080: connect: connection refused" start-of-body= Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.662927 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-8ss9b" podUID="97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.62:8080/healthz\": dial tcp 10.217.0.62:8080: connect: connection refused" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.674659 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xnb48"] Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.679631 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xnb48"] Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.688497 4810 scope.go:117] "RemoveContainer" containerID="e94d75582a7295b57e07c5299519521fd2e53fd9970afd7c17f397f524ed0fd2" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.713244 4810 scope.go:117] "RemoveContainer" containerID="0e21734ba2f73e2a661ec8a6491b122b5b97369fa45aef4adc74e825260f6b08" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.725351 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-8ss9b" podStartSLOduration=1.725334261 podStartE2EDuration="1.725334261s" podCreationTimestamp="2025-12-01 14:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:40:44.703318274 +0000 UTC m=+410.466827877" watchObservedRunningTime="2025-12-01 14:40:44.725334261 +0000 UTC m=+410.488843864" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.729054 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qvm7d"] Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.736017 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qvm7d"] Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.745908 4810 scope.go:117] "RemoveContainer" containerID="3a1d49990f7824e2a3f3cea80706c3b13951d9e30d24449a230a69cd92d6d598" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.750887 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cgzqr"] Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.754841 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-cgzqr"] Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.759388 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fm8lj"] Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.766539 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fm8lj"] Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.769039 4810 scope.go:117] "RemoveContainer" containerID="43180ba72c6bf26818cd7afe73674086f3015314ec2ac6c796a829b955537ab7" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.793154 4810 scope.go:117] "RemoveContainer" containerID="1256a0ea89cf35a3062cba66bd62af423c3acbaa80f3c7d8dc0852c087d53d8b" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.815740 4810 scope.go:117] "RemoveContainer" containerID="40544c213be6f99ca606c89d40f4a275a5aa86846696ceb5330ccb225c03c328" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.839564 4810 scope.go:117] "RemoveContainer" containerID="9b13f77523af72c9131cb10ef312970aba06063f1cefd652c3ef0ec5676afbda" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.853626 4810 scope.go:117] "RemoveContainer" containerID="9b13f77523af72c9131cb10ef312970aba06063f1cefd652c3ef0ec5676afbda" Dec 01 14:40:44 crc kubenswrapper[4810]: E1201 14:40:44.854734 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b13f77523af72c9131cb10ef312970aba06063f1cefd652c3ef0ec5676afbda\": container with ID starting with 9b13f77523af72c9131cb10ef312970aba06063f1cefd652c3ef0ec5676afbda not found: ID does not exist" containerID="9b13f77523af72c9131cb10ef312970aba06063f1cefd652c3ef0ec5676afbda" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.854809 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b13f77523af72c9131cb10ef312970aba06063f1cefd652c3ef0ec5676afbda"} err="failed to get container status \"9b13f77523af72c9131cb10ef312970aba06063f1cefd652c3ef0ec5676afbda\": rpc error: code = NotFound desc = could not find container \"9b13f77523af72c9131cb10ef312970aba06063f1cefd652c3ef0ec5676afbda\": container with ID starting with 9b13f77523af72c9131cb10ef312970aba06063f1cefd652c3ef0ec5676afbda not found: ID does not exist" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.854858 4810 scope.go:117] "RemoveContainer" containerID="c9c96013698eda0a27436b9bb842d6d859dced3ff353666d8f3d9dbcd7b2c0c3" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.872201 4810 scope.go:117] "RemoveContainer" containerID="2d08d2f1c0b268ba9a58259fe5f93283af6c1535fbe9eaaeda089e05d13363b9" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.885745 4810 scope.go:117] "RemoveContainer" containerID="699937a2d5d16c74f98d394717c8910b78d31ccedcc1f4fe5a7b4e8c8f009814" Dec 01 14:40:44 crc kubenswrapper[4810]: I1201 14:40:44.910040 4810 scope.go:117] "RemoveContainer" containerID="6342a41e762ac01621caf4778763386ce53ae43548907b267168386be609ba01" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.672421 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-8ss9b" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747442 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jmlkh"] Dec 01 14:40:45 crc kubenswrapper[4810]: E1201 14:40:45.747649 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a67f6b6-1881-4807-b719-40ca184bd5dc" containerName="registry" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747660 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a67f6b6-1881-4807-b719-40ca184bd5dc" containerName="registry" Dec 01 14:40:45 crc kubenswrapper[4810]: E1201 14:40:45.747671 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d0bac3-a43f-469e-9335-df4013f7432e" containerName="extract-utilities" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747678 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d0bac3-a43f-469e-9335-df4013f7432e" containerName="extract-utilities" Dec 01 14:40:45 crc kubenswrapper[4810]: E1201 14:40:45.747684 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca896d69-0651-48e1-a001-0b3c57e1ece4" containerName="registry-server" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747691 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca896d69-0651-48e1-a001-0b3c57e1ece4" containerName="registry-server" Dec 01 14:40:45 crc kubenswrapper[4810]: E1201 14:40:45.747701 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93bf2297-d7e7-48f2-8a96-6c33d5b3754b" containerName="extract-utilities" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747707 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="93bf2297-d7e7-48f2-8a96-6c33d5b3754b" containerName="extract-utilities" Dec 01 14:40:45 crc kubenswrapper[4810]: E1201 14:40:45.747718 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d0bac3-a43f-469e-9335-df4013f7432e" containerName="registry-server" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747724 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d0bac3-a43f-469e-9335-df4013f7432e" containerName="registry-server" Dec 01 14:40:45 crc kubenswrapper[4810]: E1201 14:40:45.747732 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca896d69-0651-48e1-a001-0b3c57e1ece4" containerName="extract-utilities" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747738 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca896d69-0651-48e1-a001-0b3c57e1ece4" containerName="extract-utilities" Dec 01 14:40:45 crc kubenswrapper[4810]: E1201 14:40:45.747746 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d0bac3-a43f-469e-9335-df4013f7432e" containerName="extract-content" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747752 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d0bac3-a43f-469e-9335-df4013f7432e" containerName="extract-content" Dec 01 14:40:45 crc kubenswrapper[4810]: E1201 14:40:45.747759 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93bf2297-d7e7-48f2-8a96-6c33d5b3754b" containerName="extract-content" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747764 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="93bf2297-d7e7-48f2-8a96-6c33d5b3754b" containerName="extract-content" Dec 01 14:40:45 crc kubenswrapper[4810]: E1201 14:40:45.747776 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fa71e97-6be0-4300-bef2-b2dffbf30076" containerName="marketplace-operator" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747781 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fa71e97-6be0-4300-bef2-b2dffbf30076" containerName="marketplace-operator" Dec 01 14:40:45 crc kubenswrapper[4810]: E1201 14:40:45.747788 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5" containerName="extract-utilities" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747796 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5" containerName="extract-utilities" Dec 01 14:40:45 crc kubenswrapper[4810]: E1201 14:40:45.747803 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93bf2297-d7e7-48f2-8a96-6c33d5b3754b" containerName="registry-server" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747809 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="93bf2297-d7e7-48f2-8a96-6c33d5b3754b" containerName="registry-server" Dec 01 14:40:45 crc kubenswrapper[4810]: E1201 14:40:45.747817 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca896d69-0651-48e1-a001-0b3c57e1ece4" containerName="extract-content" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747823 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca896d69-0651-48e1-a001-0b3c57e1ece4" containerName="extract-content" Dec 01 14:40:45 crc kubenswrapper[4810]: E1201 14:40:45.747833 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fa71e97-6be0-4300-bef2-b2dffbf30076" containerName="marketplace-operator" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747839 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fa71e97-6be0-4300-bef2-b2dffbf30076" containerName="marketplace-operator" Dec 01 14:40:45 crc kubenswrapper[4810]: E1201 14:40:45.747851 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5" containerName="extract-content" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747856 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5" containerName="extract-content" Dec 01 14:40:45 crc kubenswrapper[4810]: E1201 14:40:45.747862 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5" containerName="registry-server" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747868 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5" containerName="registry-server" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747952 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="93bf2297-d7e7-48f2-8a96-6c33d5b3754b" containerName="registry-server" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747963 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a67f6b6-1881-4807-b719-40ca184bd5dc" containerName="registry" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747970 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="00d0bac3-a43f-469e-9335-df4013f7432e" containerName="registry-server" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747978 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fa71e97-6be0-4300-bef2-b2dffbf30076" containerName="marketplace-operator" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747985 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca896d69-0651-48e1-a001-0b3c57e1ece4" containerName="registry-server" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.747992 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fa71e97-6be0-4300-bef2-b2dffbf30076" containerName="marketplace-operator" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.748002 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5" containerName="registry-server" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.748641 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jmlkh" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.753402 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.759652 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jmlkh"] Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.933839 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz6kx\" (UniqueName: \"kubernetes.io/projected/8c8f2b09-969f-4adf-8f45-8468d57e76be-kube-api-access-bz6kx\") pod \"redhat-marketplace-jmlkh\" (UID: \"8c8f2b09-969f-4adf-8f45-8468d57e76be\") " pod="openshift-marketplace/redhat-marketplace-jmlkh" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.934437 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8f2b09-969f-4adf-8f45-8468d57e76be-catalog-content\") pod \"redhat-marketplace-jmlkh\" (UID: \"8c8f2b09-969f-4adf-8f45-8468d57e76be\") " pod="openshift-marketplace/redhat-marketplace-jmlkh" Dec 01 14:40:45 crc kubenswrapper[4810]: I1201 14:40:45.934710 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8f2b09-969f-4adf-8f45-8468d57e76be-utilities\") pod \"redhat-marketplace-jmlkh\" (UID: \"8c8f2b09-969f-4adf-8f45-8468d57e76be\") " pod="openshift-marketplace/redhat-marketplace-jmlkh" Dec 01 14:40:46 crc kubenswrapper[4810]: I1201 14:40:46.036528 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8f2b09-969f-4adf-8f45-8468d57e76be-utilities\") pod \"redhat-marketplace-jmlkh\" (UID: \"8c8f2b09-969f-4adf-8f45-8468d57e76be\") " pod="openshift-marketplace/redhat-marketplace-jmlkh" Dec 01 14:40:46 crc kubenswrapper[4810]: I1201 14:40:46.037017 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz6kx\" (UniqueName: \"kubernetes.io/projected/8c8f2b09-969f-4adf-8f45-8468d57e76be-kube-api-access-bz6kx\") pod \"redhat-marketplace-jmlkh\" (UID: \"8c8f2b09-969f-4adf-8f45-8468d57e76be\") " pod="openshift-marketplace/redhat-marketplace-jmlkh" Dec 01 14:40:46 crc kubenswrapper[4810]: I1201 14:40:46.037145 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8f2b09-969f-4adf-8f45-8468d57e76be-catalog-content\") pod \"redhat-marketplace-jmlkh\" (UID: \"8c8f2b09-969f-4adf-8f45-8468d57e76be\") " pod="openshift-marketplace/redhat-marketplace-jmlkh" Dec 01 14:40:46 crc kubenswrapper[4810]: I1201 14:40:46.037366 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8f2b09-969f-4adf-8f45-8468d57e76be-utilities\") pod \"redhat-marketplace-jmlkh\" (UID: \"8c8f2b09-969f-4adf-8f45-8468d57e76be\") " pod="openshift-marketplace/redhat-marketplace-jmlkh" Dec 01 14:40:46 crc kubenswrapper[4810]: I1201 14:40:46.037786 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8f2b09-969f-4adf-8f45-8468d57e76be-catalog-content\") pod \"redhat-marketplace-jmlkh\" (UID: \"8c8f2b09-969f-4adf-8f45-8468d57e76be\") " pod="openshift-marketplace/redhat-marketplace-jmlkh" Dec 01 14:40:46 crc kubenswrapper[4810]: I1201 14:40:46.056304 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz6kx\" (UniqueName: \"kubernetes.io/projected/8c8f2b09-969f-4adf-8f45-8468d57e76be-kube-api-access-bz6kx\") pod \"redhat-marketplace-jmlkh\" (UID: \"8c8f2b09-969f-4adf-8f45-8468d57e76be\") " pod="openshift-marketplace/redhat-marketplace-jmlkh" Dec 01 14:40:46 crc kubenswrapper[4810]: I1201 14:40:46.062986 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jmlkh" Dec 01 14:40:46 crc kubenswrapper[4810]: I1201 14:40:46.511890 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00d0bac3-a43f-469e-9335-df4013f7432e" path="/var/lib/kubelet/pods/00d0bac3-a43f-469e-9335-df4013f7432e/volumes" Dec 01 14:40:46 crc kubenswrapper[4810]: I1201 14:40:46.513320 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fa71e97-6be0-4300-bef2-b2dffbf30076" path="/var/lib/kubelet/pods/0fa71e97-6be0-4300-bef2-b2dffbf30076/volumes" Dec 01 14:40:46 crc kubenswrapper[4810]: I1201 14:40:46.514196 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a67f6b6-1881-4807-b719-40ca184bd5dc" path="/var/lib/kubelet/pods/5a67f6b6-1881-4807-b719-40ca184bd5dc/volumes" Dec 01 14:40:46 crc kubenswrapper[4810]: I1201 14:40:46.515779 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93bf2297-d7e7-48f2-8a96-6c33d5b3754b" path="/var/lib/kubelet/pods/93bf2297-d7e7-48f2-8a96-6c33d5b3754b/volumes" Dec 01 14:40:46 crc kubenswrapper[4810]: W1201 14:40:46.519167 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c8f2b09_969f_4adf_8f45_8468d57e76be.slice/crio-05edea8734fda07f9db5bc7b356c81df88aff0fafd96e4826f518ce7989ad181 WatchSource:0}: Error finding container 05edea8734fda07f9db5bc7b356c81df88aff0fafd96e4826f518ce7989ad181: Status 404 returned error can't find the container with id 05edea8734fda07f9db5bc7b356c81df88aff0fafd96e4826f518ce7989ad181 Dec 01 14:40:46 crc kubenswrapper[4810]: I1201 14:40:46.525689 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5" path="/var/lib/kubelet/pods/b18b0a4d-1ac2-4ae4-a60f-73db7ec71cb5/volumes" Dec 01 14:40:46 crc kubenswrapper[4810]: I1201 14:40:46.526886 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca896d69-0651-48e1-a001-0b3c57e1ece4" path="/var/lib/kubelet/pods/ca896d69-0651-48e1-a001-0b3c57e1ece4/volumes" Dec 01 14:40:46 crc kubenswrapper[4810]: I1201 14:40:46.528769 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jmlkh"] Dec 01 14:40:46 crc kubenswrapper[4810]: I1201 14:40:46.684947 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jmlkh" event={"ID":"8c8f2b09-969f-4adf-8f45-8468d57e76be","Type":"ContainerStarted","Data":"fb52524c57a5d9be1ce5303dec6d086e8a9588caec2dede5a340556b41760739"} Dec 01 14:40:46 crc kubenswrapper[4810]: I1201 14:40:46.684993 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jmlkh" event={"ID":"8c8f2b09-969f-4adf-8f45-8468d57e76be","Type":"ContainerStarted","Data":"05edea8734fda07f9db5bc7b356c81df88aff0fafd96e4826f518ce7989ad181"} Dec 01 14:40:47 crc kubenswrapper[4810]: I1201 14:40:47.157978 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-djh66"] Dec 01 14:40:47 crc kubenswrapper[4810]: I1201 14:40:47.158981 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djh66" Dec 01 14:40:47 crc kubenswrapper[4810]: I1201 14:40:47.161856 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-djh66"] Dec 01 14:40:47 crc kubenswrapper[4810]: I1201 14:40:47.166377 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 14:40:47 crc kubenswrapper[4810]: I1201 14:40:47.353384 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8cdfd34-81c3-49ff-bba6-8b0d4192520d-catalog-content\") pod \"redhat-operators-djh66\" (UID: \"b8cdfd34-81c3-49ff-bba6-8b0d4192520d\") " pod="openshift-marketplace/redhat-operators-djh66" Dec 01 14:40:47 crc kubenswrapper[4810]: I1201 14:40:47.354090 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gnm4\" (UniqueName: \"kubernetes.io/projected/b8cdfd34-81c3-49ff-bba6-8b0d4192520d-kube-api-access-7gnm4\") pod \"redhat-operators-djh66\" (UID: \"b8cdfd34-81c3-49ff-bba6-8b0d4192520d\") " pod="openshift-marketplace/redhat-operators-djh66" Dec 01 14:40:47 crc kubenswrapper[4810]: I1201 14:40:47.354249 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8cdfd34-81c3-49ff-bba6-8b0d4192520d-utilities\") pod \"redhat-operators-djh66\" (UID: \"b8cdfd34-81c3-49ff-bba6-8b0d4192520d\") " pod="openshift-marketplace/redhat-operators-djh66" Dec 01 14:40:47 crc kubenswrapper[4810]: I1201 14:40:47.456030 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gnm4\" (UniqueName: \"kubernetes.io/projected/b8cdfd34-81c3-49ff-bba6-8b0d4192520d-kube-api-access-7gnm4\") pod \"redhat-operators-djh66\" (UID: \"b8cdfd34-81c3-49ff-bba6-8b0d4192520d\") " pod="openshift-marketplace/redhat-operators-djh66" Dec 01 14:40:47 crc kubenswrapper[4810]: I1201 14:40:47.456160 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8cdfd34-81c3-49ff-bba6-8b0d4192520d-utilities\") pod \"redhat-operators-djh66\" (UID: \"b8cdfd34-81c3-49ff-bba6-8b0d4192520d\") " pod="openshift-marketplace/redhat-operators-djh66" Dec 01 14:40:47 crc kubenswrapper[4810]: I1201 14:40:47.456287 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8cdfd34-81c3-49ff-bba6-8b0d4192520d-catalog-content\") pod \"redhat-operators-djh66\" (UID: \"b8cdfd34-81c3-49ff-bba6-8b0d4192520d\") " pod="openshift-marketplace/redhat-operators-djh66" Dec 01 14:40:47 crc kubenswrapper[4810]: I1201 14:40:47.456784 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8cdfd34-81c3-49ff-bba6-8b0d4192520d-catalog-content\") pod \"redhat-operators-djh66\" (UID: \"b8cdfd34-81c3-49ff-bba6-8b0d4192520d\") " pod="openshift-marketplace/redhat-operators-djh66" Dec 01 14:40:47 crc kubenswrapper[4810]: I1201 14:40:47.456923 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8cdfd34-81c3-49ff-bba6-8b0d4192520d-utilities\") pod \"redhat-operators-djh66\" (UID: \"b8cdfd34-81c3-49ff-bba6-8b0d4192520d\") " pod="openshift-marketplace/redhat-operators-djh66" Dec 01 14:40:47 crc kubenswrapper[4810]: I1201 14:40:47.512321 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gnm4\" (UniqueName: \"kubernetes.io/projected/b8cdfd34-81c3-49ff-bba6-8b0d4192520d-kube-api-access-7gnm4\") pod \"redhat-operators-djh66\" (UID: \"b8cdfd34-81c3-49ff-bba6-8b0d4192520d\") " pod="openshift-marketplace/redhat-operators-djh66" Dec 01 14:40:47 crc kubenswrapper[4810]: I1201 14:40:47.692922 4810 generic.go:334] "Generic (PLEG): container finished" podID="8c8f2b09-969f-4adf-8f45-8468d57e76be" containerID="fb52524c57a5d9be1ce5303dec6d086e8a9588caec2dede5a340556b41760739" exitCode=0 Dec 01 14:40:47 crc kubenswrapper[4810]: I1201 14:40:47.693095 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jmlkh" event={"ID":"8c8f2b09-969f-4adf-8f45-8468d57e76be","Type":"ContainerDied","Data":"fb52524c57a5d9be1ce5303dec6d086e8a9588caec2dede5a340556b41760739"} Dec 01 14:40:47 crc kubenswrapper[4810]: I1201 14:40:47.784531 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djh66" Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.148908 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gpx9d"] Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.150868 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gpx9d" Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.153604 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.159049 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gpx9d"] Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.164962 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n44vh\" (UniqueName: \"kubernetes.io/projected/92abf8d2-574f-4375-bc2a-3436ab810e23-kube-api-access-n44vh\") pod \"community-operators-gpx9d\" (UID: \"92abf8d2-574f-4375-bc2a-3436ab810e23\") " pod="openshift-marketplace/community-operators-gpx9d" Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.164998 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92abf8d2-574f-4375-bc2a-3436ab810e23-utilities\") pod \"community-operators-gpx9d\" (UID: \"92abf8d2-574f-4375-bc2a-3436ab810e23\") " pod="openshift-marketplace/community-operators-gpx9d" Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.165024 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92abf8d2-574f-4375-bc2a-3436ab810e23-catalog-content\") pod \"community-operators-gpx9d\" (UID: \"92abf8d2-574f-4375-bc2a-3436ab810e23\") " pod="openshift-marketplace/community-operators-gpx9d" Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.204998 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-djh66"] Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.266279 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n44vh\" (UniqueName: \"kubernetes.io/projected/92abf8d2-574f-4375-bc2a-3436ab810e23-kube-api-access-n44vh\") pod \"community-operators-gpx9d\" (UID: \"92abf8d2-574f-4375-bc2a-3436ab810e23\") " pod="openshift-marketplace/community-operators-gpx9d" Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.266329 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92abf8d2-574f-4375-bc2a-3436ab810e23-utilities\") pod \"community-operators-gpx9d\" (UID: \"92abf8d2-574f-4375-bc2a-3436ab810e23\") " pod="openshift-marketplace/community-operators-gpx9d" Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.266361 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92abf8d2-574f-4375-bc2a-3436ab810e23-catalog-content\") pod \"community-operators-gpx9d\" (UID: \"92abf8d2-574f-4375-bc2a-3436ab810e23\") " pod="openshift-marketplace/community-operators-gpx9d" Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.266885 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92abf8d2-574f-4375-bc2a-3436ab810e23-utilities\") pod \"community-operators-gpx9d\" (UID: \"92abf8d2-574f-4375-bc2a-3436ab810e23\") " pod="openshift-marketplace/community-operators-gpx9d" Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.266891 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92abf8d2-574f-4375-bc2a-3436ab810e23-catalog-content\") pod \"community-operators-gpx9d\" (UID: \"92abf8d2-574f-4375-bc2a-3436ab810e23\") " pod="openshift-marketplace/community-operators-gpx9d" Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.287225 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n44vh\" (UniqueName: \"kubernetes.io/projected/92abf8d2-574f-4375-bc2a-3436ab810e23-kube-api-access-n44vh\") pod \"community-operators-gpx9d\" (UID: \"92abf8d2-574f-4375-bc2a-3436ab810e23\") " pod="openshift-marketplace/community-operators-gpx9d" Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.468697 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gpx9d" Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.699598 4810 generic.go:334] "Generic (PLEG): container finished" podID="b8cdfd34-81c3-49ff-bba6-8b0d4192520d" containerID="7b9a2f3f0d9962a491a066dc2a2be88fb77afe5efe4a5a93c71ae3880ea887ae" exitCode=0 Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.699699 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djh66" event={"ID":"b8cdfd34-81c3-49ff-bba6-8b0d4192520d","Type":"ContainerDied","Data":"7b9a2f3f0d9962a491a066dc2a2be88fb77afe5efe4a5a93c71ae3880ea887ae"} Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.700234 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djh66" event={"ID":"b8cdfd34-81c3-49ff-bba6-8b0d4192520d","Type":"ContainerStarted","Data":"431f800501d6e071a0021ac82554a8e9afe4194a99ab181d44c97e2180241618"} Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.706666 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jmlkh" event={"ID":"8c8f2b09-969f-4adf-8f45-8468d57e76be","Type":"ContainerStarted","Data":"d7c1cf93a0a3ff0f57b293426a4848dae687567b389a9967a5bd89bb7fe91bc0"} Dec 01 14:40:48 crc kubenswrapper[4810]: I1201 14:40:48.874328 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gpx9d"] Dec 01 14:40:48 crc kubenswrapper[4810]: W1201 14:40:48.882761 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92abf8d2_574f_4375_bc2a_3436ab810e23.slice/crio-c8a898448bdc0f5b4c3ca97c943b68af6120dde46ce5ee2ee4b940c26bb9e134 WatchSource:0}: Error finding container c8a898448bdc0f5b4c3ca97c943b68af6120dde46ce5ee2ee4b940c26bb9e134: Status 404 returned error can't find the container with id c8a898448bdc0f5b4c3ca97c943b68af6120dde46ce5ee2ee4b940c26bb9e134 Dec 01 14:40:48 crc kubenswrapper[4810]: E1201 14:40:48.885285 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c8f2b09_969f_4adf_8f45_8468d57e76be.slice/crio-d7c1cf93a0a3ff0f57b293426a4848dae687567b389a9967a5bd89bb7fe91bc0.scope\": RecentStats: unable to find data in memory cache]" Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.559750 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w9g5p"] Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.561430 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w9g5p" Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.563943 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.569513 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w9g5p"] Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.580636 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5psc\" (UniqueName: \"kubernetes.io/projected/6f441f5b-94ae-4932-b534-b8afa7716e44-kube-api-access-l5psc\") pod \"certified-operators-w9g5p\" (UID: \"6f441f5b-94ae-4932-b534-b8afa7716e44\") " pod="openshift-marketplace/certified-operators-w9g5p" Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.580690 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f441f5b-94ae-4932-b534-b8afa7716e44-utilities\") pod \"certified-operators-w9g5p\" (UID: \"6f441f5b-94ae-4932-b534-b8afa7716e44\") " pod="openshift-marketplace/certified-operators-w9g5p" Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.580722 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f441f5b-94ae-4932-b534-b8afa7716e44-catalog-content\") pod \"certified-operators-w9g5p\" (UID: \"6f441f5b-94ae-4932-b534-b8afa7716e44\") " pod="openshift-marketplace/certified-operators-w9g5p" Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.681902 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5psc\" (UniqueName: \"kubernetes.io/projected/6f441f5b-94ae-4932-b534-b8afa7716e44-kube-api-access-l5psc\") pod \"certified-operators-w9g5p\" (UID: \"6f441f5b-94ae-4932-b534-b8afa7716e44\") " pod="openshift-marketplace/certified-operators-w9g5p" Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.681988 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f441f5b-94ae-4932-b534-b8afa7716e44-utilities\") pod \"certified-operators-w9g5p\" (UID: \"6f441f5b-94ae-4932-b534-b8afa7716e44\") " pod="openshift-marketplace/certified-operators-w9g5p" Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.682041 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f441f5b-94ae-4932-b534-b8afa7716e44-catalog-content\") pod \"certified-operators-w9g5p\" (UID: \"6f441f5b-94ae-4932-b534-b8afa7716e44\") " pod="openshift-marketplace/certified-operators-w9g5p" Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.683039 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f441f5b-94ae-4932-b534-b8afa7716e44-utilities\") pod \"certified-operators-w9g5p\" (UID: \"6f441f5b-94ae-4932-b534-b8afa7716e44\") " pod="openshift-marketplace/certified-operators-w9g5p" Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.683093 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f441f5b-94ae-4932-b534-b8afa7716e44-catalog-content\") pod \"certified-operators-w9g5p\" (UID: \"6f441f5b-94ae-4932-b534-b8afa7716e44\") " pod="openshift-marketplace/certified-operators-w9g5p" Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.700818 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5psc\" (UniqueName: \"kubernetes.io/projected/6f441f5b-94ae-4932-b534-b8afa7716e44-kube-api-access-l5psc\") pod \"certified-operators-w9g5p\" (UID: \"6f441f5b-94ae-4932-b534-b8afa7716e44\") " pod="openshift-marketplace/certified-operators-w9g5p" Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.712919 4810 generic.go:334] "Generic (PLEG): container finished" podID="92abf8d2-574f-4375-bc2a-3436ab810e23" containerID="76696122146d81d98b2a5332a03623b7536e9640b8b3e90b54d619d4c3e04cda" exitCode=0 Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.712989 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpx9d" event={"ID":"92abf8d2-574f-4375-bc2a-3436ab810e23","Type":"ContainerDied","Data":"76696122146d81d98b2a5332a03623b7536e9640b8b3e90b54d619d4c3e04cda"} Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.713021 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpx9d" event={"ID":"92abf8d2-574f-4375-bc2a-3436ab810e23","Type":"ContainerStarted","Data":"c8a898448bdc0f5b4c3ca97c943b68af6120dde46ce5ee2ee4b940c26bb9e134"} Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.718832 4810 generic.go:334] "Generic (PLEG): container finished" podID="8c8f2b09-969f-4adf-8f45-8468d57e76be" containerID="d7c1cf93a0a3ff0f57b293426a4848dae687567b389a9967a5bd89bb7fe91bc0" exitCode=0 Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.718865 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jmlkh" event={"ID":"8c8f2b09-969f-4adf-8f45-8468d57e76be","Type":"ContainerDied","Data":"d7c1cf93a0a3ff0f57b293426a4848dae687567b389a9967a5bd89bb7fe91bc0"} Dec 01 14:40:49 crc kubenswrapper[4810]: I1201 14:40:49.878344 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w9g5p" Dec 01 14:40:50 crc kubenswrapper[4810]: I1201 14:40:50.253017 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w9g5p"] Dec 01 14:40:50 crc kubenswrapper[4810]: I1201 14:40:50.726464 4810 generic.go:334] "Generic (PLEG): container finished" podID="6f441f5b-94ae-4932-b534-b8afa7716e44" containerID="2a7cd1ac7379b2a904a3bd510f246057ee387912db8529cbfc8f4b1e76a34e74" exitCode=0 Dec 01 14:40:50 crc kubenswrapper[4810]: I1201 14:40:50.726561 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w9g5p" event={"ID":"6f441f5b-94ae-4932-b534-b8afa7716e44","Type":"ContainerDied","Data":"2a7cd1ac7379b2a904a3bd510f246057ee387912db8529cbfc8f4b1e76a34e74"} Dec 01 14:40:50 crc kubenswrapper[4810]: I1201 14:40:50.726594 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w9g5p" event={"ID":"6f441f5b-94ae-4932-b534-b8afa7716e44","Type":"ContainerStarted","Data":"99759ba92d8172a42251ac6e878885664a782c0f7453f030d173d7587710bb6c"} Dec 01 14:40:50 crc kubenswrapper[4810]: I1201 14:40:50.729156 4810 generic.go:334] "Generic (PLEG): container finished" podID="b8cdfd34-81c3-49ff-bba6-8b0d4192520d" containerID="3fed3998eee8132d07f1ab86b0636824160c217f05d8a833e912e22fc3c49ae2" exitCode=0 Dec 01 14:40:50 crc kubenswrapper[4810]: I1201 14:40:50.729234 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djh66" event={"ID":"b8cdfd34-81c3-49ff-bba6-8b0d4192520d","Type":"ContainerDied","Data":"3fed3998eee8132d07f1ab86b0636824160c217f05d8a833e912e22fc3c49ae2"} Dec 01 14:40:51 crc kubenswrapper[4810]: I1201 14:40:51.737948 4810 generic.go:334] "Generic (PLEG): container finished" podID="92abf8d2-574f-4375-bc2a-3436ab810e23" containerID="8aa21bb82a328fd9c77663983da79f3bc90b15b398bf87e79db90bccd484dd21" exitCode=0 Dec 01 14:40:51 crc kubenswrapper[4810]: I1201 14:40:51.738197 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpx9d" event={"ID":"92abf8d2-574f-4375-bc2a-3436ab810e23","Type":"ContainerDied","Data":"8aa21bb82a328fd9c77663983da79f3bc90b15b398bf87e79db90bccd484dd21"} Dec 01 14:40:51 crc kubenswrapper[4810]: I1201 14:40:51.741442 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w9g5p" event={"ID":"6f441f5b-94ae-4932-b534-b8afa7716e44","Type":"ContainerStarted","Data":"1aeb0490544d249d321a133e660bfcc3b3b4db53baa64a6dcf093b09110af6d2"} Dec 01 14:40:51 crc kubenswrapper[4810]: I1201 14:40:51.746731 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djh66" event={"ID":"b8cdfd34-81c3-49ff-bba6-8b0d4192520d","Type":"ContainerStarted","Data":"c190916a214ccbab94317963e484b4304815933e0346dcee4a94454baa18be9a"} Dec 01 14:40:51 crc kubenswrapper[4810]: I1201 14:40:51.748963 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jmlkh" event={"ID":"8c8f2b09-969f-4adf-8f45-8468d57e76be","Type":"ContainerStarted","Data":"13274765966e426ee4e6e069dd3899cdfc352c94caf024b6c01a9dffccb6a6d4"} Dec 01 14:40:51 crc kubenswrapper[4810]: I1201 14:40:51.788238 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jmlkh" podStartSLOduration=3.826484508 podStartE2EDuration="6.788222473s" podCreationTimestamp="2025-12-01 14:40:45 +0000 UTC" firstStartedPulling="2025-12-01 14:40:47.696619609 +0000 UTC m=+413.460129212" lastFinishedPulling="2025-12-01 14:40:50.658357584 +0000 UTC m=+416.421867177" observedRunningTime="2025-12-01 14:40:51.783950995 +0000 UTC m=+417.547460598" watchObservedRunningTime="2025-12-01 14:40:51.788222473 +0000 UTC m=+417.551732076" Dec 01 14:40:51 crc kubenswrapper[4810]: I1201 14:40:51.803224 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-djh66" podStartSLOduration=2.237340822 podStartE2EDuration="4.803194206s" podCreationTimestamp="2025-12-01 14:40:47 +0000 UTC" firstStartedPulling="2025-12-01 14:40:48.70136988 +0000 UTC m=+414.464879483" lastFinishedPulling="2025-12-01 14:40:51.267223254 +0000 UTC m=+417.030732867" observedRunningTime="2025-12-01 14:40:51.801568321 +0000 UTC m=+417.565077924" watchObservedRunningTime="2025-12-01 14:40:51.803194206 +0000 UTC m=+417.566703809" Dec 01 14:40:52 crc kubenswrapper[4810]: I1201 14:40:52.757616 4810 generic.go:334] "Generic (PLEG): container finished" podID="6f441f5b-94ae-4932-b534-b8afa7716e44" containerID="1aeb0490544d249d321a133e660bfcc3b3b4db53baa64a6dcf093b09110af6d2" exitCode=0 Dec 01 14:40:52 crc kubenswrapper[4810]: I1201 14:40:52.758061 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w9g5p" event={"ID":"6f441f5b-94ae-4932-b534-b8afa7716e44","Type":"ContainerDied","Data":"1aeb0490544d249d321a133e660bfcc3b3b4db53baa64a6dcf093b09110af6d2"} Dec 01 14:40:53 crc kubenswrapper[4810]: I1201 14:40:53.767549 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpx9d" event={"ID":"92abf8d2-574f-4375-bc2a-3436ab810e23","Type":"ContainerStarted","Data":"50b016a6c36e3f48c87f1fed5a95478a1c90242abf91ff3ef84deaf10fc17c95"} Dec 01 14:40:53 crc kubenswrapper[4810]: I1201 14:40:53.769548 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w9g5p" event={"ID":"6f441f5b-94ae-4932-b534-b8afa7716e44","Type":"ContainerStarted","Data":"69a4f4484793971aed60c5adede195f46aefdd27c0e3c367b2f8f03d48705816"} Dec 01 14:40:53 crc kubenswrapper[4810]: I1201 14:40:53.790854 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gpx9d" podStartSLOduration=2.991689902 podStartE2EDuration="5.790832286s" podCreationTimestamp="2025-12-01 14:40:48 +0000 UTC" firstStartedPulling="2025-12-01 14:40:49.714518952 +0000 UTC m=+415.478028555" lastFinishedPulling="2025-12-01 14:40:52.513661326 +0000 UTC m=+418.277170939" observedRunningTime="2025-12-01 14:40:53.787879205 +0000 UTC m=+419.551388828" watchObservedRunningTime="2025-12-01 14:40:53.790832286 +0000 UTC m=+419.554341889" Dec 01 14:40:53 crc kubenswrapper[4810]: I1201 14:40:53.814724 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w9g5p" podStartSLOduration=2.247847792 podStartE2EDuration="4.814702634s" podCreationTimestamp="2025-12-01 14:40:49 +0000 UTC" firstStartedPulling="2025-12-01 14:40:50.728789196 +0000 UTC m=+416.492298799" lastFinishedPulling="2025-12-01 14:40:53.295644038 +0000 UTC m=+419.059153641" observedRunningTime="2025-12-01 14:40:53.81419704 +0000 UTC m=+419.577706643" watchObservedRunningTime="2025-12-01 14:40:53.814702634 +0000 UTC m=+419.578212237" Dec 01 14:40:56 crc kubenswrapper[4810]: I1201 14:40:56.063964 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jmlkh" Dec 01 14:40:56 crc kubenswrapper[4810]: I1201 14:40:56.064239 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jmlkh" Dec 01 14:40:56 crc kubenswrapper[4810]: I1201 14:40:56.124185 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jmlkh" Dec 01 14:40:56 crc kubenswrapper[4810]: I1201 14:40:56.819399 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jmlkh" Dec 01 14:40:57 crc kubenswrapper[4810]: I1201 14:40:57.784766 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-djh66" Dec 01 14:40:57 crc kubenswrapper[4810]: I1201 14:40:57.785228 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-djh66" Dec 01 14:40:57 crc kubenswrapper[4810]: I1201 14:40:57.825981 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-djh66" Dec 01 14:40:58 crc kubenswrapper[4810]: I1201 14:40:58.469345 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gpx9d" Dec 01 14:40:58 crc kubenswrapper[4810]: I1201 14:40:58.470295 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gpx9d" Dec 01 14:40:58 crc kubenswrapper[4810]: I1201 14:40:58.507216 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gpx9d" Dec 01 14:40:58 crc kubenswrapper[4810]: I1201 14:40:58.826026 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-djh66" Dec 01 14:40:58 crc kubenswrapper[4810]: I1201 14:40:58.841532 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gpx9d" Dec 01 14:40:59 crc kubenswrapper[4810]: I1201 14:40:59.879519 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w9g5p" Dec 01 14:40:59 crc kubenswrapper[4810]: I1201 14:40:59.880361 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w9g5p" Dec 01 14:40:59 crc kubenswrapper[4810]: I1201 14:40:59.918405 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w9g5p" Dec 01 14:41:00 crc kubenswrapper[4810]: I1201 14:41:00.842785 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w9g5p" Dec 01 14:41:02 crc kubenswrapper[4810]: I1201 14:41:02.972302 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:41:02 crc kubenswrapper[4810]: I1201 14:41:02.972879 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:41:02 crc kubenswrapper[4810]: I1201 14:41:02.972994 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:41:02 crc kubenswrapper[4810]: I1201 14:41:02.973666 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3e2bceffac8b9b74a53ba6e40ccf495e62023cf75155f8f803b43b64a708ed90"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 14:41:02 crc kubenswrapper[4810]: I1201 14:41:02.973818 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://3e2bceffac8b9b74a53ba6e40ccf495e62023cf75155f8f803b43b64a708ed90" gracePeriod=600 Dec 01 14:41:03 crc kubenswrapper[4810]: I1201 14:41:03.819238 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="3e2bceffac8b9b74a53ba6e40ccf495e62023cf75155f8f803b43b64a708ed90" exitCode=0 Dec 01 14:41:03 crc kubenswrapper[4810]: I1201 14:41:03.819341 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"3e2bceffac8b9b74a53ba6e40ccf495e62023cf75155f8f803b43b64a708ed90"} Dec 01 14:41:03 crc kubenswrapper[4810]: I1201 14:41:03.819616 4810 scope.go:117] "RemoveContainer" containerID="3a6c23b9105e635d3c56e2f0b2b5a2b33ca47c238dd8b1f39a9fc914514a79a8" Dec 01 14:41:04 crc kubenswrapper[4810]: I1201 14:41:04.826998 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"d0afd8bc99702762b2f13098276fc28dc1bc7bfe7816bf6a33175529f7ac1768"} Dec 01 14:43:32 crc kubenswrapper[4810]: I1201 14:43:32.972417 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:43:32 crc kubenswrapper[4810]: I1201 14:43:32.973084 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:44:02 crc kubenswrapper[4810]: I1201 14:44:02.971703 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:44:02 crc kubenswrapper[4810]: I1201 14:44:02.972374 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:44:32 crc kubenswrapper[4810]: I1201 14:44:32.972523 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:44:32 crc kubenswrapper[4810]: I1201 14:44:32.973276 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:44:32 crc kubenswrapper[4810]: I1201 14:44:32.973332 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:44:32 crc kubenswrapper[4810]: I1201 14:44:32.974136 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d0afd8bc99702762b2f13098276fc28dc1bc7bfe7816bf6a33175529f7ac1768"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 14:44:32 crc kubenswrapper[4810]: I1201 14:44:32.974215 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://d0afd8bc99702762b2f13098276fc28dc1bc7bfe7816bf6a33175529f7ac1768" gracePeriod=600 Dec 01 14:44:34 crc kubenswrapper[4810]: I1201 14:44:34.067894 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="d0afd8bc99702762b2f13098276fc28dc1bc7bfe7816bf6a33175529f7ac1768" exitCode=0 Dec 01 14:44:34 crc kubenswrapper[4810]: I1201 14:44:34.068006 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"d0afd8bc99702762b2f13098276fc28dc1bc7bfe7816bf6a33175529f7ac1768"} Dec 01 14:44:34 crc kubenswrapper[4810]: I1201 14:44:34.068399 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"d9ccbffc9cbcd02b32617e3a533f267bf2f98615c6b47efb95676ef8e399d454"} Dec 01 14:44:34 crc kubenswrapper[4810]: I1201 14:44:34.068440 4810 scope.go:117] "RemoveContainer" containerID="3e2bceffac8b9b74a53ba6e40ccf495e62023cf75155f8f803b43b64a708ed90" Dec 01 14:45:00 crc kubenswrapper[4810]: I1201 14:45:00.180827 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt"] Dec 01 14:45:00 crc kubenswrapper[4810]: I1201 14:45:00.182057 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt" Dec 01 14:45:00 crc kubenswrapper[4810]: I1201 14:45:00.184902 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 14:45:00 crc kubenswrapper[4810]: I1201 14:45:00.184938 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 14:45:00 crc kubenswrapper[4810]: I1201 14:45:00.202121 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt"] Dec 01 14:45:00 crc kubenswrapper[4810]: I1201 14:45:00.303140 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2067075-0beb-49b1-be3f-2a1577801d4a-secret-volume\") pod \"collect-profiles-29410005-fcbwt\" (UID: \"f2067075-0beb-49b1-be3f-2a1577801d4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt" Dec 01 14:45:00 crc kubenswrapper[4810]: I1201 14:45:00.303418 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bn95\" (UniqueName: \"kubernetes.io/projected/f2067075-0beb-49b1-be3f-2a1577801d4a-kube-api-access-8bn95\") pod \"collect-profiles-29410005-fcbwt\" (UID: \"f2067075-0beb-49b1-be3f-2a1577801d4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt" Dec 01 14:45:00 crc kubenswrapper[4810]: I1201 14:45:00.303566 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2067075-0beb-49b1-be3f-2a1577801d4a-config-volume\") pod \"collect-profiles-29410005-fcbwt\" (UID: \"f2067075-0beb-49b1-be3f-2a1577801d4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt" Dec 01 14:45:00 crc kubenswrapper[4810]: I1201 14:45:00.405202 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bn95\" (UniqueName: \"kubernetes.io/projected/f2067075-0beb-49b1-be3f-2a1577801d4a-kube-api-access-8bn95\") pod \"collect-profiles-29410005-fcbwt\" (UID: \"f2067075-0beb-49b1-be3f-2a1577801d4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt" Dec 01 14:45:00 crc kubenswrapper[4810]: I1201 14:45:00.405252 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2067075-0beb-49b1-be3f-2a1577801d4a-config-volume\") pod \"collect-profiles-29410005-fcbwt\" (UID: \"f2067075-0beb-49b1-be3f-2a1577801d4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt" Dec 01 14:45:00 crc kubenswrapper[4810]: I1201 14:45:00.405347 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2067075-0beb-49b1-be3f-2a1577801d4a-secret-volume\") pod \"collect-profiles-29410005-fcbwt\" (UID: \"f2067075-0beb-49b1-be3f-2a1577801d4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt" Dec 01 14:45:00 crc kubenswrapper[4810]: I1201 14:45:00.406248 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2067075-0beb-49b1-be3f-2a1577801d4a-config-volume\") pod \"collect-profiles-29410005-fcbwt\" (UID: \"f2067075-0beb-49b1-be3f-2a1577801d4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt" Dec 01 14:45:00 crc kubenswrapper[4810]: I1201 14:45:00.420785 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2067075-0beb-49b1-be3f-2a1577801d4a-secret-volume\") pod \"collect-profiles-29410005-fcbwt\" (UID: \"f2067075-0beb-49b1-be3f-2a1577801d4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt" Dec 01 14:45:00 crc kubenswrapper[4810]: I1201 14:45:00.422014 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bn95\" (UniqueName: \"kubernetes.io/projected/f2067075-0beb-49b1-be3f-2a1577801d4a-kube-api-access-8bn95\") pod \"collect-profiles-29410005-fcbwt\" (UID: \"f2067075-0beb-49b1-be3f-2a1577801d4a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt" Dec 01 14:45:00 crc kubenswrapper[4810]: I1201 14:45:00.503136 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt" Dec 01 14:45:00 crc kubenswrapper[4810]: I1201 14:45:00.692671 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt"] Dec 01 14:45:01 crc kubenswrapper[4810]: I1201 14:45:01.248394 4810 generic.go:334] "Generic (PLEG): container finished" podID="f2067075-0beb-49b1-be3f-2a1577801d4a" containerID="c55bddc8877be42932cecfaf18ff65a25c0c87b44045e74379df3031e2f55a6b" exitCode=0 Dec 01 14:45:01 crc kubenswrapper[4810]: I1201 14:45:01.248543 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt" event={"ID":"f2067075-0beb-49b1-be3f-2a1577801d4a","Type":"ContainerDied","Data":"c55bddc8877be42932cecfaf18ff65a25c0c87b44045e74379df3031e2f55a6b"} Dec 01 14:45:01 crc kubenswrapper[4810]: I1201 14:45:01.248719 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt" event={"ID":"f2067075-0beb-49b1-be3f-2a1577801d4a","Type":"ContainerStarted","Data":"c7ff34fdda3924ae2f3ac839bd7084f29d958d8c6ab87e1b62fe0c5f731b9f3e"} Dec 01 14:45:02 crc kubenswrapper[4810]: I1201 14:45:02.532427 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt" Dec 01 14:45:02 crc kubenswrapper[4810]: I1201 14:45:02.632787 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bn95\" (UniqueName: \"kubernetes.io/projected/f2067075-0beb-49b1-be3f-2a1577801d4a-kube-api-access-8bn95\") pod \"f2067075-0beb-49b1-be3f-2a1577801d4a\" (UID: \"f2067075-0beb-49b1-be3f-2a1577801d4a\") " Dec 01 14:45:02 crc kubenswrapper[4810]: I1201 14:45:02.632912 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2067075-0beb-49b1-be3f-2a1577801d4a-config-volume\") pod \"f2067075-0beb-49b1-be3f-2a1577801d4a\" (UID: \"f2067075-0beb-49b1-be3f-2a1577801d4a\") " Dec 01 14:45:02 crc kubenswrapper[4810]: I1201 14:45:02.632974 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2067075-0beb-49b1-be3f-2a1577801d4a-secret-volume\") pod \"f2067075-0beb-49b1-be3f-2a1577801d4a\" (UID: \"f2067075-0beb-49b1-be3f-2a1577801d4a\") " Dec 01 14:45:02 crc kubenswrapper[4810]: I1201 14:45:02.634011 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2067075-0beb-49b1-be3f-2a1577801d4a-config-volume" (OuterVolumeSpecName: "config-volume") pod "f2067075-0beb-49b1-be3f-2a1577801d4a" (UID: "f2067075-0beb-49b1-be3f-2a1577801d4a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:45:02 crc kubenswrapper[4810]: I1201 14:45:02.638778 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2067075-0beb-49b1-be3f-2a1577801d4a-kube-api-access-8bn95" (OuterVolumeSpecName: "kube-api-access-8bn95") pod "f2067075-0beb-49b1-be3f-2a1577801d4a" (UID: "f2067075-0beb-49b1-be3f-2a1577801d4a"). InnerVolumeSpecName "kube-api-access-8bn95". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:45:02 crc kubenswrapper[4810]: I1201 14:45:02.639293 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2067075-0beb-49b1-be3f-2a1577801d4a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f2067075-0beb-49b1-be3f-2a1577801d4a" (UID: "f2067075-0beb-49b1-be3f-2a1577801d4a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:45:02 crc kubenswrapper[4810]: I1201 14:45:02.734905 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bn95\" (UniqueName: \"kubernetes.io/projected/f2067075-0beb-49b1-be3f-2a1577801d4a-kube-api-access-8bn95\") on node \"crc\" DevicePath \"\"" Dec 01 14:45:02 crc kubenswrapper[4810]: I1201 14:45:02.734954 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2067075-0beb-49b1-be3f-2a1577801d4a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 14:45:02 crc kubenswrapper[4810]: I1201 14:45:02.734968 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2067075-0beb-49b1-be3f-2a1577801d4a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 14:45:03 crc kubenswrapper[4810]: I1201 14:45:03.265402 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt" event={"ID":"f2067075-0beb-49b1-be3f-2a1577801d4a","Type":"ContainerDied","Data":"c7ff34fdda3924ae2f3ac839bd7084f29d958d8c6ab87e1b62fe0c5f731b9f3e"} Dec 01 14:45:03 crc kubenswrapper[4810]: I1201 14:45:03.265460 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7ff34fdda3924ae2f3ac839bd7084f29d958d8c6ab87e1b62fe0c5f731b9f3e" Dec 01 14:45:03 crc kubenswrapper[4810]: I1201 14:45:03.265621 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt" Dec 01 14:46:33 crc kubenswrapper[4810]: I1201 14:46:33.413393 4810 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 14:47:02 crc kubenswrapper[4810]: I1201 14:47:02.971943 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:47:02 crc kubenswrapper[4810]: I1201 14:47:02.972763 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:47:32 crc kubenswrapper[4810]: I1201 14:47:32.972087 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:47:32 crc kubenswrapper[4810]: I1201 14:47:32.972978 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.494364 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p7pvm"] Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.495152 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovn-controller" containerID="cri-o://a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0" gracePeriod=30 Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.495208 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce" gracePeriod=30 Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.495256 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="kube-rbac-proxy-node" containerID="cri-o://b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f" gracePeriod=30 Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.495231 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="northd" containerID="cri-o://31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb" gracePeriod=30 Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.495300 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovn-acl-logging" containerID="cri-o://727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026" gracePeriod=30 Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.495208 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="nbdb" containerID="cri-o://9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8" gracePeriod=30 Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.495422 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="sbdb" containerID="cri-o://abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d" gracePeriod=30 Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.583280 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovnkube-controller" containerID="cri-o://4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4" gracePeriod=30 Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.882322 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovnkube-controller/3.log" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.885126 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovn-acl-logging/0.log" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.885522 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovn-controller/0.log" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.886105 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.957729 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zjqcs"] Dec 01 14:47:35 crc kubenswrapper[4810]: E1201 14:47:35.957991 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="northd" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958009 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="northd" Dec 01 14:47:35 crc kubenswrapper[4810]: E1201 14:47:35.958027 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovnkube-controller" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958038 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovnkube-controller" Dec 01 14:47:35 crc kubenswrapper[4810]: E1201 14:47:35.958050 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovn-controller" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958059 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovn-controller" Dec 01 14:47:35 crc kubenswrapper[4810]: E1201 14:47:35.958070 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958078 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 14:47:35 crc kubenswrapper[4810]: E1201 14:47:35.958089 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovnkube-controller" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958097 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovnkube-controller" Dec 01 14:47:35 crc kubenswrapper[4810]: E1201 14:47:35.958108 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovnkube-controller" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958116 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovnkube-controller" Dec 01 14:47:35 crc kubenswrapper[4810]: E1201 14:47:35.958125 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2067075-0beb-49b1-be3f-2a1577801d4a" containerName="collect-profiles" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958133 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2067075-0beb-49b1-be3f-2a1577801d4a" containerName="collect-profiles" Dec 01 14:47:35 crc kubenswrapper[4810]: E1201 14:47:35.958147 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovn-acl-logging" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958156 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovn-acl-logging" Dec 01 14:47:35 crc kubenswrapper[4810]: E1201 14:47:35.958167 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="nbdb" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958176 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="nbdb" Dec 01 14:47:35 crc kubenswrapper[4810]: E1201 14:47:35.958188 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="kubecfg-setup" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958197 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="kubecfg-setup" Dec 01 14:47:35 crc kubenswrapper[4810]: E1201 14:47:35.958208 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="kube-rbac-proxy-node" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958216 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="kube-rbac-proxy-node" Dec 01 14:47:35 crc kubenswrapper[4810]: E1201 14:47:35.958232 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovnkube-controller" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958241 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovnkube-controller" Dec 01 14:47:35 crc kubenswrapper[4810]: E1201 14:47:35.958250 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="sbdb" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958259 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="sbdb" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958370 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2067075-0beb-49b1-be3f-2a1577801d4a" containerName="collect-profiles" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958384 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovnkube-controller" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958395 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovn-acl-logging" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958407 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="sbdb" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958419 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovnkube-controller" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958429 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovnkube-controller" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958438 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="kube-rbac-proxy-node" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958448 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovnkube-controller" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958459 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovn-controller" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958495 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958504 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="northd" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958514 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="nbdb" Dec 01 14:47:35 crc kubenswrapper[4810]: E1201 14:47:35.958631 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovnkube-controller" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958640 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovnkube-controller" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.958763 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerName="ovnkube-controller" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.960670 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.973266 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-node-log\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.973365 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/efba5556-97da-447f-9dda-40c2ea6d3e3d-ovnkube-config\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.973461 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-kubelet\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.973515 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.973544 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-cni-netd\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.973573 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-log-socket\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.973621 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7xmf\" (UniqueName: \"kubernetes.io/projected/efba5556-97da-447f-9dda-40c2ea6d3e3d-kube-api-access-q7xmf\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.973651 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-run-systemd\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.973693 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/efba5556-97da-447f-9dda-40c2ea6d3e3d-ovn-node-metrics-cert\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.973725 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-cni-bin\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.973765 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-var-lib-openvswitch\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.973791 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-etc-openvswitch\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.973830 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/efba5556-97da-447f-9dda-40c2ea6d3e3d-env-overrides\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.973859 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-slash\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.973891 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-run-netns\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.973915 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-run-openvswitch\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.973951 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-run-ovn\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.973985 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-run-ovn-kubernetes\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.974023 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/efba5556-97da-447f-9dda-40c2ea6d3e3d-ovnkube-script-lib\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.974066 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-systemd-units\") pod \"efba5556-97da-447f-9dda-40c2ea6d3e3d\" (UID: \"efba5556-97da-447f-9dda-40c2ea6d3e3d\") " Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.974353 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.974411 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-node-log" (OuterVolumeSpecName: "node-log") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.974598 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-log-socket" (OuterVolumeSpecName: "log-socket") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.974598 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.974646 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.974658 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.974695 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.974726 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.974753 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.974804 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.974887 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-slash" (OuterVolumeSpecName: "host-slash") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.974926 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.974966 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.975004 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.975200 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efba5556-97da-447f-9dda-40c2ea6d3e3d-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.975226 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efba5556-97da-447f-9dda-40c2ea6d3e3d-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.975632 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efba5556-97da-447f-9dda-40c2ea6d3e3d-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.980939 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efba5556-97da-447f-9dda-40c2ea6d3e3d-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.981515 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efba5556-97da-447f-9dda-40c2ea6d3e3d-kube-api-access-q7xmf" (OuterVolumeSpecName: "kube-api-access-q7xmf") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "kube-api-access-q7xmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:47:35 crc kubenswrapper[4810]: I1201 14:47:35.990347 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "efba5556-97da-447f-9dda-40c2ea6d3e3d" (UID: "efba5556-97da-447f-9dda-40c2ea6d3e3d"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075171 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-slash\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075246 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075288 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-log-socket\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075337 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a0197eb5-7117-4bc1-bc91-d29166873573-env-overrides\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075365 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-run-netns\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075389 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-run-ovn-kubernetes\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075414 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-run-ovn\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075436 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-etc-openvswitch\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075493 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a0197eb5-7117-4bc1-bc91-d29166873573-ovnkube-config\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075517 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r4wr\" (UniqueName: \"kubernetes.io/projected/a0197eb5-7117-4bc1-bc91-d29166873573-kube-api-access-5r4wr\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075543 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-cni-netd\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075563 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-run-systemd\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075592 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-kubelet\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075613 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-run-openvswitch\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075635 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-systemd-units\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075663 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a0197eb5-7117-4bc1-bc91-d29166873573-ovn-node-metrics-cert\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075692 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-var-lib-openvswitch\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075715 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-cni-bin\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075734 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-node-log\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075764 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a0197eb5-7117-4bc1-bc91-d29166873573-ovnkube-script-lib\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075819 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/efba5556-97da-447f-9dda-40c2ea6d3e3d-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075835 4810 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075848 4810 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075859 4810 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075872 4810 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/efba5556-97da-447f-9dda-40c2ea6d3e3d-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075885 4810 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-slash\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075896 4810 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075908 4810 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075920 4810 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075931 4810 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075942 4810 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/efba5556-97da-447f-9dda-40c2ea6d3e3d-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075954 4810 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075965 4810 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-node-log\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075976 4810 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/efba5556-97da-447f-9dda-40c2ea6d3e3d-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.075988 4810 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.076005 4810 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.076035 4810 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.076055 4810 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-log-socket\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.076073 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7xmf\" (UniqueName: \"kubernetes.io/projected/efba5556-97da-447f-9dda-40c2ea6d3e3d-kube-api-access-q7xmf\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.076089 4810 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/efba5556-97da-447f-9dda-40c2ea6d3e3d-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.089616 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-s4cc7_df7ecc53-6478-49cd-bfb6-6ee80e850a19/kube-multus/2.log" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.090121 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-s4cc7_df7ecc53-6478-49cd-bfb6-6ee80e850a19/kube-multus/1.log" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.090146 4810 generic.go:334] "Generic (PLEG): container finished" podID="df7ecc53-6478-49cd-bfb6-6ee80e850a19" containerID="13444fe245e31b72033991aab5b46b4bb2bbb61d1f5f07f09473175dfa7c1464" exitCode=2 Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.090189 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-s4cc7" event={"ID":"df7ecc53-6478-49cd-bfb6-6ee80e850a19","Type":"ContainerDied","Data":"13444fe245e31b72033991aab5b46b4bb2bbb61d1f5f07f09473175dfa7c1464"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.090227 4810 scope.go:117] "RemoveContainer" containerID="bd96824c21af3e98dc4565197e8c1e036e7a99ce8855d92ff01b023cea04e7d5" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.090724 4810 scope.go:117] "RemoveContainer" containerID="13444fe245e31b72033991aab5b46b4bb2bbb61d1f5f07f09473175dfa7c1464" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.096113 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovnkube-controller/3.log" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.105149 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovn-acl-logging/0.log" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.105907 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-p7pvm_efba5556-97da-447f-9dda-40c2ea6d3e3d/ovn-controller/0.log" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106460 4810 generic.go:334] "Generic (PLEG): container finished" podID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerID="4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4" exitCode=0 Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106515 4810 generic.go:334] "Generic (PLEG): container finished" podID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerID="abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d" exitCode=0 Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106528 4810 generic.go:334] "Generic (PLEG): container finished" podID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerID="9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8" exitCode=0 Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106537 4810 generic.go:334] "Generic (PLEG): container finished" podID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerID="31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb" exitCode=0 Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106580 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerDied","Data":"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106611 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106642 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerDied","Data":"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106662 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerDied","Data":"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106674 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerDied","Data":"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106685 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerDied","Data":"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106597 4810 generic.go:334] "Generic (PLEG): container finished" podID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerID="7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce" exitCode=0 Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106723 4810 generic.go:334] "Generic (PLEG): container finished" podID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerID="b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f" exitCode=0 Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106746 4810 generic.go:334] "Generic (PLEG): container finished" podID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerID="727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026" exitCode=143 Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106755 4810 generic.go:334] "Generic (PLEG): container finished" podID="efba5556-97da-447f-9dda-40c2ea6d3e3d" containerID="a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0" exitCode=143 Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106855 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerDied","Data":"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106905 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106920 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106926 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106933 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106939 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106945 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.106951 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.107035 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.107042 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.107048 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.107060 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerDied","Data":"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.107952 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.107965 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108112 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108123 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108129 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108137 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108143 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108148 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108154 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108160 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108232 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerDied","Data":"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108245 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108252 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108258 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108264 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108270 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108276 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108281 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108287 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108314 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108321 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108330 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-p7pvm" event={"ID":"efba5556-97da-447f-9dda-40c2ea6d3e3d","Type":"ContainerDied","Data":"5410ffd03955b8b5558821baec0e9cef19e443be2c486539b91a5a1a35cec10d"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108342 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108349 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108354 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108361 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108367 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108393 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108398 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108404 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108409 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.108416 4810 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96"} Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.141880 4810 scope.go:117] "RemoveContainer" containerID="4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.149647 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p7pvm"] Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.154965 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-p7pvm"] Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.160216 4810 scope.go:117] "RemoveContainer" containerID="d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.177216 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a0197eb5-7117-4bc1-bc91-d29166873573-ovnkube-config\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.177658 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r4wr\" (UniqueName: \"kubernetes.io/projected/a0197eb5-7117-4bc1-bc91-d29166873573-kube-api-access-5r4wr\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.177690 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-cni-netd\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.177709 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-run-systemd\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.177764 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-kubelet\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.177789 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-run-openvswitch\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.177834 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-systemd-units\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.177844 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-cni-netd\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.177859 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a0197eb5-7117-4bc1-bc91-d29166873573-ovn-node-metrics-cert\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.177906 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-run-systemd\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.177922 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-run-openvswitch\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.177971 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-kubelet\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.178007 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-var-lib-openvswitch\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.178039 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-cni-bin\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.178056 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-var-lib-openvswitch\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.178048 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-systemd-units\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.178063 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-node-log\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.178123 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-cni-bin\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.178200 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-node-log\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.178213 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a0197eb5-7117-4bc1-bc91-d29166873573-ovnkube-script-lib\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.178290 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-slash\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.178319 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.178344 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-slash\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.178369 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-log-socket\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.178401 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-log-socket\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.178371 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.178412 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a0197eb5-7117-4bc1-bc91-d29166873573-env-overrides\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.178529 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a0197eb5-7117-4bc1-bc91-d29166873573-ovnkube-config\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.178674 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-run-netns\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.178977 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a0197eb5-7117-4bc1-bc91-d29166873573-env-overrides\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.179005 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-run-netns\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.179041 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-run-ovn-kubernetes\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.179100 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-run-ovn\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.179119 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-etc-openvswitch\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.179156 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a0197eb5-7117-4bc1-bc91-d29166873573-ovnkube-script-lib\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.179250 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-host-run-ovn-kubernetes\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.179261 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-etc-openvswitch\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.179278 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a0197eb5-7117-4bc1-bc91-d29166873573-run-ovn\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.186421 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a0197eb5-7117-4bc1-bc91-d29166873573-ovn-node-metrics-cert\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.186586 4810 scope.go:117] "RemoveContainer" containerID="abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.199246 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r4wr\" (UniqueName: \"kubernetes.io/projected/a0197eb5-7117-4bc1-bc91-d29166873573-kube-api-access-5r4wr\") pod \"ovnkube-node-zjqcs\" (UID: \"a0197eb5-7117-4bc1-bc91-d29166873573\") " pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.209505 4810 scope.go:117] "RemoveContainer" containerID="9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.226040 4810 scope.go:117] "RemoveContainer" containerID="31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.241891 4810 scope.go:117] "RemoveContainer" containerID="7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.257610 4810 scope.go:117] "RemoveContainer" containerID="b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.274997 4810 scope.go:117] "RemoveContainer" containerID="727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.290453 4810 scope.go:117] "RemoveContainer" containerID="a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.305164 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.311977 4810 scope.go:117] "RemoveContainer" containerID="13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.332164 4810 scope.go:117] "RemoveContainer" containerID="4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4" Dec 01 14:47:36 crc kubenswrapper[4810]: E1201 14:47:36.333266 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4\": container with ID starting with 4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4 not found: ID does not exist" containerID="4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.333312 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4"} err="failed to get container status \"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4\": rpc error: code = NotFound desc = could not find container \"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4\": container with ID starting with 4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.333336 4810 scope.go:117] "RemoveContainer" containerID="d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe" Dec 01 14:47:36 crc kubenswrapper[4810]: E1201 14:47:36.333762 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe\": container with ID starting with d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe not found: ID does not exist" containerID="d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.333783 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe"} err="failed to get container status \"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe\": rpc error: code = NotFound desc = could not find container \"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe\": container with ID starting with d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.333797 4810 scope.go:117] "RemoveContainer" containerID="abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d" Dec 01 14:47:36 crc kubenswrapper[4810]: E1201 14:47:36.334162 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\": container with ID starting with abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d not found: ID does not exist" containerID="abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.334203 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d"} err="failed to get container status \"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\": rpc error: code = NotFound desc = could not find container \"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\": container with ID starting with abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.334222 4810 scope.go:117] "RemoveContainer" containerID="9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8" Dec 01 14:47:36 crc kubenswrapper[4810]: E1201 14:47:36.334442 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\": container with ID starting with 9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8 not found: ID does not exist" containerID="9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.334462 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8"} err="failed to get container status \"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\": rpc error: code = NotFound desc = could not find container \"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\": container with ID starting with 9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.334668 4810 scope.go:117] "RemoveContainer" containerID="31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb" Dec 01 14:47:36 crc kubenswrapper[4810]: E1201 14:47:36.334948 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\": container with ID starting with 31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb not found: ID does not exist" containerID="31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.334988 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb"} err="failed to get container status \"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\": rpc error: code = NotFound desc = could not find container \"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\": container with ID starting with 31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.335012 4810 scope.go:117] "RemoveContainer" containerID="7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce" Dec 01 14:47:36 crc kubenswrapper[4810]: E1201 14:47:36.335263 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\": container with ID starting with 7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce not found: ID does not exist" containerID="7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.335305 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce"} err="failed to get container status \"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\": rpc error: code = NotFound desc = could not find container \"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\": container with ID starting with 7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.335332 4810 scope.go:117] "RemoveContainer" containerID="b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f" Dec 01 14:47:36 crc kubenswrapper[4810]: E1201 14:47:36.335708 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\": container with ID starting with b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f not found: ID does not exist" containerID="b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.335740 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f"} err="failed to get container status \"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\": rpc error: code = NotFound desc = could not find container \"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\": container with ID starting with b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.335764 4810 scope.go:117] "RemoveContainer" containerID="727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026" Dec 01 14:47:36 crc kubenswrapper[4810]: E1201 14:47:36.336049 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\": container with ID starting with 727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026 not found: ID does not exist" containerID="727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.336089 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026"} err="failed to get container status \"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\": rpc error: code = NotFound desc = could not find container \"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\": container with ID starting with 727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.336112 4810 scope.go:117] "RemoveContainer" containerID="a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0" Dec 01 14:47:36 crc kubenswrapper[4810]: E1201 14:47:36.336936 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\": container with ID starting with a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0 not found: ID does not exist" containerID="a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.336972 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0"} err="failed to get container status \"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\": rpc error: code = NotFound desc = could not find container \"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\": container with ID starting with a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.337011 4810 scope.go:117] "RemoveContainer" containerID="13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96" Dec 01 14:47:36 crc kubenswrapper[4810]: E1201 14:47:36.337346 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\": container with ID starting with 13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96 not found: ID does not exist" containerID="13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.337372 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96"} err="failed to get container status \"13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\": rpc error: code = NotFound desc = could not find container \"13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\": container with ID starting with 13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.337389 4810 scope.go:117] "RemoveContainer" containerID="4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.337750 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4"} err="failed to get container status \"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4\": rpc error: code = NotFound desc = could not find container \"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4\": container with ID starting with 4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.337783 4810 scope.go:117] "RemoveContainer" containerID="d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.338112 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe"} err="failed to get container status \"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe\": rpc error: code = NotFound desc = could not find container \"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe\": container with ID starting with d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.338145 4810 scope.go:117] "RemoveContainer" containerID="abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.338540 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d"} err="failed to get container status \"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\": rpc error: code = NotFound desc = could not find container \"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\": container with ID starting with abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.338573 4810 scope.go:117] "RemoveContainer" containerID="9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.338930 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8"} err="failed to get container status \"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\": rpc error: code = NotFound desc = could not find container \"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\": container with ID starting with 9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.338964 4810 scope.go:117] "RemoveContainer" containerID="31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.339259 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb"} err="failed to get container status \"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\": rpc error: code = NotFound desc = could not find container \"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\": container with ID starting with 31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.339294 4810 scope.go:117] "RemoveContainer" containerID="7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.339676 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce"} err="failed to get container status \"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\": rpc error: code = NotFound desc = could not find container \"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\": container with ID starting with 7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.339699 4810 scope.go:117] "RemoveContainer" containerID="b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.339986 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f"} err="failed to get container status \"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\": rpc error: code = NotFound desc = could not find container \"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\": container with ID starting with b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.340011 4810 scope.go:117] "RemoveContainer" containerID="727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.340453 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026"} err="failed to get container status \"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\": rpc error: code = NotFound desc = could not find container \"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\": container with ID starting with 727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.340504 4810 scope.go:117] "RemoveContainer" containerID="a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.341798 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0"} err="failed to get container status \"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\": rpc error: code = NotFound desc = could not find container \"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\": container with ID starting with a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.341831 4810 scope.go:117] "RemoveContainer" containerID="13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.342184 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96"} err="failed to get container status \"13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\": rpc error: code = NotFound desc = could not find container \"13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\": container with ID starting with 13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.342216 4810 scope.go:117] "RemoveContainer" containerID="4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.342804 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4"} err="failed to get container status \"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4\": rpc error: code = NotFound desc = could not find container \"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4\": container with ID starting with 4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.342833 4810 scope.go:117] "RemoveContainer" containerID="d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.343241 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe"} err="failed to get container status \"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe\": rpc error: code = NotFound desc = could not find container \"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe\": container with ID starting with d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.343274 4810 scope.go:117] "RemoveContainer" containerID="abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d" Dec 01 14:47:36 crc kubenswrapper[4810]: W1201 14:47:36.343638 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0197eb5_7117_4bc1_bc91_d29166873573.slice/crio-c1bdb0d3469bc1041bdd659a9d916f20758ac4e7685a93476d63247a848d0838 WatchSource:0}: Error finding container c1bdb0d3469bc1041bdd659a9d916f20758ac4e7685a93476d63247a848d0838: Status 404 returned error can't find the container with id c1bdb0d3469bc1041bdd659a9d916f20758ac4e7685a93476d63247a848d0838 Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.343705 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d"} err="failed to get container status \"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\": rpc error: code = NotFound desc = could not find container \"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\": container with ID starting with abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.343726 4810 scope.go:117] "RemoveContainer" containerID="9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.344121 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8"} err="failed to get container status \"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\": rpc error: code = NotFound desc = could not find container \"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\": container with ID starting with 9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.344143 4810 scope.go:117] "RemoveContainer" containerID="31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.344410 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb"} err="failed to get container status \"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\": rpc error: code = NotFound desc = could not find container \"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\": container with ID starting with 31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.344436 4810 scope.go:117] "RemoveContainer" containerID="7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.344881 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce"} err="failed to get container status \"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\": rpc error: code = NotFound desc = could not find container \"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\": container with ID starting with 7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.344904 4810 scope.go:117] "RemoveContainer" containerID="b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.345404 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f"} err="failed to get container status \"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\": rpc error: code = NotFound desc = could not find container \"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\": container with ID starting with b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.345428 4810 scope.go:117] "RemoveContainer" containerID="727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.346058 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026"} err="failed to get container status \"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\": rpc error: code = NotFound desc = could not find container \"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\": container with ID starting with 727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.346093 4810 scope.go:117] "RemoveContainer" containerID="a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.346729 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0"} err="failed to get container status \"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\": rpc error: code = NotFound desc = could not find container \"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\": container with ID starting with a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.346784 4810 scope.go:117] "RemoveContainer" containerID="13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.347115 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96"} err="failed to get container status \"13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\": rpc error: code = NotFound desc = could not find container \"13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\": container with ID starting with 13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.347138 4810 scope.go:117] "RemoveContainer" containerID="4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.348189 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4"} err="failed to get container status \"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4\": rpc error: code = NotFound desc = could not find container \"4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4\": container with ID starting with 4f576bbc08d768e69a5d3cc345674425587135285e6909f2bbd2bcd687ecf1b4 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.348226 4810 scope.go:117] "RemoveContainer" containerID="d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.362538 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe"} err="failed to get container status \"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe\": rpc error: code = NotFound desc = could not find container \"d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe\": container with ID starting with d6694d3fc3cf7809b39212e5db0e137ecdb27ca0d80ad8663543ed9fa00df4fe not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.362640 4810 scope.go:117] "RemoveContainer" containerID="abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.363087 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d"} err="failed to get container status \"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\": rpc error: code = NotFound desc = could not find container \"abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d\": container with ID starting with abba0c7c3c4a8cf68e944636213ab103752d3ed3d91b88cb3e81a14dc7f8021d not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.363117 4810 scope.go:117] "RemoveContainer" containerID="9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.363488 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8"} err="failed to get container status \"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\": rpc error: code = NotFound desc = could not find container \"9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8\": container with ID starting with 9d74d38b074f806b91c9fc1d183d2a4cc68f70c7e6b25cba955a107d4c1bbbc8 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.363515 4810 scope.go:117] "RemoveContainer" containerID="31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.363987 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb"} err="failed to get container status \"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\": rpc error: code = NotFound desc = could not find container \"31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb\": container with ID starting with 31c61ac1fa11188a1c16293da10336149cf800bf00d3decb28d6a24941f2a4bb not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.364014 4810 scope.go:117] "RemoveContainer" containerID="7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.364742 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce"} err="failed to get container status \"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\": rpc error: code = NotFound desc = could not find container \"7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce\": container with ID starting with 7078394e6fdbf040d355e1634615718e37065d4dfe74179ec0ba2ba8ad2964ce not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.364774 4810 scope.go:117] "RemoveContainer" containerID="b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.366019 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f"} err="failed to get container status \"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\": rpc error: code = NotFound desc = could not find container \"b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f\": container with ID starting with b17557f26778093418129fffc9ed2ff46006649e984ad501a05e0a027d29ae2f not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.366057 4810 scope.go:117] "RemoveContainer" containerID="727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.366530 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026"} err="failed to get container status \"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\": rpc error: code = NotFound desc = could not find container \"727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026\": container with ID starting with 727c9ef132fbd7af4ad3d55778b90761ce44e65f965f8d6eb75ee52b009df026 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.366564 4810 scope.go:117] "RemoveContainer" containerID="a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.366998 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0"} err="failed to get container status \"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\": rpc error: code = NotFound desc = could not find container \"a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0\": container with ID starting with a86a8b595f6652d23773595a68785dbcaaf8b038b2005c4ae0cff1cea734aaf0 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.367031 4810 scope.go:117] "RemoveContainer" containerID="13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.367402 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96"} err="failed to get container status \"13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\": rpc error: code = NotFound desc = could not find container \"13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96\": container with ID starting with 13326f3ad75d0f1b1a498d436fd574848d1524ce42b7a1f4bfa589d48fc87d96 not found: ID does not exist" Dec 01 14:47:36 crc kubenswrapper[4810]: I1201 14:47:36.499683 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efba5556-97da-447f-9dda-40c2ea6d3e3d" path="/var/lib/kubelet/pods/efba5556-97da-447f-9dda-40c2ea6d3e3d/volumes" Dec 01 14:47:37 crc kubenswrapper[4810]: I1201 14:47:37.116514 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-s4cc7_df7ecc53-6478-49cd-bfb6-6ee80e850a19/kube-multus/2.log" Dec 01 14:47:37 crc kubenswrapper[4810]: I1201 14:47:37.116927 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-s4cc7" event={"ID":"df7ecc53-6478-49cd-bfb6-6ee80e850a19","Type":"ContainerStarted","Data":"636e89e3c76a61b408395eb4a5b899669ea54a83bf72115c1ea5f2181ef8eeb6"} Dec 01 14:47:37 crc kubenswrapper[4810]: I1201 14:47:37.120803 4810 generic.go:334] "Generic (PLEG): container finished" podID="a0197eb5-7117-4bc1-bc91-d29166873573" containerID="30bc4603d8a757fa74855930dc2a7ebfc6f07580404699eb61014c739431a5bd" exitCode=0 Dec 01 14:47:37 crc kubenswrapper[4810]: I1201 14:47:37.120852 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" event={"ID":"a0197eb5-7117-4bc1-bc91-d29166873573","Type":"ContainerDied","Data":"30bc4603d8a757fa74855930dc2a7ebfc6f07580404699eb61014c739431a5bd"} Dec 01 14:47:37 crc kubenswrapper[4810]: I1201 14:47:37.120900 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" event={"ID":"a0197eb5-7117-4bc1-bc91-d29166873573","Type":"ContainerStarted","Data":"c1bdb0d3469bc1041bdd659a9d916f20758ac4e7685a93476d63247a848d0838"} Dec 01 14:47:38 crc kubenswrapper[4810]: I1201 14:47:38.129501 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" event={"ID":"a0197eb5-7117-4bc1-bc91-d29166873573","Type":"ContainerStarted","Data":"5881bb0a0325b04b7da5461fa83fb6e5d9513cbe6c7f397e7f6743845d94155e"} Dec 01 14:47:38 crc kubenswrapper[4810]: I1201 14:47:38.130117 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" event={"ID":"a0197eb5-7117-4bc1-bc91-d29166873573","Type":"ContainerStarted","Data":"0d444a5f51dbd21797987aae6a2fb4049ee26a1f7777eb130f5d67e384ff6da1"} Dec 01 14:47:38 crc kubenswrapper[4810]: I1201 14:47:38.130128 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" event={"ID":"a0197eb5-7117-4bc1-bc91-d29166873573","Type":"ContainerStarted","Data":"ff487c31df96e09a694cf4a680474ca447fe094b55cda70aeeeb367358402436"} Dec 01 14:47:38 crc kubenswrapper[4810]: I1201 14:47:38.130138 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" event={"ID":"a0197eb5-7117-4bc1-bc91-d29166873573","Type":"ContainerStarted","Data":"b2f9db852ab73de47a2922b09ea3ea62373a339977abc1064dc5dc741a692f7c"} Dec 01 14:47:38 crc kubenswrapper[4810]: I1201 14:47:38.130146 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" event={"ID":"a0197eb5-7117-4bc1-bc91-d29166873573","Type":"ContainerStarted","Data":"1a5400a8ff1c90633a562e86c9e1cc84041fa55b68a3f249ca7f3a92b947106b"} Dec 01 14:47:38 crc kubenswrapper[4810]: I1201 14:47:38.130154 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" event={"ID":"a0197eb5-7117-4bc1-bc91-d29166873573","Type":"ContainerStarted","Data":"f648561cf91dde89b79c1a555e8b73ff10bc5adb9e97e5d90a74c7806aeda60e"} Dec 01 14:47:40 crc kubenswrapper[4810]: I1201 14:47:40.143730 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" event={"ID":"a0197eb5-7117-4bc1-bc91-d29166873573","Type":"ContainerStarted","Data":"8a539757b455d4304477248ae2a9e0ee838c562b81a7872815e34afeae579284"} Dec 01 14:47:42 crc kubenswrapper[4810]: I1201 14:47:42.517837 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-ss5nn"] Dec 01 14:47:42 crc kubenswrapper[4810]: I1201 14:47:42.518791 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:42 crc kubenswrapper[4810]: I1201 14:47:42.521329 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 01 14:47:42 crc kubenswrapper[4810]: I1201 14:47:42.521396 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 01 14:47:42 crc kubenswrapper[4810]: I1201 14:47:42.521640 4810 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-zgjj8" Dec 01 14:47:42 crc kubenswrapper[4810]: I1201 14:47:42.521693 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 01 14:47:42 crc kubenswrapper[4810]: I1201 14:47:42.659657 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e-crc-storage\") pod \"crc-storage-crc-ss5nn\" (UID: \"0a67e6d2-3f65-4e27-aefe-27f04f7ca75e\") " pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:42 crc kubenswrapper[4810]: I1201 14:47:42.659735 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e-node-mnt\") pod \"crc-storage-crc-ss5nn\" (UID: \"0a67e6d2-3f65-4e27-aefe-27f04f7ca75e\") " pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:42 crc kubenswrapper[4810]: I1201 14:47:42.659780 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b47c9\" (UniqueName: \"kubernetes.io/projected/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e-kube-api-access-b47c9\") pod \"crc-storage-crc-ss5nn\" (UID: \"0a67e6d2-3f65-4e27-aefe-27f04f7ca75e\") " pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:42 crc kubenswrapper[4810]: I1201 14:47:42.761148 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e-crc-storage\") pod \"crc-storage-crc-ss5nn\" (UID: \"0a67e6d2-3f65-4e27-aefe-27f04f7ca75e\") " pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:42 crc kubenswrapper[4810]: I1201 14:47:42.761252 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e-node-mnt\") pod \"crc-storage-crc-ss5nn\" (UID: \"0a67e6d2-3f65-4e27-aefe-27f04f7ca75e\") " pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:42 crc kubenswrapper[4810]: I1201 14:47:42.761861 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b47c9\" (UniqueName: \"kubernetes.io/projected/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e-kube-api-access-b47c9\") pod \"crc-storage-crc-ss5nn\" (UID: \"0a67e6d2-3f65-4e27-aefe-27f04f7ca75e\") " pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:42 crc kubenswrapper[4810]: I1201 14:47:42.762357 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e-node-mnt\") pod \"crc-storage-crc-ss5nn\" (UID: \"0a67e6d2-3f65-4e27-aefe-27f04f7ca75e\") " pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:42 crc kubenswrapper[4810]: I1201 14:47:42.762627 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e-crc-storage\") pod \"crc-storage-crc-ss5nn\" (UID: \"0a67e6d2-3f65-4e27-aefe-27f04f7ca75e\") " pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:42 crc kubenswrapper[4810]: I1201 14:47:42.782186 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b47c9\" (UniqueName: \"kubernetes.io/projected/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e-kube-api-access-b47c9\") pod \"crc-storage-crc-ss5nn\" (UID: \"0a67e6d2-3f65-4e27-aefe-27f04f7ca75e\") " pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:42 crc kubenswrapper[4810]: I1201 14:47:42.836286 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:42 crc kubenswrapper[4810]: E1201 14:47:42.860148 4810 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-ss5nn_crc-storage_0a67e6d2-3f65-4e27-aefe-27f04f7ca75e_0(8e7f02694f530101bcb616b20af4e4da2f18abb5a404af764179c98b6cd36879): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 14:47:42 crc kubenswrapper[4810]: E1201 14:47:42.860235 4810 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-ss5nn_crc-storage_0a67e6d2-3f65-4e27-aefe-27f04f7ca75e_0(8e7f02694f530101bcb616b20af4e4da2f18abb5a404af764179c98b6cd36879): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:42 crc kubenswrapper[4810]: E1201 14:47:42.860263 4810 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-ss5nn_crc-storage_0a67e6d2-3f65-4e27-aefe-27f04f7ca75e_0(8e7f02694f530101bcb616b20af4e4da2f18abb5a404af764179c98b6cd36879): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:42 crc kubenswrapper[4810]: E1201 14:47:42.860325 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-ss5nn_crc-storage(0a67e6d2-3f65-4e27-aefe-27f04f7ca75e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-ss5nn_crc-storage(0a67e6d2-3f65-4e27-aefe-27f04f7ca75e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-ss5nn_crc-storage_0a67e6d2-3f65-4e27-aefe-27f04f7ca75e_0(8e7f02694f530101bcb616b20af4e4da2f18abb5a404af764179c98b6cd36879): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-ss5nn" podUID="0a67e6d2-3f65-4e27-aefe-27f04f7ca75e" Dec 01 14:47:43 crc kubenswrapper[4810]: I1201 14:47:43.168897 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" event={"ID":"a0197eb5-7117-4bc1-bc91-d29166873573","Type":"ContainerStarted","Data":"d11f7be97ef4fc1280ff05a4fab4064cec7ca5184c802758bd819cccb3ade787"} Dec 01 14:47:43 crc kubenswrapper[4810]: I1201 14:47:43.169333 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:43 crc kubenswrapper[4810]: I1201 14:47:43.169350 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:43 crc kubenswrapper[4810]: I1201 14:47:43.169363 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:43 crc kubenswrapper[4810]: I1201 14:47:43.201185 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" podStartSLOduration=8.201165469 podStartE2EDuration="8.201165469s" podCreationTimestamp="2025-12-01 14:47:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:47:43.197524089 +0000 UTC m=+828.961033692" watchObservedRunningTime="2025-12-01 14:47:43.201165469 +0000 UTC m=+828.964675072" Dec 01 14:47:43 crc kubenswrapper[4810]: I1201 14:47:43.208211 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:43 crc kubenswrapper[4810]: I1201 14:47:43.213062 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:47:43 crc kubenswrapper[4810]: I1201 14:47:43.784028 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-ss5nn"] Dec 01 14:47:43 crc kubenswrapper[4810]: I1201 14:47:43.784137 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:43 crc kubenswrapper[4810]: I1201 14:47:43.784597 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:43 crc kubenswrapper[4810]: E1201 14:47:43.816278 4810 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-ss5nn_crc-storage_0a67e6d2-3f65-4e27-aefe-27f04f7ca75e_0(3ce748a793e93e6cc34c498a487f7f3ee7bffb6530c11fc1eb70a284d03737a2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 14:47:43 crc kubenswrapper[4810]: E1201 14:47:43.816339 4810 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-ss5nn_crc-storage_0a67e6d2-3f65-4e27-aefe-27f04f7ca75e_0(3ce748a793e93e6cc34c498a487f7f3ee7bffb6530c11fc1eb70a284d03737a2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:43 crc kubenswrapper[4810]: E1201 14:47:43.816358 4810 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-ss5nn_crc-storage_0a67e6d2-3f65-4e27-aefe-27f04f7ca75e_0(3ce748a793e93e6cc34c498a487f7f3ee7bffb6530c11fc1eb70a284d03737a2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:43 crc kubenswrapper[4810]: E1201 14:47:43.816394 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-ss5nn_crc-storage(0a67e6d2-3f65-4e27-aefe-27f04f7ca75e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-ss5nn_crc-storage(0a67e6d2-3f65-4e27-aefe-27f04f7ca75e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-ss5nn_crc-storage_0a67e6d2-3f65-4e27-aefe-27f04f7ca75e_0(3ce748a793e93e6cc34c498a487f7f3ee7bffb6530c11fc1eb70a284d03737a2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-ss5nn" podUID="0a67e6d2-3f65-4e27-aefe-27f04f7ca75e" Dec 01 14:47:57 crc kubenswrapper[4810]: I1201 14:47:57.490890 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:57 crc kubenswrapper[4810]: I1201 14:47:57.493075 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:47:57 crc kubenswrapper[4810]: I1201 14:47:57.672802 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-ss5nn"] Dec 01 14:47:57 crc kubenswrapper[4810]: I1201 14:47:57.693171 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 14:47:58 crc kubenswrapper[4810]: I1201 14:47:58.262107 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-ss5nn" event={"ID":"0a67e6d2-3f65-4e27-aefe-27f04f7ca75e","Type":"ContainerStarted","Data":"9c5cb33897c4a5dee2bba33044593c535bb873453b1d9b985a51b7365d4b6fe9"} Dec 01 14:47:59 crc kubenswrapper[4810]: I1201 14:47:59.270614 4810 generic.go:334] "Generic (PLEG): container finished" podID="0a67e6d2-3f65-4e27-aefe-27f04f7ca75e" containerID="66eed9ca83c9c0f52c6ec00f5d3a71b9b0613bb7ef079010557011bf014c6898" exitCode=0 Dec 01 14:47:59 crc kubenswrapper[4810]: I1201 14:47:59.270711 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-ss5nn" event={"ID":"0a67e6d2-3f65-4e27-aefe-27f04f7ca75e","Type":"ContainerDied","Data":"66eed9ca83c9c0f52c6ec00f5d3a71b9b0613bb7ef079010557011bf014c6898"} Dec 01 14:48:00 crc kubenswrapper[4810]: I1201 14:48:00.564098 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:48:00 crc kubenswrapper[4810]: I1201 14:48:00.619222 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e-crc-storage\") pod \"0a67e6d2-3f65-4e27-aefe-27f04f7ca75e\" (UID: \"0a67e6d2-3f65-4e27-aefe-27f04f7ca75e\") " Dec 01 14:48:00 crc kubenswrapper[4810]: I1201 14:48:00.619293 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b47c9\" (UniqueName: \"kubernetes.io/projected/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e-kube-api-access-b47c9\") pod \"0a67e6d2-3f65-4e27-aefe-27f04f7ca75e\" (UID: \"0a67e6d2-3f65-4e27-aefe-27f04f7ca75e\") " Dec 01 14:48:00 crc kubenswrapper[4810]: I1201 14:48:00.619321 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e-node-mnt\") pod \"0a67e6d2-3f65-4e27-aefe-27f04f7ca75e\" (UID: \"0a67e6d2-3f65-4e27-aefe-27f04f7ca75e\") " Dec 01 14:48:00 crc kubenswrapper[4810]: I1201 14:48:00.619766 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "0a67e6d2-3f65-4e27-aefe-27f04f7ca75e" (UID: "0a67e6d2-3f65-4e27-aefe-27f04f7ca75e"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:48:00 crc kubenswrapper[4810]: I1201 14:48:00.624217 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e-kube-api-access-b47c9" (OuterVolumeSpecName: "kube-api-access-b47c9") pod "0a67e6d2-3f65-4e27-aefe-27f04f7ca75e" (UID: "0a67e6d2-3f65-4e27-aefe-27f04f7ca75e"). InnerVolumeSpecName "kube-api-access-b47c9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:48:00 crc kubenswrapper[4810]: I1201 14:48:00.638567 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "0a67e6d2-3f65-4e27-aefe-27f04f7ca75e" (UID: "0a67e6d2-3f65-4e27-aefe-27f04f7ca75e"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:48:00 crc kubenswrapper[4810]: I1201 14:48:00.721237 4810 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 01 14:48:00 crc kubenswrapper[4810]: I1201 14:48:00.721274 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b47c9\" (UniqueName: \"kubernetes.io/projected/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e-kube-api-access-b47c9\") on node \"crc\" DevicePath \"\"" Dec 01 14:48:00 crc kubenswrapper[4810]: I1201 14:48:00.721285 4810 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 01 14:48:01 crc kubenswrapper[4810]: I1201 14:48:01.284555 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-ss5nn" event={"ID":"0a67e6d2-3f65-4e27-aefe-27f04f7ca75e","Type":"ContainerDied","Data":"9c5cb33897c4a5dee2bba33044593c535bb873453b1d9b985a51b7365d4b6fe9"} Dec 01 14:48:01 crc kubenswrapper[4810]: I1201 14:48:01.284619 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c5cb33897c4a5dee2bba33044593c535bb873453b1d9b985a51b7365d4b6fe9" Dec 01 14:48:01 crc kubenswrapper[4810]: I1201 14:48:01.284581 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-ss5nn" Dec 01 14:48:02 crc kubenswrapper[4810]: I1201 14:48:02.971839 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:48:02 crc kubenswrapper[4810]: I1201 14:48:02.971910 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:48:02 crc kubenswrapper[4810]: I1201 14:48:02.971953 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:48:02 crc kubenswrapper[4810]: I1201 14:48:02.972523 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d9ccbffc9cbcd02b32617e3a533f267bf2f98615c6b47efb95676ef8e399d454"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 14:48:02 crc kubenswrapper[4810]: I1201 14:48:02.972587 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://d9ccbffc9cbcd02b32617e3a533f267bf2f98615c6b47efb95676ef8e399d454" gracePeriod=600 Dec 01 14:48:03 crc kubenswrapper[4810]: I1201 14:48:03.295837 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="d9ccbffc9cbcd02b32617e3a533f267bf2f98615c6b47efb95676ef8e399d454" exitCode=0 Dec 01 14:48:03 crc kubenswrapper[4810]: I1201 14:48:03.295909 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"d9ccbffc9cbcd02b32617e3a533f267bf2f98615c6b47efb95676ef8e399d454"} Dec 01 14:48:03 crc kubenswrapper[4810]: I1201 14:48:03.296089 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"7d9c11d61c66e363668182137baa230beaee1296f3efd5bf440ccc7a13833e7c"} Dec 01 14:48:03 crc kubenswrapper[4810]: I1201 14:48:03.296109 4810 scope.go:117] "RemoveContainer" containerID="d0afd8bc99702762b2f13098276fc28dc1bc7bfe7816bf6a33175529f7ac1768" Dec 01 14:48:06 crc kubenswrapper[4810]: I1201 14:48:06.335187 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zjqcs" Dec 01 14:48:08 crc kubenswrapper[4810]: I1201 14:48:08.391307 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz"] Dec 01 14:48:08 crc kubenswrapper[4810]: E1201 14:48:08.391836 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a67e6d2-3f65-4e27-aefe-27f04f7ca75e" containerName="storage" Dec 01 14:48:08 crc kubenswrapper[4810]: I1201 14:48:08.391848 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a67e6d2-3f65-4e27-aefe-27f04f7ca75e" containerName="storage" Dec 01 14:48:08 crc kubenswrapper[4810]: I1201 14:48:08.391936 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a67e6d2-3f65-4e27-aefe-27f04f7ca75e" containerName="storage" Dec 01 14:48:08 crc kubenswrapper[4810]: I1201 14:48:08.392591 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" Dec 01 14:48:08 crc kubenswrapper[4810]: I1201 14:48:08.394589 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 14:48:08 crc kubenswrapper[4810]: I1201 14:48:08.403680 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz"] Dec 01 14:48:08 crc kubenswrapper[4810]: I1201 14:48:08.419855 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b0ed0930-4327-4772-9a1c-2a0ae55f529c-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz\" (UID: \"b0ed0930-4327-4772-9a1c-2a0ae55f529c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" Dec 01 14:48:08 crc kubenswrapper[4810]: I1201 14:48:08.419903 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl9s6\" (UniqueName: \"kubernetes.io/projected/b0ed0930-4327-4772-9a1c-2a0ae55f529c-kube-api-access-rl9s6\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz\" (UID: \"b0ed0930-4327-4772-9a1c-2a0ae55f529c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" Dec 01 14:48:08 crc kubenswrapper[4810]: I1201 14:48:08.419976 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b0ed0930-4327-4772-9a1c-2a0ae55f529c-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz\" (UID: \"b0ed0930-4327-4772-9a1c-2a0ae55f529c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" Dec 01 14:48:08 crc kubenswrapper[4810]: I1201 14:48:08.520981 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b0ed0930-4327-4772-9a1c-2a0ae55f529c-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz\" (UID: \"b0ed0930-4327-4772-9a1c-2a0ae55f529c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" Dec 01 14:48:08 crc kubenswrapper[4810]: I1201 14:48:08.521060 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b0ed0930-4327-4772-9a1c-2a0ae55f529c-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz\" (UID: \"b0ed0930-4327-4772-9a1c-2a0ae55f529c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" Dec 01 14:48:08 crc kubenswrapper[4810]: I1201 14:48:08.521089 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl9s6\" (UniqueName: \"kubernetes.io/projected/b0ed0930-4327-4772-9a1c-2a0ae55f529c-kube-api-access-rl9s6\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz\" (UID: \"b0ed0930-4327-4772-9a1c-2a0ae55f529c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" Dec 01 14:48:08 crc kubenswrapper[4810]: I1201 14:48:08.522581 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b0ed0930-4327-4772-9a1c-2a0ae55f529c-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz\" (UID: \"b0ed0930-4327-4772-9a1c-2a0ae55f529c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" Dec 01 14:48:08 crc kubenswrapper[4810]: I1201 14:48:08.522863 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b0ed0930-4327-4772-9a1c-2a0ae55f529c-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz\" (UID: \"b0ed0930-4327-4772-9a1c-2a0ae55f529c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" Dec 01 14:48:08 crc kubenswrapper[4810]: I1201 14:48:08.548444 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl9s6\" (UniqueName: \"kubernetes.io/projected/b0ed0930-4327-4772-9a1c-2a0ae55f529c-kube-api-access-rl9s6\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz\" (UID: \"b0ed0930-4327-4772-9a1c-2a0ae55f529c\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" Dec 01 14:48:08 crc kubenswrapper[4810]: I1201 14:48:08.712096 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" Dec 01 14:48:08 crc kubenswrapper[4810]: I1201 14:48:08.930044 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz"] Dec 01 14:48:09 crc kubenswrapper[4810]: I1201 14:48:09.329634 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" event={"ID":"b0ed0930-4327-4772-9a1c-2a0ae55f529c","Type":"ContainerStarted","Data":"764db8f8d8159ec4f9f8cf717d9aca7edfd047e0b47456bb2fb10aabb07f8f31"} Dec 01 14:48:09 crc kubenswrapper[4810]: I1201 14:48:09.329689 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" event={"ID":"b0ed0930-4327-4772-9a1c-2a0ae55f529c","Type":"ContainerStarted","Data":"2da0dcf4cd5f38ed16a844569a54382d92d8925b8727100a27ed0b3b8eb80a33"} Dec 01 14:48:10 crc kubenswrapper[4810]: I1201 14:48:10.335820 4810 generic.go:334] "Generic (PLEG): container finished" podID="b0ed0930-4327-4772-9a1c-2a0ae55f529c" containerID="764db8f8d8159ec4f9f8cf717d9aca7edfd047e0b47456bb2fb10aabb07f8f31" exitCode=0 Dec 01 14:48:10 crc kubenswrapper[4810]: I1201 14:48:10.336164 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" event={"ID":"b0ed0930-4327-4772-9a1c-2a0ae55f529c","Type":"ContainerDied","Data":"764db8f8d8159ec4f9f8cf717d9aca7edfd047e0b47456bb2fb10aabb07f8f31"} Dec 01 14:48:10 crc kubenswrapper[4810]: I1201 14:48:10.364235 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sq6nr"] Dec 01 14:48:10 crc kubenswrapper[4810]: I1201 14:48:10.368287 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sq6nr" Dec 01 14:48:10 crc kubenswrapper[4810]: I1201 14:48:10.375795 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sq6nr"] Dec 01 14:48:10 crc kubenswrapper[4810]: I1201 14:48:10.446133 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7svbb\" (UniqueName: \"kubernetes.io/projected/19288c1f-8d88-4e3a-85e3-efbaea495b77-kube-api-access-7svbb\") pod \"redhat-operators-sq6nr\" (UID: \"19288c1f-8d88-4e3a-85e3-efbaea495b77\") " pod="openshift-marketplace/redhat-operators-sq6nr" Dec 01 14:48:10 crc kubenswrapper[4810]: I1201 14:48:10.446230 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19288c1f-8d88-4e3a-85e3-efbaea495b77-utilities\") pod \"redhat-operators-sq6nr\" (UID: \"19288c1f-8d88-4e3a-85e3-efbaea495b77\") " pod="openshift-marketplace/redhat-operators-sq6nr" Dec 01 14:48:10 crc kubenswrapper[4810]: I1201 14:48:10.446252 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19288c1f-8d88-4e3a-85e3-efbaea495b77-catalog-content\") pod \"redhat-operators-sq6nr\" (UID: \"19288c1f-8d88-4e3a-85e3-efbaea495b77\") " pod="openshift-marketplace/redhat-operators-sq6nr" Dec 01 14:48:10 crc kubenswrapper[4810]: I1201 14:48:10.547779 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19288c1f-8d88-4e3a-85e3-efbaea495b77-utilities\") pod \"redhat-operators-sq6nr\" (UID: \"19288c1f-8d88-4e3a-85e3-efbaea495b77\") " pod="openshift-marketplace/redhat-operators-sq6nr" Dec 01 14:48:10 crc kubenswrapper[4810]: I1201 14:48:10.547842 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19288c1f-8d88-4e3a-85e3-efbaea495b77-catalog-content\") pod \"redhat-operators-sq6nr\" (UID: \"19288c1f-8d88-4e3a-85e3-efbaea495b77\") " pod="openshift-marketplace/redhat-operators-sq6nr" Dec 01 14:48:10 crc kubenswrapper[4810]: I1201 14:48:10.547872 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7svbb\" (UniqueName: \"kubernetes.io/projected/19288c1f-8d88-4e3a-85e3-efbaea495b77-kube-api-access-7svbb\") pod \"redhat-operators-sq6nr\" (UID: \"19288c1f-8d88-4e3a-85e3-efbaea495b77\") " pod="openshift-marketplace/redhat-operators-sq6nr" Dec 01 14:48:10 crc kubenswrapper[4810]: I1201 14:48:10.548591 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19288c1f-8d88-4e3a-85e3-efbaea495b77-utilities\") pod \"redhat-operators-sq6nr\" (UID: \"19288c1f-8d88-4e3a-85e3-efbaea495b77\") " pod="openshift-marketplace/redhat-operators-sq6nr" Dec 01 14:48:10 crc kubenswrapper[4810]: I1201 14:48:10.548650 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19288c1f-8d88-4e3a-85e3-efbaea495b77-catalog-content\") pod \"redhat-operators-sq6nr\" (UID: \"19288c1f-8d88-4e3a-85e3-efbaea495b77\") " pod="openshift-marketplace/redhat-operators-sq6nr" Dec 01 14:48:10 crc kubenswrapper[4810]: I1201 14:48:10.567139 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7svbb\" (UniqueName: \"kubernetes.io/projected/19288c1f-8d88-4e3a-85e3-efbaea495b77-kube-api-access-7svbb\") pod \"redhat-operators-sq6nr\" (UID: \"19288c1f-8d88-4e3a-85e3-efbaea495b77\") " pod="openshift-marketplace/redhat-operators-sq6nr" Dec 01 14:48:10 crc kubenswrapper[4810]: I1201 14:48:10.698771 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sq6nr" Dec 01 14:48:10 crc kubenswrapper[4810]: I1201 14:48:10.894891 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sq6nr"] Dec 01 14:48:11 crc kubenswrapper[4810]: I1201 14:48:11.342517 4810 generic.go:334] "Generic (PLEG): container finished" podID="19288c1f-8d88-4e3a-85e3-efbaea495b77" containerID="7a4bc98ea751885f6d257cfc4ed4579cdd450ea55b5ac511cb6fac373903806f" exitCode=0 Dec 01 14:48:11 crc kubenswrapper[4810]: I1201 14:48:11.342559 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sq6nr" event={"ID":"19288c1f-8d88-4e3a-85e3-efbaea495b77","Type":"ContainerDied","Data":"7a4bc98ea751885f6d257cfc4ed4579cdd450ea55b5ac511cb6fac373903806f"} Dec 01 14:48:11 crc kubenswrapper[4810]: I1201 14:48:11.342585 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sq6nr" event={"ID":"19288c1f-8d88-4e3a-85e3-efbaea495b77","Type":"ContainerStarted","Data":"8e9b8f5e0cb6ac32aa2558854ff2888dab0cbd4196ac34baff45b2f0e454b67f"} Dec 01 14:48:12 crc kubenswrapper[4810]: I1201 14:48:12.359421 4810 generic.go:334] "Generic (PLEG): container finished" podID="b0ed0930-4327-4772-9a1c-2a0ae55f529c" containerID="71f309ec4b7aed6f378aa8011cba94882a681af836486cac6651ef495f0285bd" exitCode=0 Dec 01 14:48:12 crc kubenswrapper[4810]: I1201 14:48:12.359544 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" event={"ID":"b0ed0930-4327-4772-9a1c-2a0ae55f529c","Type":"ContainerDied","Data":"71f309ec4b7aed6f378aa8011cba94882a681af836486cac6651ef495f0285bd"} Dec 01 14:48:13 crc kubenswrapper[4810]: I1201 14:48:13.377719 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sq6nr" event={"ID":"19288c1f-8d88-4e3a-85e3-efbaea495b77","Type":"ContainerStarted","Data":"873ead448951c49d8683e5125cc1fbbe9a5347283041d08824f5ff296ad4617e"} Dec 01 14:48:13 crc kubenswrapper[4810]: I1201 14:48:13.386829 4810 generic.go:334] "Generic (PLEG): container finished" podID="b0ed0930-4327-4772-9a1c-2a0ae55f529c" containerID="1394a95320ff7e96ca0e8b4b7394cbb391bdea6982a1b2588837f2b840322a8a" exitCode=0 Dec 01 14:48:13 crc kubenswrapper[4810]: I1201 14:48:13.386926 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" event={"ID":"b0ed0930-4327-4772-9a1c-2a0ae55f529c","Type":"ContainerDied","Data":"1394a95320ff7e96ca0e8b4b7394cbb391bdea6982a1b2588837f2b840322a8a"} Dec 01 14:48:14 crc kubenswrapper[4810]: I1201 14:48:14.400109 4810 generic.go:334] "Generic (PLEG): container finished" podID="19288c1f-8d88-4e3a-85e3-efbaea495b77" containerID="873ead448951c49d8683e5125cc1fbbe9a5347283041d08824f5ff296ad4617e" exitCode=0 Dec 01 14:48:14 crc kubenswrapper[4810]: I1201 14:48:14.400464 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sq6nr" event={"ID":"19288c1f-8d88-4e3a-85e3-efbaea495b77","Type":"ContainerDied","Data":"873ead448951c49d8683e5125cc1fbbe9a5347283041d08824f5ff296ad4617e"} Dec 01 14:48:14 crc kubenswrapper[4810]: I1201 14:48:14.652797 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" Dec 01 14:48:14 crc kubenswrapper[4810]: I1201 14:48:14.713227 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b0ed0930-4327-4772-9a1c-2a0ae55f529c-util\") pod \"b0ed0930-4327-4772-9a1c-2a0ae55f529c\" (UID: \"b0ed0930-4327-4772-9a1c-2a0ae55f529c\") " Dec 01 14:48:14 crc kubenswrapper[4810]: I1201 14:48:14.713349 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b0ed0930-4327-4772-9a1c-2a0ae55f529c-bundle\") pod \"b0ed0930-4327-4772-9a1c-2a0ae55f529c\" (UID: \"b0ed0930-4327-4772-9a1c-2a0ae55f529c\") " Dec 01 14:48:14 crc kubenswrapper[4810]: I1201 14:48:14.713394 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rl9s6\" (UniqueName: \"kubernetes.io/projected/b0ed0930-4327-4772-9a1c-2a0ae55f529c-kube-api-access-rl9s6\") pod \"b0ed0930-4327-4772-9a1c-2a0ae55f529c\" (UID: \"b0ed0930-4327-4772-9a1c-2a0ae55f529c\") " Dec 01 14:48:14 crc kubenswrapper[4810]: I1201 14:48:14.715324 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0ed0930-4327-4772-9a1c-2a0ae55f529c-bundle" (OuterVolumeSpecName: "bundle") pod "b0ed0930-4327-4772-9a1c-2a0ae55f529c" (UID: "b0ed0930-4327-4772-9a1c-2a0ae55f529c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:48:14 crc kubenswrapper[4810]: I1201 14:48:14.720958 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0ed0930-4327-4772-9a1c-2a0ae55f529c-kube-api-access-rl9s6" (OuterVolumeSpecName: "kube-api-access-rl9s6") pod "b0ed0930-4327-4772-9a1c-2a0ae55f529c" (UID: "b0ed0930-4327-4772-9a1c-2a0ae55f529c"). InnerVolumeSpecName "kube-api-access-rl9s6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:48:14 crc kubenswrapper[4810]: I1201 14:48:14.815383 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rl9s6\" (UniqueName: \"kubernetes.io/projected/b0ed0930-4327-4772-9a1c-2a0ae55f529c-kube-api-access-rl9s6\") on node \"crc\" DevicePath \"\"" Dec 01 14:48:14 crc kubenswrapper[4810]: I1201 14:48:14.815437 4810 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b0ed0930-4327-4772-9a1c-2a0ae55f529c-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:48:14 crc kubenswrapper[4810]: I1201 14:48:14.847615 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0ed0930-4327-4772-9a1c-2a0ae55f529c-util" (OuterVolumeSpecName: "util") pod "b0ed0930-4327-4772-9a1c-2a0ae55f529c" (UID: "b0ed0930-4327-4772-9a1c-2a0ae55f529c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:48:14 crc kubenswrapper[4810]: I1201 14:48:14.916493 4810 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b0ed0930-4327-4772-9a1c-2a0ae55f529c-util\") on node \"crc\" DevicePath \"\"" Dec 01 14:48:15 crc kubenswrapper[4810]: I1201 14:48:15.409243 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" event={"ID":"b0ed0930-4327-4772-9a1c-2a0ae55f529c","Type":"ContainerDied","Data":"2da0dcf4cd5f38ed16a844569a54382d92d8925b8727100a27ed0b3b8eb80a33"} Dec 01 14:48:15 crc kubenswrapper[4810]: I1201 14:48:15.409849 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2da0dcf4cd5f38ed16a844569a54382d92d8925b8727100a27ed0b3b8eb80a33" Dec 01 14:48:15 crc kubenswrapper[4810]: I1201 14:48:15.409287 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz" Dec 01 14:48:15 crc kubenswrapper[4810]: I1201 14:48:15.413057 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sq6nr" event={"ID":"19288c1f-8d88-4e3a-85e3-efbaea495b77","Type":"ContainerStarted","Data":"a261523d03b8b689360552768260cc8889852cb22a75019bd77abe95ed255734"} Dec 01 14:48:15 crc kubenswrapper[4810]: I1201 14:48:15.702317 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sq6nr" podStartSLOduration=1.897834251 podStartE2EDuration="5.702285237s" podCreationTimestamp="2025-12-01 14:48:10 +0000 UTC" firstStartedPulling="2025-12-01 14:48:11.344245267 +0000 UTC m=+857.107754870" lastFinishedPulling="2025-12-01 14:48:15.148696253 +0000 UTC m=+860.912205856" observedRunningTime="2025-12-01 14:48:15.442877852 +0000 UTC m=+861.206387455" watchObservedRunningTime="2025-12-01 14:48:15.702285237 +0000 UTC m=+861.465794840" Dec 01 14:48:19 crc kubenswrapper[4810]: I1201 14:48:19.841115 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-6k5cn"] Dec 01 14:48:19 crc kubenswrapper[4810]: E1201 14:48:19.841954 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ed0930-4327-4772-9a1c-2a0ae55f529c" containerName="extract" Dec 01 14:48:19 crc kubenswrapper[4810]: I1201 14:48:19.841975 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ed0930-4327-4772-9a1c-2a0ae55f529c" containerName="extract" Dec 01 14:48:19 crc kubenswrapper[4810]: E1201 14:48:19.841984 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ed0930-4327-4772-9a1c-2a0ae55f529c" containerName="util" Dec 01 14:48:19 crc kubenswrapper[4810]: I1201 14:48:19.841990 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ed0930-4327-4772-9a1c-2a0ae55f529c" containerName="util" Dec 01 14:48:19 crc kubenswrapper[4810]: E1201 14:48:19.842007 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ed0930-4327-4772-9a1c-2a0ae55f529c" containerName="pull" Dec 01 14:48:19 crc kubenswrapper[4810]: I1201 14:48:19.842014 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ed0930-4327-4772-9a1c-2a0ae55f529c" containerName="pull" Dec 01 14:48:19 crc kubenswrapper[4810]: I1201 14:48:19.842150 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0ed0930-4327-4772-9a1c-2a0ae55f529c" containerName="extract" Dec 01 14:48:19 crc kubenswrapper[4810]: I1201 14:48:19.842720 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-6k5cn" Dec 01 14:48:19 crc kubenswrapper[4810]: I1201 14:48:19.845530 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-htmm4" Dec 01 14:48:19 crc kubenswrapper[4810]: I1201 14:48:19.846133 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 01 14:48:19 crc kubenswrapper[4810]: I1201 14:48:19.847492 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 01 14:48:19 crc kubenswrapper[4810]: I1201 14:48:19.854262 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-6k5cn"] Dec 01 14:48:19 crc kubenswrapper[4810]: I1201 14:48:19.913813 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxc9d\" (UniqueName: \"kubernetes.io/projected/33232e23-1ded-4abb-9df2-247452f3f6bf-kube-api-access-qxc9d\") pod \"nmstate-operator-5b5b58f5c8-6k5cn\" (UID: \"33232e23-1ded-4abb-9df2-247452f3f6bf\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-6k5cn" Dec 01 14:48:20 crc kubenswrapper[4810]: I1201 14:48:20.015774 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxc9d\" (UniqueName: \"kubernetes.io/projected/33232e23-1ded-4abb-9df2-247452f3f6bf-kube-api-access-qxc9d\") pod \"nmstate-operator-5b5b58f5c8-6k5cn\" (UID: \"33232e23-1ded-4abb-9df2-247452f3f6bf\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-6k5cn" Dec 01 14:48:20 crc kubenswrapper[4810]: I1201 14:48:20.037916 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxc9d\" (UniqueName: \"kubernetes.io/projected/33232e23-1ded-4abb-9df2-247452f3f6bf-kube-api-access-qxc9d\") pod \"nmstate-operator-5b5b58f5c8-6k5cn\" (UID: \"33232e23-1ded-4abb-9df2-247452f3f6bf\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-6k5cn" Dec 01 14:48:20 crc kubenswrapper[4810]: I1201 14:48:20.157992 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-6k5cn" Dec 01 14:48:20 crc kubenswrapper[4810]: I1201 14:48:20.499569 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-6k5cn"] Dec 01 14:48:20 crc kubenswrapper[4810]: I1201 14:48:20.699118 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sq6nr" Dec 01 14:48:20 crc kubenswrapper[4810]: I1201 14:48:20.699328 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sq6nr" Dec 01 14:48:20 crc kubenswrapper[4810]: I1201 14:48:20.744020 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sq6nr" Dec 01 14:48:21 crc kubenswrapper[4810]: I1201 14:48:21.452978 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-6k5cn" event={"ID":"33232e23-1ded-4abb-9df2-247452f3f6bf","Type":"ContainerStarted","Data":"94359295e166685570d039c5240a71fc55f2018f798a546e32352a3384c6291d"} Dec 01 14:48:21 crc kubenswrapper[4810]: I1201 14:48:21.496620 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sq6nr" Dec 01 14:48:23 crc kubenswrapper[4810]: I1201 14:48:23.359316 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sq6nr"] Dec 01 14:48:24 crc kubenswrapper[4810]: I1201 14:48:24.479500 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sq6nr" podUID="19288c1f-8d88-4e3a-85e3-efbaea495b77" containerName="registry-server" containerID="cri-o://a261523d03b8b689360552768260cc8889852cb22a75019bd77abe95ed255734" gracePeriod=2 Dec 01 14:48:24 crc kubenswrapper[4810]: I1201 14:48:24.480012 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-6k5cn" event={"ID":"33232e23-1ded-4abb-9df2-247452f3f6bf","Type":"ContainerStarted","Data":"63511349f5cc6f86113c7237ae4e525cb285bd7ccb66704031b67660305f4a5b"} Dec 01 14:48:24 crc kubenswrapper[4810]: I1201 14:48:24.505676 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-6k5cn" podStartSLOduration=2.757092043 podStartE2EDuration="5.5056509s" podCreationTimestamp="2025-12-01 14:48:19 +0000 UTC" firstStartedPulling="2025-12-01 14:48:20.509774763 +0000 UTC m=+866.273284366" lastFinishedPulling="2025-12-01 14:48:23.25833362 +0000 UTC m=+869.021843223" observedRunningTime="2025-12-01 14:48:24.504407956 +0000 UTC m=+870.267917559" watchObservedRunningTime="2025-12-01 14:48:24.5056509 +0000 UTC m=+870.269160503" Dec 01 14:48:25 crc kubenswrapper[4810]: I1201 14:48:25.487266 4810 generic.go:334] "Generic (PLEG): container finished" podID="19288c1f-8d88-4e3a-85e3-efbaea495b77" containerID="a261523d03b8b689360552768260cc8889852cb22a75019bd77abe95ed255734" exitCode=0 Dec 01 14:48:25 crc kubenswrapper[4810]: I1201 14:48:25.487327 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sq6nr" event={"ID":"19288c1f-8d88-4e3a-85e3-efbaea495b77","Type":"ContainerDied","Data":"a261523d03b8b689360552768260cc8889852cb22a75019bd77abe95ed255734"} Dec 01 14:48:26 crc kubenswrapper[4810]: I1201 14:48:26.018982 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sq6nr" Dec 01 14:48:26 crc kubenswrapper[4810]: I1201 14:48:26.115033 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19288c1f-8d88-4e3a-85e3-efbaea495b77-catalog-content\") pod \"19288c1f-8d88-4e3a-85e3-efbaea495b77\" (UID: \"19288c1f-8d88-4e3a-85e3-efbaea495b77\") " Dec 01 14:48:26 crc kubenswrapper[4810]: I1201 14:48:26.115079 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19288c1f-8d88-4e3a-85e3-efbaea495b77-utilities\") pod \"19288c1f-8d88-4e3a-85e3-efbaea495b77\" (UID: \"19288c1f-8d88-4e3a-85e3-efbaea495b77\") " Dec 01 14:48:26 crc kubenswrapper[4810]: I1201 14:48:26.115191 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7svbb\" (UniqueName: \"kubernetes.io/projected/19288c1f-8d88-4e3a-85e3-efbaea495b77-kube-api-access-7svbb\") pod \"19288c1f-8d88-4e3a-85e3-efbaea495b77\" (UID: \"19288c1f-8d88-4e3a-85e3-efbaea495b77\") " Dec 01 14:48:26 crc kubenswrapper[4810]: I1201 14:48:26.116438 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19288c1f-8d88-4e3a-85e3-efbaea495b77-utilities" (OuterVolumeSpecName: "utilities") pod "19288c1f-8d88-4e3a-85e3-efbaea495b77" (UID: "19288c1f-8d88-4e3a-85e3-efbaea495b77"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:48:26 crc kubenswrapper[4810]: I1201 14:48:26.122643 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19288c1f-8d88-4e3a-85e3-efbaea495b77-kube-api-access-7svbb" (OuterVolumeSpecName: "kube-api-access-7svbb") pod "19288c1f-8d88-4e3a-85e3-efbaea495b77" (UID: "19288c1f-8d88-4e3a-85e3-efbaea495b77"). InnerVolumeSpecName "kube-api-access-7svbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:48:26 crc kubenswrapper[4810]: I1201 14:48:26.216783 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19288c1f-8d88-4e3a-85e3-efbaea495b77-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 14:48:26 crc kubenswrapper[4810]: I1201 14:48:26.217098 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7svbb\" (UniqueName: \"kubernetes.io/projected/19288c1f-8d88-4e3a-85e3-efbaea495b77-kube-api-access-7svbb\") on node \"crc\" DevicePath \"\"" Dec 01 14:48:26 crc kubenswrapper[4810]: I1201 14:48:26.236667 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19288c1f-8d88-4e3a-85e3-efbaea495b77-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19288c1f-8d88-4e3a-85e3-efbaea495b77" (UID: "19288c1f-8d88-4e3a-85e3-efbaea495b77"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:48:26 crc kubenswrapper[4810]: I1201 14:48:26.318874 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19288c1f-8d88-4e3a-85e3-efbaea495b77-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 14:48:26 crc kubenswrapper[4810]: I1201 14:48:26.497521 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sq6nr" Dec 01 14:48:26 crc kubenswrapper[4810]: I1201 14:48:26.499178 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sq6nr" event={"ID":"19288c1f-8d88-4e3a-85e3-efbaea495b77","Type":"ContainerDied","Data":"8e9b8f5e0cb6ac32aa2558854ff2888dab0cbd4196ac34baff45b2f0e454b67f"} Dec 01 14:48:26 crc kubenswrapper[4810]: I1201 14:48:26.499226 4810 scope.go:117] "RemoveContainer" containerID="a261523d03b8b689360552768260cc8889852cb22a75019bd77abe95ed255734" Dec 01 14:48:26 crc kubenswrapper[4810]: I1201 14:48:26.545125 4810 scope.go:117] "RemoveContainer" containerID="873ead448951c49d8683e5125cc1fbbe9a5347283041d08824f5ff296ad4617e" Dec 01 14:48:26 crc kubenswrapper[4810]: I1201 14:48:26.553508 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sq6nr"] Dec 01 14:48:26 crc kubenswrapper[4810]: I1201 14:48:26.558906 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sq6nr"] Dec 01 14:48:26 crc kubenswrapper[4810]: I1201 14:48:26.562618 4810 scope.go:117] "RemoveContainer" containerID="7a4bc98ea751885f6d257cfc4ed4579cdd450ea55b5ac511cb6fac373903806f" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.500111 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19288c1f-8d88-4e3a-85e3-efbaea495b77" path="/var/lib/kubelet/pods/19288c1f-8d88-4e3a-85e3-efbaea495b77/volumes" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.745416 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-hsl7k"] Dec 01 14:48:28 crc kubenswrapper[4810]: E1201 14:48:28.746591 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19288c1f-8d88-4e3a-85e3-efbaea495b77" containerName="extract-utilities" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.746672 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="19288c1f-8d88-4e3a-85e3-efbaea495b77" containerName="extract-utilities" Dec 01 14:48:28 crc kubenswrapper[4810]: E1201 14:48:28.746731 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19288c1f-8d88-4e3a-85e3-efbaea495b77" containerName="extract-content" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.746781 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="19288c1f-8d88-4e3a-85e3-efbaea495b77" containerName="extract-content" Dec 01 14:48:28 crc kubenswrapper[4810]: E1201 14:48:28.746836 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19288c1f-8d88-4e3a-85e3-efbaea495b77" containerName="registry-server" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.746882 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="19288c1f-8d88-4e3a-85e3-efbaea495b77" containerName="registry-server" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.747043 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="19288c1f-8d88-4e3a-85e3-efbaea495b77" containerName="registry-server" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.747787 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hsl7k" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.751881 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-h86m4" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.765189 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hzsts"] Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.766346 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hzsts" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.768860 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-hsl7k"] Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.769570 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.789765 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hzsts"] Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.798150 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-n8r5h"] Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.799149 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-n8r5h" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.851084 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52976\" (UniqueName: \"kubernetes.io/projected/3e4852d7-004b-462b-9b99-c9a13097292d-kube-api-access-52976\") pod \"nmstate-handler-n8r5h\" (UID: \"3e4852d7-004b-462b-9b99-c9a13097292d\") " pod="openshift-nmstate/nmstate-handler-n8r5h" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.851153 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/3e4852d7-004b-462b-9b99-c9a13097292d-ovs-socket\") pod \"nmstate-handler-n8r5h\" (UID: \"3e4852d7-004b-462b-9b99-c9a13097292d\") " pod="openshift-nmstate/nmstate-handler-n8r5h" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.851187 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/3e4852d7-004b-462b-9b99-c9a13097292d-dbus-socket\") pod \"nmstate-handler-n8r5h\" (UID: \"3e4852d7-004b-462b-9b99-c9a13097292d\") " pod="openshift-nmstate/nmstate-handler-n8r5h" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.851217 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdj6n\" (UniqueName: \"kubernetes.io/projected/3cda0965-f2f6-4a96-9e5a-91938cf20b0f-kube-api-access-xdj6n\") pod \"nmstate-webhook-5f6d4c5ccb-hzsts\" (UID: \"3cda0965-f2f6-4a96-9e5a-91938cf20b0f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hzsts" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.851266 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/3e4852d7-004b-462b-9b99-c9a13097292d-nmstate-lock\") pod \"nmstate-handler-n8r5h\" (UID: \"3e4852d7-004b-462b-9b99-c9a13097292d\") " pod="openshift-nmstate/nmstate-handler-n8r5h" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.851296 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3cda0965-f2f6-4a96-9e5a-91938cf20b0f-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-hzsts\" (UID: \"3cda0965-f2f6-4a96-9e5a-91938cf20b0f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hzsts" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.851325 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxv7g\" (UniqueName: \"kubernetes.io/projected/c46e2836-e450-4375-bc15-3295c3772671-kube-api-access-wxv7g\") pod \"nmstate-metrics-7f946cbc9-hsl7k\" (UID: \"c46e2836-e450-4375-bc15-3295c3772671\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hsl7k" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.935136 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qxwtv"] Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.935988 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qxwtv" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.938849 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-2vfst" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.939162 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.941989 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.951987 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52976\" (UniqueName: \"kubernetes.io/projected/3e4852d7-004b-462b-9b99-c9a13097292d-kube-api-access-52976\") pod \"nmstate-handler-n8r5h\" (UID: \"3e4852d7-004b-462b-9b99-c9a13097292d\") " pod="openshift-nmstate/nmstate-handler-n8r5h" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.952032 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/3e4852d7-004b-462b-9b99-c9a13097292d-ovs-socket\") pod \"nmstate-handler-n8r5h\" (UID: \"3e4852d7-004b-462b-9b99-c9a13097292d\") " pod="openshift-nmstate/nmstate-handler-n8r5h" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.952074 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/3e4852d7-004b-462b-9b99-c9a13097292d-dbus-socket\") pod \"nmstate-handler-n8r5h\" (UID: \"3e4852d7-004b-462b-9b99-c9a13097292d\") " pod="openshift-nmstate/nmstate-handler-n8r5h" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.952104 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdj6n\" (UniqueName: \"kubernetes.io/projected/3cda0965-f2f6-4a96-9e5a-91938cf20b0f-kube-api-access-xdj6n\") pod \"nmstate-webhook-5f6d4c5ccb-hzsts\" (UID: \"3cda0965-f2f6-4a96-9e5a-91938cf20b0f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hzsts" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.952144 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/3e4852d7-004b-462b-9b99-c9a13097292d-nmstate-lock\") pod \"nmstate-handler-n8r5h\" (UID: \"3e4852d7-004b-462b-9b99-c9a13097292d\") " pod="openshift-nmstate/nmstate-handler-n8r5h" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.952166 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3cda0965-f2f6-4a96-9e5a-91938cf20b0f-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-hzsts\" (UID: \"3cda0965-f2f6-4a96-9e5a-91938cf20b0f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hzsts" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.952186 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxv7g\" (UniqueName: \"kubernetes.io/projected/c46e2836-e450-4375-bc15-3295c3772671-kube-api-access-wxv7g\") pod \"nmstate-metrics-7f946cbc9-hsl7k\" (UID: \"c46e2836-e450-4375-bc15-3295c3772671\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hsl7k" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.952728 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/3e4852d7-004b-462b-9b99-c9a13097292d-ovs-socket\") pod \"nmstate-handler-n8r5h\" (UID: \"3e4852d7-004b-462b-9b99-c9a13097292d\") " pod="openshift-nmstate/nmstate-handler-n8r5h" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.953101 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/3e4852d7-004b-462b-9b99-c9a13097292d-dbus-socket\") pod \"nmstate-handler-n8r5h\" (UID: \"3e4852d7-004b-462b-9b99-c9a13097292d\") " pod="openshift-nmstate/nmstate-handler-n8r5h" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.953275 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/3e4852d7-004b-462b-9b99-c9a13097292d-nmstate-lock\") pod \"nmstate-handler-n8r5h\" (UID: \"3e4852d7-004b-462b-9b99-c9a13097292d\") " pod="openshift-nmstate/nmstate-handler-n8r5h" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.954096 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qxwtv"] Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.970406 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3cda0965-f2f6-4a96-9e5a-91938cf20b0f-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-hzsts\" (UID: \"3cda0965-f2f6-4a96-9e5a-91938cf20b0f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hzsts" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.972896 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdj6n\" (UniqueName: \"kubernetes.io/projected/3cda0965-f2f6-4a96-9e5a-91938cf20b0f-kube-api-access-xdj6n\") pod \"nmstate-webhook-5f6d4c5ccb-hzsts\" (UID: \"3cda0965-f2f6-4a96-9e5a-91938cf20b0f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hzsts" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.977866 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxv7g\" (UniqueName: \"kubernetes.io/projected/c46e2836-e450-4375-bc15-3295c3772671-kube-api-access-wxv7g\") pod \"nmstate-metrics-7f946cbc9-hsl7k\" (UID: \"c46e2836-e450-4375-bc15-3295c3772671\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hsl7k" Dec 01 14:48:28 crc kubenswrapper[4810]: I1201 14:48:28.978687 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52976\" (UniqueName: \"kubernetes.io/projected/3e4852d7-004b-462b-9b99-c9a13097292d-kube-api-access-52976\") pod \"nmstate-handler-n8r5h\" (UID: \"3e4852d7-004b-462b-9b99-c9a13097292d\") " pod="openshift-nmstate/nmstate-handler-n8r5h" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.055451 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e16a3390-afd1-4812-b571-883e7c1515f7-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-qxwtv\" (UID: \"e16a3390-afd1-4812-b571-883e7c1515f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qxwtv" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.055563 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e16a3390-afd1-4812-b571-883e7c1515f7-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-qxwtv\" (UID: \"e16a3390-afd1-4812-b571-883e7c1515f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qxwtv" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.055609 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5md8x\" (UniqueName: \"kubernetes.io/projected/e16a3390-afd1-4812-b571-883e7c1515f7-kube-api-access-5md8x\") pod \"nmstate-console-plugin-7fbb5f6569-qxwtv\" (UID: \"e16a3390-afd1-4812-b571-883e7c1515f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qxwtv" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.066199 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hsl7k" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.090968 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hzsts" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.138968 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-n8r5h" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.165561 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e16a3390-afd1-4812-b571-883e7c1515f7-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-qxwtv\" (UID: \"e16a3390-afd1-4812-b571-883e7c1515f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qxwtv" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.165832 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5md8x\" (UniqueName: \"kubernetes.io/projected/e16a3390-afd1-4812-b571-883e7c1515f7-kube-api-access-5md8x\") pod \"nmstate-console-plugin-7fbb5f6569-qxwtv\" (UID: \"e16a3390-afd1-4812-b571-883e7c1515f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qxwtv" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.166048 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e16a3390-afd1-4812-b571-883e7c1515f7-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-qxwtv\" (UID: \"e16a3390-afd1-4812-b571-883e7c1515f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qxwtv" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.167220 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e16a3390-afd1-4812-b571-883e7c1515f7-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-qxwtv\" (UID: \"e16a3390-afd1-4812-b571-883e7c1515f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qxwtv" Dec 01 14:48:29 crc kubenswrapper[4810]: E1201 14:48:29.165841 4810 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 01 14:48:29 crc kubenswrapper[4810]: E1201 14:48:29.172535 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e16a3390-afd1-4812-b571-883e7c1515f7-plugin-serving-cert podName:e16a3390-afd1-4812-b571-883e7c1515f7 nodeName:}" failed. No retries permitted until 2025-12-01 14:48:29.672454778 +0000 UTC m=+875.435964371 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/e16a3390-afd1-4812-b571-883e7c1515f7-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-qxwtv" (UID: "e16a3390-afd1-4812-b571-883e7c1515f7") : secret "plugin-serving-cert" not found Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.182138 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-57d8d6f6c4-6qd59"] Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.183349 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.197286 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5md8x\" (UniqueName: \"kubernetes.io/projected/e16a3390-afd1-4812-b571-883e7c1515f7-kube-api-access-5md8x\") pod \"nmstate-console-plugin-7fbb5f6569-qxwtv\" (UID: \"e16a3390-afd1-4812-b571-883e7c1515f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qxwtv" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.202716 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-57d8d6f6c4-6qd59"] Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.376523 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2c362309-97b6-4241-9812-bf58a2c13551-console-oauth-config\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.376584 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kl9r\" (UniqueName: \"kubernetes.io/projected/2c362309-97b6-4241-9812-bf58a2c13551-kube-api-access-8kl9r\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.376628 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2c362309-97b6-4241-9812-bf58a2c13551-oauth-serving-cert\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.376649 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c362309-97b6-4241-9812-bf58a2c13551-console-serving-cert\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.376691 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c362309-97b6-4241-9812-bf58a2c13551-trusted-ca-bundle\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.376719 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2c362309-97b6-4241-9812-bf58a2c13551-console-config\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.376787 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2c362309-97b6-4241-9812-bf58a2c13551-service-ca\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.478219 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c362309-97b6-4241-9812-bf58a2c13551-trusted-ca-bundle\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.478269 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2c362309-97b6-4241-9812-bf58a2c13551-console-config\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.478318 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2c362309-97b6-4241-9812-bf58a2c13551-service-ca\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.478349 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2c362309-97b6-4241-9812-bf58a2c13551-console-oauth-config\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.478374 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kl9r\" (UniqueName: \"kubernetes.io/projected/2c362309-97b6-4241-9812-bf58a2c13551-kube-api-access-8kl9r\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.478407 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2c362309-97b6-4241-9812-bf58a2c13551-oauth-serving-cert\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.478428 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c362309-97b6-4241-9812-bf58a2c13551-console-serving-cert\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.479495 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2c362309-97b6-4241-9812-bf58a2c13551-oauth-serving-cert\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.479682 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2c362309-97b6-4241-9812-bf58a2c13551-console-config\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.479888 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2c362309-97b6-4241-9812-bf58a2c13551-service-ca\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.480988 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2c362309-97b6-4241-9812-bf58a2c13551-trusted-ca-bundle\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.483598 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2c362309-97b6-4241-9812-bf58a2c13551-console-oauth-config\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.486287 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2c362309-97b6-4241-9812-bf58a2c13551-console-serving-cert\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.496246 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kl9r\" (UniqueName: \"kubernetes.io/projected/2c362309-97b6-4241-9812-bf58a2c13551-kube-api-access-8kl9r\") pod \"console-57d8d6f6c4-6qd59\" (UID: \"2c362309-97b6-4241-9812-bf58a2c13551\") " pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.510099 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.516392 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-n8r5h" event={"ID":"3e4852d7-004b-462b-9b99-c9a13097292d","Type":"ContainerStarted","Data":"f9781b1bfdc666c8e41f15a5b3d39d920b06da7b39a0f4062f3edf94728646d7"} Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.635660 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hzsts"] Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.655922 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-hsl7k"] Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.681318 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e16a3390-afd1-4812-b571-883e7c1515f7-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-qxwtv\" (UID: \"e16a3390-afd1-4812-b571-883e7c1515f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qxwtv" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.695147 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e16a3390-afd1-4812-b571-883e7c1515f7-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-qxwtv\" (UID: \"e16a3390-afd1-4812-b571-883e7c1515f7\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qxwtv" Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.753022 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-57d8d6f6c4-6qd59"] Dec 01 14:48:29 crc kubenswrapper[4810]: W1201 14:48:29.762456 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c362309_97b6_4241_9812_bf58a2c13551.slice/crio-9f7614b503b336216815b310db2cf77b50e598d0db20f76e9565f1063b065831 WatchSource:0}: Error finding container 9f7614b503b336216815b310db2cf77b50e598d0db20f76e9565f1063b065831: Status 404 returned error can't find the container with id 9f7614b503b336216815b310db2cf77b50e598d0db20f76e9565f1063b065831 Dec 01 14:48:29 crc kubenswrapper[4810]: I1201 14:48:29.852035 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qxwtv" Dec 01 14:48:30 crc kubenswrapper[4810]: I1201 14:48:30.128558 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qxwtv"] Dec 01 14:48:30 crc kubenswrapper[4810]: I1201 14:48:30.526197 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qxwtv" event={"ID":"e16a3390-afd1-4812-b571-883e7c1515f7","Type":"ContainerStarted","Data":"06d6ae5903f83d268f8c715dceb1089c7b7f5d11ca2bb09aea0488db88ffd5ef"} Dec 01 14:48:30 crc kubenswrapper[4810]: I1201 14:48:30.530411 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-57d8d6f6c4-6qd59" event={"ID":"2c362309-97b6-4241-9812-bf58a2c13551","Type":"ContainerStarted","Data":"54653655b4ddea3eef803f6c684573ca495c3f386694a700140c59c7332b7120"} Dec 01 14:48:30 crc kubenswrapper[4810]: I1201 14:48:30.530449 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-57d8d6f6c4-6qd59" event={"ID":"2c362309-97b6-4241-9812-bf58a2c13551","Type":"ContainerStarted","Data":"9f7614b503b336216815b310db2cf77b50e598d0db20f76e9565f1063b065831"} Dec 01 14:48:30 crc kubenswrapper[4810]: I1201 14:48:30.532359 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hsl7k" event={"ID":"c46e2836-e450-4375-bc15-3295c3772671","Type":"ContainerStarted","Data":"ee5bff251800789b7f6442502face8586cbf0aa2aabfc446a294f4d728fbe4a3"} Dec 01 14:48:30 crc kubenswrapper[4810]: I1201 14:48:30.533731 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hzsts" event={"ID":"3cda0965-f2f6-4a96-9e5a-91938cf20b0f","Type":"ContainerStarted","Data":"ebaa34f41f9ea98a86a95694667d6453eb2fe8e2181698d3fb50c4753a02f154"} Dec 01 14:48:30 crc kubenswrapper[4810]: I1201 14:48:30.564298 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-57d8d6f6c4-6qd59" podStartSLOduration=1.564248791 podStartE2EDuration="1.564248791s" podCreationTimestamp="2025-12-01 14:48:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:48:30.557483746 +0000 UTC m=+876.320993339" watchObservedRunningTime="2025-12-01 14:48:30.564248791 +0000 UTC m=+876.327758394" Dec 01 14:48:32 crc kubenswrapper[4810]: I1201 14:48:32.552637 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-n8r5h" event={"ID":"3e4852d7-004b-462b-9b99-c9a13097292d","Type":"ContainerStarted","Data":"493d0e7219d4dc5ab8c8f822c3c27cc2af0cc36f905c2c279995c6075136d6c9"} Dec 01 14:48:32 crc kubenswrapper[4810]: I1201 14:48:32.553647 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-n8r5h" Dec 01 14:48:32 crc kubenswrapper[4810]: I1201 14:48:32.558925 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hsl7k" event={"ID":"c46e2836-e450-4375-bc15-3295c3772671","Type":"ContainerStarted","Data":"6075e584acbe8e1a3ba8f3820d9cfb1ad35e47df9a37af5dd93fb9a4237045e1"} Dec 01 14:48:32 crc kubenswrapper[4810]: I1201 14:48:32.561040 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hzsts" event={"ID":"3cda0965-f2f6-4a96-9e5a-91938cf20b0f","Type":"ContainerStarted","Data":"73e082c3de36e63e34e338621e76cd97acb266d3818f8bdf0170f25f2a1ed62a"} Dec 01 14:48:32 crc kubenswrapper[4810]: I1201 14:48:32.561391 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hzsts" Dec 01 14:48:32 crc kubenswrapper[4810]: I1201 14:48:32.577282 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-n8r5h" podStartSLOduration=1.7830017439999999 podStartE2EDuration="4.577258293s" podCreationTimestamp="2025-12-01 14:48:28 +0000 UTC" firstStartedPulling="2025-12-01 14:48:29.20717041 +0000 UTC m=+874.970680013" lastFinishedPulling="2025-12-01 14:48:32.001426949 +0000 UTC m=+877.764936562" observedRunningTime="2025-12-01 14:48:32.569945902 +0000 UTC m=+878.333455525" watchObservedRunningTime="2025-12-01 14:48:32.577258293 +0000 UTC m=+878.340767896" Dec 01 14:48:32 crc kubenswrapper[4810]: I1201 14:48:32.592057 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hzsts" podStartSLOduration=2.246845075 podStartE2EDuration="4.592031708s" podCreationTimestamp="2025-12-01 14:48:28 +0000 UTC" firstStartedPulling="2025-12-01 14:48:29.656437122 +0000 UTC m=+875.419946725" lastFinishedPulling="2025-12-01 14:48:32.001623755 +0000 UTC m=+877.765133358" observedRunningTime="2025-12-01 14:48:32.589138249 +0000 UTC m=+878.352647872" watchObservedRunningTime="2025-12-01 14:48:32.592031708 +0000 UTC m=+878.355541311" Dec 01 14:48:33 crc kubenswrapper[4810]: I1201 14:48:33.569377 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qxwtv" event={"ID":"e16a3390-afd1-4812-b571-883e7c1515f7","Type":"ContainerStarted","Data":"7fbe82edd800227b8706aa4166b9bcd7ea6e10857c9fd1c4b6041df548cde056"} Dec 01 14:48:33 crc kubenswrapper[4810]: I1201 14:48:33.589711 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-qxwtv" podStartSLOduration=2.705869615 podStartE2EDuration="5.589688901s" podCreationTimestamp="2025-12-01 14:48:28 +0000 UTC" firstStartedPulling="2025-12-01 14:48:30.143704017 +0000 UTC m=+875.907213630" lastFinishedPulling="2025-12-01 14:48:33.027523303 +0000 UTC m=+878.791032916" observedRunningTime="2025-12-01 14:48:33.586563335 +0000 UTC m=+879.350072938" watchObservedRunningTime="2025-12-01 14:48:33.589688901 +0000 UTC m=+879.353198504" Dec 01 14:48:35 crc kubenswrapper[4810]: I1201 14:48:35.591763 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hsl7k" event={"ID":"c46e2836-e450-4375-bc15-3295c3772671","Type":"ContainerStarted","Data":"2abf8b0d5a25e39fa97dc211217bae4c88a6eea535fbec78cbcf1eb7d4b6af39"} Dec 01 14:48:35 crc kubenswrapper[4810]: I1201 14:48:35.624005 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hsl7k" podStartSLOduration=2.31154849 podStartE2EDuration="7.623974246s" podCreationTimestamp="2025-12-01 14:48:28 +0000 UTC" firstStartedPulling="2025-12-01 14:48:29.66399768 +0000 UTC m=+875.427507283" lastFinishedPulling="2025-12-01 14:48:34.976423446 +0000 UTC m=+880.739933039" observedRunningTime="2025-12-01 14:48:35.621773566 +0000 UTC m=+881.385283239" watchObservedRunningTime="2025-12-01 14:48:35.623974246 +0000 UTC m=+881.387483879" Dec 01 14:48:39 crc kubenswrapper[4810]: I1201 14:48:39.173885 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-n8r5h" Dec 01 14:48:39 crc kubenswrapper[4810]: I1201 14:48:39.511294 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:39 crc kubenswrapper[4810]: I1201 14:48:39.511794 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:39 crc kubenswrapper[4810]: I1201 14:48:39.515837 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:39 crc kubenswrapper[4810]: I1201 14:48:39.618522 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-57d8d6f6c4-6qd59" Dec 01 14:48:39 crc kubenswrapper[4810]: I1201 14:48:39.711250 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-8tcbx"] Dec 01 14:48:49 crc kubenswrapper[4810]: I1201 14:48:49.097121 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-hzsts" Dec 01 14:49:01 crc kubenswrapper[4810]: I1201 14:49:01.197981 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv"] Dec 01 14:49:01 crc kubenswrapper[4810]: I1201 14:49:01.200332 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" Dec 01 14:49:01 crc kubenswrapper[4810]: I1201 14:49:01.202682 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 14:49:01 crc kubenswrapper[4810]: I1201 14:49:01.207721 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv"] Dec 01 14:49:01 crc kubenswrapper[4810]: I1201 14:49:01.293810 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/673fd73e-0701-4cf7-a7dd-273db851c2ba-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv\" (UID: \"673fd73e-0701-4cf7-a7dd-273db851c2ba\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" Dec 01 14:49:01 crc kubenswrapper[4810]: I1201 14:49:01.293870 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4d5m\" (UniqueName: \"kubernetes.io/projected/673fd73e-0701-4cf7-a7dd-273db851c2ba-kube-api-access-p4d5m\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv\" (UID: \"673fd73e-0701-4cf7-a7dd-273db851c2ba\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" Dec 01 14:49:01 crc kubenswrapper[4810]: I1201 14:49:01.293979 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/673fd73e-0701-4cf7-a7dd-273db851c2ba-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv\" (UID: \"673fd73e-0701-4cf7-a7dd-273db851c2ba\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" Dec 01 14:49:01 crc kubenswrapper[4810]: I1201 14:49:01.395194 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/673fd73e-0701-4cf7-a7dd-273db851c2ba-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv\" (UID: \"673fd73e-0701-4cf7-a7dd-273db851c2ba\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" Dec 01 14:49:01 crc kubenswrapper[4810]: I1201 14:49:01.395245 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4d5m\" (UniqueName: \"kubernetes.io/projected/673fd73e-0701-4cf7-a7dd-273db851c2ba-kube-api-access-p4d5m\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv\" (UID: \"673fd73e-0701-4cf7-a7dd-273db851c2ba\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" Dec 01 14:49:01 crc kubenswrapper[4810]: I1201 14:49:01.395301 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/673fd73e-0701-4cf7-a7dd-273db851c2ba-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv\" (UID: \"673fd73e-0701-4cf7-a7dd-273db851c2ba\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" Dec 01 14:49:01 crc kubenswrapper[4810]: I1201 14:49:01.395862 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/673fd73e-0701-4cf7-a7dd-273db851c2ba-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv\" (UID: \"673fd73e-0701-4cf7-a7dd-273db851c2ba\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" Dec 01 14:49:01 crc kubenswrapper[4810]: I1201 14:49:01.395892 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/673fd73e-0701-4cf7-a7dd-273db851c2ba-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv\" (UID: \"673fd73e-0701-4cf7-a7dd-273db851c2ba\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" Dec 01 14:49:01 crc kubenswrapper[4810]: I1201 14:49:01.413557 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4d5m\" (UniqueName: \"kubernetes.io/projected/673fd73e-0701-4cf7-a7dd-273db851c2ba-kube-api-access-p4d5m\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv\" (UID: \"673fd73e-0701-4cf7-a7dd-273db851c2ba\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" Dec 01 14:49:01 crc kubenswrapper[4810]: I1201 14:49:01.517969 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" Dec 01 14:49:01 crc kubenswrapper[4810]: I1201 14:49:01.694407 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv"] Dec 01 14:49:01 crc kubenswrapper[4810]: I1201 14:49:01.801379 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" event={"ID":"673fd73e-0701-4cf7-a7dd-273db851c2ba","Type":"ContainerStarted","Data":"d7c0be0bf38df9656a17178a8e793f197ac9133797ae95e7f74737f507155b58"} Dec 01 14:49:02 crc kubenswrapper[4810]: I1201 14:49:02.808058 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" event={"ID":"673fd73e-0701-4cf7-a7dd-273db851c2ba","Type":"ContainerStarted","Data":"9cdfe9c7f608ddb760e5c290eaf396ce001af69d392173f2f6e9a5e0f8fc6450"} Dec 01 14:49:03 crc kubenswrapper[4810]: I1201 14:49:03.162085 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rwmnj"] Dec 01 14:49:03 crc kubenswrapper[4810]: I1201 14:49:03.163377 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rwmnj" Dec 01 14:49:03 crc kubenswrapper[4810]: I1201 14:49:03.174259 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rwmnj"] Dec 01 14:49:03 crc kubenswrapper[4810]: I1201 14:49:03.215603 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6e0ff27-df96-47ca-897a-53b8ad395f24-catalog-content\") pod \"community-operators-rwmnj\" (UID: \"a6e0ff27-df96-47ca-897a-53b8ad395f24\") " pod="openshift-marketplace/community-operators-rwmnj" Dec 01 14:49:03 crc kubenswrapper[4810]: I1201 14:49:03.215669 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6e0ff27-df96-47ca-897a-53b8ad395f24-utilities\") pod \"community-operators-rwmnj\" (UID: \"a6e0ff27-df96-47ca-897a-53b8ad395f24\") " pod="openshift-marketplace/community-operators-rwmnj" Dec 01 14:49:03 crc kubenswrapper[4810]: I1201 14:49:03.215709 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt9hp\" (UniqueName: \"kubernetes.io/projected/a6e0ff27-df96-47ca-897a-53b8ad395f24-kube-api-access-wt9hp\") pod \"community-operators-rwmnj\" (UID: \"a6e0ff27-df96-47ca-897a-53b8ad395f24\") " pod="openshift-marketplace/community-operators-rwmnj" Dec 01 14:49:03 crc kubenswrapper[4810]: I1201 14:49:03.316430 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6e0ff27-df96-47ca-897a-53b8ad395f24-catalog-content\") pod \"community-operators-rwmnj\" (UID: \"a6e0ff27-df96-47ca-897a-53b8ad395f24\") " pod="openshift-marketplace/community-operators-rwmnj" Dec 01 14:49:03 crc kubenswrapper[4810]: I1201 14:49:03.316510 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6e0ff27-df96-47ca-897a-53b8ad395f24-utilities\") pod \"community-operators-rwmnj\" (UID: \"a6e0ff27-df96-47ca-897a-53b8ad395f24\") " pod="openshift-marketplace/community-operators-rwmnj" Dec 01 14:49:03 crc kubenswrapper[4810]: I1201 14:49:03.316568 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt9hp\" (UniqueName: \"kubernetes.io/projected/a6e0ff27-df96-47ca-897a-53b8ad395f24-kube-api-access-wt9hp\") pod \"community-operators-rwmnj\" (UID: \"a6e0ff27-df96-47ca-897a-53b8ad395f24\") " pod="openshift-marketplace/community-operators-rwmnj" Dec 01 14:49:03 crc kubenswrapper[4810]: I1201 14:49:03.317323 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6e0ff27-df96-47ca-897a-53b8ad395f24-catalog-content\") pod \"community-operators-rwmnj\" (UID: \"a6e0ff27-df96-47ca-897a-53b8ad395f24\") " pod="openshift-marketplace/community-operators-rwmnj" Dec 01 14:49:03 crc kubenswrapper[4810]: I1201 14:49:03.317612 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6e0ff27-df96-47ca-897a-53b8ad395f24-utilities\") pod \"community-operators-rwmnj\" (UID: \"a6e0ff27-df96-47ca-897a-53b8ad395f24\") " pod="openshift-marketplace/community-operators-rwmnj" Dec 01 14:49:03 crc kubenswrapper[4810]: I1201 14:49:03.337442 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt9hp\" (UniqueName: \"kubernetes.io/projected/a6e0ff27-df96-47ca-897a-53b8ad395f24-kube-api-access-wt9hp\") pod \"community-operators-rwmnj\" (UID: \"a6e0ff27-df96-47ca-897a-53b8ad395f24\") " pod="openshift-marketplace/community-operators-rwmnj" Dec 01 14:49:03 crc kubenswrapper[4810]: I1201 14:49:03.492436 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rwmnj" Dec 01 14:49:03 crc kubenswrapper[4810]: I1201 14:49:03.775344 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rwmnj"] Dec 01 14:49:03 crc kubenswrapper[4810]: I1201 14:49:03.815713 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwmnj" event={"ID":"a6e0ff27-df96-47ca-897a-53b8ad395f24","Type":"ContainerStarted","Data":"3b5f0ac0e154e0ebac942ec19bae23f4f290da0730230e63e6f849eaf4c34dc2"} Dec 01 14:49:03 crc kubenswrapper[4810]: I1201 14:49:03.816883 4810 generic.go:334] "Generic (PLEG): container finished" podID="673fd73e-0701-4cf7-a7dd-273db851c2ba" containerID="9cdfe9c7f608ddb760e5c290eaf396ce001af69d392173f2f6e9a5e0f8fc6450" exitCode=0 Dec 01 14:49:03 crc kubenswrapper[4810]: I1201 14:49:03.816910 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" event={"ID":"673fd73e-0701-4cf7-a7dd-273db851c2ba","Type":"ContainerDied","Data":"9cdfe9c7f608ddb760e5c290eaf396ce001af69d392173f2f6e9a5e0f8fc6450"} Dec 01 14:49:04 crc kubenswrapper[4810]: I1201 14:49:04.766457 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-8tcbx" podUID="f65c3828-a372-4660-97dd-e01d6efcb78c" containerName="console" containerID="cri-o://3dd349b806f5cbfea3fb30a21c954f8fa69445b9882244563dba5a8d0a3d14fe" gracePeriod=15 Dec 01 14:49:04 crc kubenswrapper[4810]: I1201 14:49:04.823806 4810 generic.go:334] "Generic (PLEG): container finished" podID="a6e0ff27-df96-47ca-897a-53b8ad395f24" containerID="6abcac1d9b60b611a551194295030c089df5e7acff760afe8c02ecf36f9e6832" exitCode=0 Dec 01 14:49:04 crc kubenswrapper[4810]: I1201 14:49:04.823855 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwmnj" event={"ID":"a6e0ff27-df96-47ca-897a-53b8ad395f24","Type":"ContainerDied","Data":"6abcac1d9b60b611a551194295030c089df5e7acff760afe8c02ecf36f9e6832"} Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.093709 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-8tcbx_f65c3828-a372-4660-97dd-e01d6efcb78c/console/0.log" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.093777 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.252105 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxnbz\" (UniqueName: \"kubernetes.io/projected/f65c3828-a372-4660-97dd-e01d6efcb78c-kube-api-access-xxnbz\") pod \"f65c3828-a372-4660-97dd-e01d6efcb78c\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.252193 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-service-ca\") pod \"f65c3828-a372-4660-97dd-e01d6efcb78c\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.252264 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-console-config\") pod \"f65c3828-a372-4660-97dd-e01d6efcb78c\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.252336 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-oauth-serving-cert\") pod \"f65c3828-a372-4660-97dd-e01d6efcb78c\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.252435 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f65c3828-a372-4660-97dd-e01d6efcb78c-console-oauth-config\") pod \"f65c3828-a372-4660-97dd-e01d6efcb78c\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.252505 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f65c3828-a372-4660-97dd-e01d6efcb78c-console-serving-cert\") pod \"f65c3828-a372-4660-97dd-e01d6efcb78c\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.252575 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-trusted-ca-bundle\") pod \"f65c3828-a372-4660-97dd-e01d6efcb78c\" (UID: \"f65c3828-a372-4660-97dd-e01d6efcb78c\") " Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.253035 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-console-config" (OuterVolumeSpecName: "console-config") pod "f65c3828-a372-4660-97dd-e01d6efcb78c" (UID: "f65c3828-a372-4660-97dd-e01d6efcb78c"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.253116 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-service-ca" (OuterVolumeSpecName: "service-ca") pod "f65c3828-a372-4660-97dd-e01d6efcb78c" (UID: "f65c3828-a372-4660-97dd-e01d6efcb78c"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.253428 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "f65c3828-a372-4660-97dd-e01d6efcb78c" (UID: "f65c3828-a372-4660-97dd-e01d6efcb78c"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.253706 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "f65c3828-a372-4660-97dd-e01d6efcb78c" (UID: "f65c3828-a372-4660-97dd-e01d6efcb78c"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.258270 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f65c3828-a372-4660-97dd-e01d6efcb78c-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "f65c3828-a372-4660-97dd-e01d6efcb78c" (UID: "f65c3828-a372-4660-97dd-e01d6efcb78c"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.258501 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f65c3828-a372-4660-97dd-e01d6efcb78c-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "f65c3828-a372-4660-97dd-e01d6efcb78c" (UID: "f65c3828-a372-4660-97dd-e01d6efcb78c"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.258742 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f65c3828-a372-4660-97dd-e01d6efcb78c-kube-api-access-xxnbz" (OuterVolumeSpecName: "kube-api-access-xxnbz") pod "f65c3828-a372-4660-97dd-e01d6efcb78c" (UID: "f65c3828-a372-4660-97dd-e01d6efcb78c"). InnerVolumeSpecName "kube-api-access-xxnbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.353591 4810 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-console-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.353626 4810 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.353636 4810 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f65c3828-a372-4660-97dd-e01d6efcb78c-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.353647 4810 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f65c3828-a372-4660-97dd-e01d6efcb78c-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.353655 4810 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.353664 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxnbz\" (UniqueName: \"kubernetes.io/projected/f65c3828-a372-4660-97dd-e01d6efcb78c-kube-api-access-xxnbz\") on node \"crc\" DevicePath \"\"" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.353674 4810 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f65c3828-a372-4660-97dd-e01d6efcb78c-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.832538 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwmnj" event={"ID":"a6e0ff27-df96-47ca-897a-53b8ad395f24","Type":"ContainerStarted","Data":"e20119685fb4e6e91afa4119e68bf96c37e6fe373287a7f7d981b8021d2670ec"} Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.834324 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-8tcbx_f65c3828-a372-4660-97dd-e01d6efcb78c/console/0.log" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.834358 4810 generic.go:334] "Generic (PLEG): container finished" podID="f65c3828-a372-4660-97dd-e01d6efcb78c" containerID="3dd349b806f5cbfea3fb30a21c954f8fa69445b9882244563dba5a8d0a3d14fe" exitCode=2 Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.834413 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8tcbx" event={"ID":"f65c3828-a372-4660-97dd-e01d6efcb78c","Type":"ContainerDied","Data":"3dd349b806f5cbfea3fb30a21c954f8fa69445b9882244563dba5a8d0a3d14fe"} Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.834425 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8tcbx" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.834449 4810 scope.go:117] "RemoveContainer" containerID="3dd349b806f5cbfea3fb30a21c954f8fa69445b9882244563dba5a8d0a3d14fe" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.834436 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8tcbx" event={"ID":"f65c3828-a372-4660-97dd-e01d6efcb78c","Type":"ContainerDied","Data":"9de934c9dd0cbb259cfc8bbaa915e0e65562752463c71daca07931128b663989"} Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.860731 4810 scope.go:117] "RemoveContainer" containerID="3dd349b806f5cbfea3fb30a21c954f8fa69445b9882244563dba5a8d0a3d14fe" Dec 01 14:49:05 crc kubenswrapper[4810]: E1201 14:49:05.861074 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dd349b806f5cbfea3fb30a21c954f8fa69445b9882244563dba5a8d0a3d14fe\": container with ID starting with 3dd349b806f5cbfea3fb30a21c954f8fa69445b9882244563dba5a8d0a3d14fe not found: ID does not exist" containerID="3dd349b806f5cbfea3fb30a21c954f8fa69445b9882244563dba5a8d0a3d14fe" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.861159 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dd349b806f5cbfea3fb30a21c954f8fa69445b9882244563dba5a8d0a3d14fe"} err="failed to get container status \"3dd349b806f5cbfea3fb30a21c954f8fa69445b9882244563dba5a8d0a3d14fe\": rpc error: code = NotFound desc = could not find container \"3dd349b806f5cbfea3fb30a21c954f8fa69445b9882244563dba5a8d0a3d14fe\": container with ID starting with 3dd349b806f5cbfea3fb30a21c954f8fa69445b9882244563dba5a8d0a3d14fe not found: ID does not exist" Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.877670 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-8tcbx"] Dec 01 14:49:05 crc kubenswrapper[4810]: I1201 14:49:05.877716 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-8tcbx"] Dec 01 14:49:06 crc kubenswrapper[4810]: I1201 14:49:06.499111 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f65c3828-a372-4660-97dd-e01d6efcb78c" path="/var/lib/kubelet/pods/f65c3828-a372-4660-97dd-e01d6efcb78c/volumes" Dec 01 14:49:06 crc kubenswrapper[4810]: I1201 14:49:06.842573 4810 generic.go:334] "Generic (PLEG): container finished" podID="a6e0ff27-df96-47ca-897a-53b8ad395f24" containerID="e20119685fb4e6e91afa4119e68bf96c37e6fe373287a7f7d981b8021d2670ec" exitCode=0 Dec 01 14:49:06 crc kubenswrapper[4810]: I1201 14:49:06.842640 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwmnj" event={"ID":"a6e0ff27-df96-47ca-897a-53b8ad395f24","Type":"ContainerDied","Data":"e20119685fb4e6e91afa4119e68bf96c37e6fe373287a7f7d981b8021d2670ec"} Dec 01 14:49:06 crc kubenswrapper[4810]: I1201 14:49:06.844141 4810 generic.go:334] "Generic (PLEG): container finished" podID="673fd73e-0701-4cf7-a7dd-273db851c2ba" containerID="fc2239e3513ee5e2487d401f998cf0116615d4686b8d4d36a5f2f4e6072c3195" exitCode=0 Dec 01 14:49:06 crc kubenswrapper[4810]: I1201 14:49:06.844185 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" event={"ID":"673fd73e-0701-4cf7-a7dd-273db851c2ba","Type":"ContainerDied","Data":"fc2239e3513ee5e2487d401f998cf0116615d4686b8d4d36a5f2f4e6072c3195"} Dec 01 14:49:07 crc kubenswrapper[4810]: I1201 14:49:07.851109 4810 generic.go:334] "Generic (PLEG): container finished" podID="673fd73e-0701-4cf7-a7dd-273db851c2ba" containerID="773faedeb9f9b967dc434003e07d19dc014e4085dd5483ace5b05912a0768904" exitCode=0 Dec 01 14:49:07 crc kubenswrapper[4810]: I1201 14:49:07.851188 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" event={"ID":"673fd73e-0701-4cf7-a7dd-273db851c2ba","Type":"ContainerDied","Data":"773faedeb9f9b967dc434003e07d19dc014e4085dd5483ace5b05912a0768904"} Dec 01 14:49:07 crc kubenswrapper[4810]: I1201 14:49:07.853181 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwmnj" event={"ID":"a6e0ff27-df96-47ca-897a-53b8ad395f24","Type":"ContainerStarted","Data":"d5db9e855e187ffa3c7eff931e584be5f989d879fcf276d5dbef1e28666df967"} Dec 01 14:49:07 crc kubenswrapper[4810]: I1201 14:49:07.896385 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rwmnj" podStartSLOduration=2.135516275 podStartE2EDuration="4.896357978s" podCreationTimestamp="2025-12-01 14:49:03 +0000 UTC" firstStartedPulling="2025-12-01 14:49:04.826368215 +0000 UTC m=+910.589877818" lastFinishedPulling="2025-12-01 14:49:07.587209908 +0000 UTC m=+913.350719521" observedRunningTime="2025-12-01 14:49:07.894874767 +0000 UTC m=+913.658384420" watchObservedRunningTime="2025-12-01 14:49:07.896357978 +0000 UTC m=+913.659867571" Dec 01 14:49:09 crc kubenswrapper[4810]: I1201 14:49:09.219145 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" Dec 01 14:49:09 crc kubenswrapper[4810]: I1201 14:49:09.358179 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/673fd73e-0701-4cf7-a7dd-273db851c2ba-bundle\") pod \"673fd73e-0701-4cf7-a7dd-273db851c2ba\" (UID: \"673fd73e-0701-4cf7-a7dd-273db851c2ba\") " Dec 01 14:49:09 crc kubenswrapper[4810]: I1201 14:49:09.358352 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/673fd73e-0701-4cf7-a7dd-273db851c2ba-util\") pod \"673fd73e-0701-4cf7-a7dd-273db851c2ba\" (UID: \"673fd73e-0701-4cf7-a7dd-273db851c2ba\") " Dec 01 14:49:09 crc kubenswrapper[4810]: I1201 14:49:09.358392 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4d5m\" (UniqueName: \"kubernetes.io/projected/673fd73e-0701-4cf7-a7dd-273db851c2ba-kube-api-access-p4d5m\") pod \"673fd73e-0701-4cf7-a7dd-273db851c2ba\" (UID: \"673fd73e-0701-4cf7-a7dd-273db851c2ba\") " Dec 01 14:49:09 crc kubenswrapper[4810]: I1201 14:49:09.359258 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/673fd73e-0701-4cf7-a7dd-273db851c2ba-bundle" (OuterVolumeSpecName: "bundle") pod "673fd73e-0701-4cf7-a7dd-273db851c2ba" (UID: "673fd73e-0701-4cf7-a7dd-273db851c2ba"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:49:09 crc kubenswrapper[4810]: I1201 14:49:09.368280 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/673fd73e-0701-4cf7-a7dd-273db851c2ba-util" (OuterVolumeSpecName: "util") pod "673fd73e-0701-4cf7-a7dd-273db851c2ba" (UID: "673fd73e-0701-4cf7-a7dd-273db851c2ba"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:49:09 crc kubenswrapper[4810]: I1201 14:49:09.371734 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/673fd73e-0701-4cf7-a7dd-273db851c2ba-kube-api-access-p4d5m" (OuterVolumeSpecName: "kube-api-access-p4d5m") pod "673fd73e-0701-4cf7-a7dd-273db851c2ba" (UID: "673fd73e-0701-4cf7-a7dd-273db851c2ba"). InnerVolumeSpecName "kube-api-access-p4d5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:49:09 crc kubenswrapper[4810]: I1201 14:49:09.459532 4810 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/673fd73e-0701-4cf7-a7dd-273db851c2ba-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:49:09 crc kubenswrapper[4810]: I1201 14:49:09.459577 4810 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/673fd73e-0701-4cf7-a7dd-273db851c2ba-util\") on node \"crc\" DevicePath \"\"" Dec 01 14:49:09 crc kubenswrapper[4810]: I1201 14:49:09.459590 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4d5m\" (UniqueName: \"kubernetes.io/projected/673fd73e-0701-4cf7-a7dd-273db851c2ba-kube-api-access-p4d5m\") on node \"crc\" DevicePath \"\"" Dec 01 14:49:09 crc kubenswrapper[4810]: I1201 14:49:09.866740 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" event={"ID":"673fd73e-0701-4cf7-a7dd-273db851c2ba","Type":"ContainerDied","Data":"d7c0be0bf38df9656a17178a8e793f197ac9133797ae95e7f74737f507155b58"} Dec 01 14:49:09 crc kubenswrapper[4810]: I1201 14:49:09.867107 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7c0be0bf38df9656a17178a8e793f197ac9133797ae95e7f74737f507155b58" Dec 01 14:49:09 crc kubenswrapper[4810]: I1201 14:49:09.866969 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv" Dec 01 14:49:10 crc kubenswrapper[4810]: I1201 14:49:10.959030 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8d79z"] Dec 01 14:49:10 crc kubenswrapper[4810]: E1201 14:49:10.959295 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="673fd73e-0701-4cf7-a7dd-273db851c2ba" containerName="pull" Dec 01 14:49:10 crc kubenswrapper[4810]: I1201 14:49:10.959309 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="673fd73e-0701-4cf7-a7dd-273db851c2ba" containerName="pull" Dec 01 14:49:10 crc kubenswrapper[4810]: E1201 14:49:10.959323 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="673fd73e-0701-4cf7-a7dd-273db851c2ba" containerName="extract" Dec 01 14:49:10 crc kubenswrapper[4810]: I1201 14:49:10.959331 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="673fd73e-0701-4cf7-a7dd-273db851c2ba" containerName="extract" Dec 01 14:49:10 crc kubenswrapper[4810]: E1201 14:49:10.959342 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="673fd73e-0701-4cf7-a7dd-273db851c2ba" containerName="util" Dec 01 14:49:10 crc kubenswrapper[4810]: I1201 14:49:10.959352 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="673fd73e-0701-4cf7-a7dd-273db851c2ba" containerName="util" Dec 01 14:49:10 crc kubenswrapper[4810]: E1201 14:49:10.959371 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f65c3828-a372-4660-97dd-e01d6efcb78c" containerName="console" Dec 01 14:49:10 crc kubenswrapper[4810]: I1201 14:49:10.959379 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f65c3828-a372-4660-97dd-e01d6efcb78c" containerName="console" Dec 01 14:49:10 crc kubenswrapper[4810]: I1201 14:49:10.959504 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f65c3828-a372-4660-97dd-e01d6efcb78c" containerName="console" Dec 01 14:49:10 crc kubenswrapper[4810]: I1201 14:49:10.959522 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="673fd73e-0701-4cf7-a7dd-273db851c2ba" containerName="extract" Dec 01 14:49:10 crc kubenswrapper[4810]: I1201 14:49:10.961682 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8d79z" Dec 01 14:49:10 crc kubenswrapper[4810]: I1201 14:49:10.976356 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8d79z"] Dec 01 14:49:11 crc kubenswrapper[4810]: I1201 14:49:11.092728 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76135880-43c5-43c3-bb4d-7591e82a0a29-utilities\") pod \"redhat-marketplace-8d79z\" (UID: \"76135880-43c5-43c3-bb4d-7591e82a0a29\") " pod="openshift-marketplace/redhat-marketplace-8d79z" Dec 01 14:49:11 crc kubenswrapper[4810]: I1201 14:49:11.092798 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kzj5\" (UniqueName: \"kubernetes.io/projected/76135880-43c5-43c3-bb4d-7591e82a0a29-kube-api-access-8kzj5\") pod \"redhat-marketplace-8d79z\" (UID: \"76135880-43c5-43c3-bb4d-7591e82a0a29\") " pod="openshift-marketplace/redhat-marketplace-8d79z" Dec 01 14:49:11 crc kubenswrapper[4810]: I1201 14:49:11.092931 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76135880-43c5-43c3-bb4d-7591e82a0a29-catalog-content\") pod \"redhat-marketplace-8d79z\" (UID: \"76135880-43c5-43c3-bb4d-7591e82a0a29\") " pod="openshift-marketplace/redhat-marketplace-8d79z" Dec 01 14:49:11 crc kubenswrapper[4810]: I1201 14:49:11.193670 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76135880-43c5-43c3-bb4d-7591e82a0a29-catalog-content\") pod \"redhat-marketplace-8d79z\" (UID: \"76135880-43c5-43c3-bb4d-7591e82a0a29\") " pod="openshift-marketplace/redhat-marketplace-8d79z" Dec 01 14:49:11 crc kubenswrapper[4810]: I1201 14:49:11.193722 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76135880-43c5-43c3-bb4d-7591e82a0a29-utilities\") pod \"redhat-marketplace-8d79z\" (UID: \"76135880-43c5-43c3-bb4d-7591e82a0a29\") " pod="openshift-marketplace/redhat-marketplace-8d79z" Dec 01 14:49:11 crc kubenswrapper[4810]: I1201 14:49:11.193743 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kzj5\" (UniqueName: \"kubernetes.io/projected/76135880-43c5-43c3-bb4d-7591e82a0a29-kube-api-access-8kzj5\") pod \"redhat-marketplace-8d79z\" (UID: \"76135880-43c5-43c3-bb4d-7591e82a0a29\") " pod="openshift-marketplace/redhat-marketplace-8d79z" Dec 01 14:49:11 crc kubenswrapper[4810]: I1201 14:49:11.194165 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76135880-43c5-43c3-bb4d-7591e82a0a29-utilities\") pod \"redhat-marketplace-8d79z\" (UID: \"76135880-43c5-43c3-bb4d-7591e82a0a29\") " pod="openshift-marketplace/redhat-marketplace-8d79z" Dec 01 14:49:11 crc kubenswrapper[4810]: I1201 14:49:11.194242 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76135880-43c5-43c3-bb4d-7591e82a0a29-catalog-content\") pod \"redhat-marketplace-8d79z\" (UID: \"76135880-43c5-43c3-bb4d-7591e82a0a29\") " pod="openshift-marketplace/redhat-marketplace-8d79z" Dec 01 14:49:11 crc kubenswrapper[4810]: I1201 14:49:11.213361 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kzj5\" (UniqueName: \"kubernetes.io/projected/76135880-43c5-43c3-bb4d-7591e82a0a29-kube-api-access-8kzj5\") pod \"redhat-marketplace-8d79z\" (UID: \"76135880-43c5-43c3-bb4d-7591e82a0a29\") " pod="openshift-marketplace/redhat-marketplace-8d79z" Dec 01 14:49:11 crc kubenswrapper[4810]: I1201 14:49:11.276807 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8d79z" Dec 01 14:49:11 crc kubenswrapper[4810]: I1201 14:49:11.499353 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8d79z"] Dec 01 14:49:11 crc kubenswrapper[4810]: I1201 14:49:11.878262 4810 generic.go:334] "Generic (PLEG): container finished" podID="76135880-43c5-43c3-bb4d-7591e82a0a29" containerID="4ebe0e45ef0dfc4e919fdea6b705b9d8ba3ab57e0841bd904ef357ccd91f1620" exitCode=0 Dec 01 14:49:11 crc kubenswrapper[4810]: I1201 14:49:11.878364 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8d79z" event={"ID":"76135880-43c5-43c3-bb4d-7591e82a0a29","Type":"ContainerDied","Data":"4ebe0e45ef0dfc4e919fdea6b705b9d8ba3ab57e0841bd904ef357ccd91f1620"} Dec 01 14:49:11 crc kubenswrapper[4810]: I1201 14:49:11.878663 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8d79z" event={"ID":"76135880-43c5-43c3-bb4d-7591e82a0a29","Type":"ContainerStarted","Data":"71d6dbd26523050c4fc3ee0bf0673741b80d3934b6e240a1f92f7588430d4ed4"} Dec 01 14:49:12 crc kubenswrapper[4810]: I1201 14:49:12.886113 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8d79z" event={"ID":"76135880-43c5-43c3-bb4d-7591e82a0a29","Type":"ContainerStarted","Data":"f832ce0f2b9efeb7bbb97a6637f468b889439fb27ad4979bc4b6aafc49e04680"} Dec 01 14:49:13 crc kubenswrapper[4810]: I1201 14:49:13.492849 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rwmnj" Dec 01 14:49:13 crc kubenswrapper[4810]: I1201 14:49:13.492899 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rwmnj" Dec 01 14:49:13 crc kubenswrapper[4810]: I1201 14:49:13.534321 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rwmnj" Dec 01 14:49:13 crc kubenswrapper[4810]: I1201 14:49:13.913141 4810 generic.go:334] "Generic (PLEG): container finished" podID="76135880-43c5-43c3-bb4d-7591e82a0a29" containerID="f832ce0f2b9efeb7bbb97a6637f468b889439fb27ad4979bc4b6aafc49e04680" exitCode=0 Dec 01 14:49:13 crc kubenswrapper[4810]: I1201 14:49:13.915242 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8d79z" event={"ID":"76135880-43c5-43c3-bb4d-7591e82a0a29","Type":"ContainerDied","Data":"f832ce0f2b9efeb7bbb97a6637f468b889439fb27ad4979bc4b6aafc49e04680"} Dec 01 14:49:14 crc kubenswrapper[4810]: I1201 14:49:14.024608 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rwmnj" Dec 01 14:49:14 crc kubenswrapper[4810]: I1201 14:49:14.920700 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8d79z" event={"ID":"76135880-43c5-43c3-bb4d-7591e82a0a29","Type":"ContainerStarted","Data":"ed4a69e43f531cafb475cd93c40be7672b67ede892d91e7c12789669e9c1ccb3"} Dec 01 14:49:15 crc kubenswrapper[4810]: I1201 14:49:15.551647 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8d79z" podStartSLOduration=2.9715102460000002 podStartE2EDuration="5.551621649s" podCreationTimestamp="2025-12-01 14:49:10 +0000 UTC" firstStartedPulling="2025-12-01 14:49:11.880542147 +0000 UTC m=+917.644051750" lastFinishedPulling="2025-12-01 14:49:14.46065355 +0000 UTC m=+920.224163153" observedRunningTime="2025-12-01 14:49:14.938794439 +0000 UTC m=+920.702304042" watchObservedRunningTime="2025-12-01 14:49:15.551621649 +0000 UTC m=+921.315131292" Dec 01 14:49:15 crc kubenswrapper[4810]: I1201 14:49:15.552285 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rwmnj"] Dec 01 14:49:15 crc kubenswrapper[4810]: I1201 14:49:15.927133 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rwmnj" podUID="a6e0ff27-df96-47ca-897a-53b8ad395f24" containerName="registry-server" containerID="cri-o://d5db9e855e187ffa3c7eff931e584be5f989d879fcf276d5dbef1e28666df967" gracePeriod=2 Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.287542 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rwmnj" Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.374644 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6e0ff27-df96-47ca-897a-53b8ad395f24-utilities\") pod \"a6e0ff27-df96-47ca-897a-53b8ad395f24\" (UID: \"a6e0ff27-df96-47ca-897a-53b8ad395f24\") " Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.374756 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6e0ff27-df96-47ca-897a-53b8ad395f24-catalog-content\") pod \"a6e0ff27-df96-47ca-897a-53b8ad395f24\" (UID: \"a6e0ff27-df96-47ca-897a-53b8ad395f24\") " Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.374793 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wt9hp\" (UniqueName: \"kubernetes.io/projected/a6e0ff27-df96-47ca-897a-53b8ad395f24-kube-api-access-wt9hp\") pod \"a6e0ff27-df96-47ca-897a-53b8ad395f24\" (UID: \"a6e0ff27-df96-47ca-897a-53b8ad395f24\") " Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.375516 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6e0ff27-df96-47ca-897a-53b8ad395f24-utilities" (OuterVolumeSpecName: "utilities") pod "a6e0ff27-df96-47ca-897a-53b8ad395f24" (UID: "a6e0ff27-df96-47ca-897a-53b8ad395f24"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.380905 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6e0ff27-df96-47ca-897a-53b8ad395f24-kube-api-access-wt9hp" (OuterVolumeSpecName: "kube-api-access-wt9hp") pod "a6e0ff27-df96-47ca-897a-53b8ad395f24" (UID: "a6e0ff27-df96-47ca-897a-53b8ad395f24"). InnerVolumeSpecName "kube-api-access-wt9hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.437119 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6e0ff27-df96-47ca-897a-53b8ad395f24-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a6e0ff27-df96-47ca-897a-53b8ad395f24" (UID: "a6e0ff27-df96-47ca-897a-53b8ad395f24"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.475717 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6e0ff27-df96-47ca-897a-53b8ad395f24-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.475752 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wt9hp\" (UniqueName: \"kubernetes.io/projected/a6e0ff27-df96-47ca-897a-53b8ad395f24-kube-api-access-wt9hp\") on node \"crc\" DevicePath \"\"" Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.475762 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6e0ff27-df96-47ca-897a-53b8ad395f24-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.934519 4810 generic.go:334] "Generic (PLEG): container finished" podID="a6e0ff27-df96-47ca-897a-53b8ad395f24" containerID="d5db9e855e187ffa3c7eff931e584be5f989d879fcf276d5dbef1e28666df967" exitCode=0 Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.934561 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwmnj" event={"ID":"a6e0ff27-df96-47ca-897a-53b8ad395f24","Type":"ContainerDied","Data":"d5db9e855e187ffa3c7eff931e584be5f989d879fcf276d5dbef1e28666df967"} Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.934589 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rwmnj" Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.934606 4810 scope.go:117] "RemoveContainer" containerID="d5db9e855e187ffa3c7eff931e584be5f989d879fcf276d5dbef1e28666df967" Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.934594 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rwmnj" event={"ID":"a6e0ff27-df96-47ca-897a-53b8ad395f24","Type":"ContainerDied","Data":"3b5f0ac0e154e0ebac942ec19bae23f4f290da0730230e63e6f849eaf4c34dc2"} Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.952166 4810 scope.go:117] "RemoveContainer" containerID="e20119685fb4e6e91afa4119e68bf96c37e6fe373287a7f7d981b8021d2670ec" Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.960215 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rwmnj"] Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.967428 4810 scope.go:117] "RemoveContainer" containerID="6abcac1d9b60b611a551194295030c089df5e7acff760afe8c02ecf36f9e6832" Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.969402 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rwmnj"] Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.991217 4810 scope.go:117] "RemoveContainer" containerID="d5db9e855e187ffa3c7eff931e584be5f989d879fcf276d5dbef1e28666df967" Dec 01 14:49:16 crc kubenswrapper[4810]: E1201 14:49:16.991683 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5db9e855e187ffa3c7eff931e584be5f989d879fcf276d5dbef1e28666df967\": container with ID starting with d5db9e855e187ffa3c7eff931e584be5f989d879fcf276d5dbef1e28666df967 not found: ID does not exist" containerID="d5db9e855e187ffa3c7eff931e584be5f989d879fcf276d5dbef1e28666df967" Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.991715 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5db9e855e187ffa3c7eff931e584be5f989d879fcf276d5dbef1e28666df967"} err="failed to get container status \"d5db9e855e187ffa3c7eff931e584be5f989d879fcf276d5dbef1e28666df967\": rpc error: code = NotFound desc = could not find container \"d5db9e855e187ffa3c7eff931e584be5f989d879fcf276d5dbef1e28666df967\": container with ID starting with d5db9e855e187ffa3c7eff931e584be5f989d879fcf276d5dbef1e28666df967 not found: ID does not exist" Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.991739 4810 scope.go:117] "RemoveContainer" containerID="e20119685fb4e6e91afa4119e68bf96c37e6fe373287a7f7d981b8021d2670ec" Dec 01 14:49:16 crc kubenswrapper[4810]: E1201 14:49:16.992060 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e20119685fb4e6e91afa4119e68bf96c37e6fe373287a7f7d981b8021d2670ec\": container with ID starting with e20119685fb4e6e91afa4119e68bf96c37e6fe373287a7f7d981b8021d2670ec not found: ID does not exist" containerID="e20119685fb4e6e91afa4119e68bf96c37e6fe373287a7f7d981b8021d2670ec" Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.992078 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e20119685fb4e6e91afa4119e68bf96c37e6fe373287a7f7d981b8021d2670ec"} err="failed to get container status \"e20119685fb4e6e91afa4119e68bf96c37e6fe373287a7f7d981b8021d2670ec\": rpc error: code = NotFound desc = could not find container \"e20119685fb4e6e91afa4119e68bf96c37e6fe373287a7f7d981b8021d2670ec\": container with ID starting with e20119685fb4e6e91afa4119e68bf96c37e6fe373287a7f7d981b8021d2670ec not found: ID does not exist" Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.992089 4810 scope.go:117] "RemoveContainer" containerID="6abcac1d9b60b611a551194295030c089df5e7acff760afe8c02ecf36f9e6832" Dec 01 14:49:16 crc kubenswrapper[4810]: E1201 14:49:16.992409 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6abcac1d9b60b611a551194295030c089df5e7acff760afe8c02ecf36f9e6832\": container with ID starting with 6abcac1d9b60b611a551194295030c089df5e7acff760afe8c02ecf36f9e6832 not found: ID does not exist" containerID="6abcac1d9b60b611a551194295030c089df5e7acff760afe8c02ecf36f9e6832" Dec 01 14:49:16 crc kubenswrapper[4810]: I1201 14:49:16.992429 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6abcac1d9b60b611a551194295030c089df5e7acff760afe8c02ecf36f9e6832"} err="failed to get container status \"6abcac1d9b60b611a551194295030c089df5e7acff760afe8c02ecf36f9e6832\": rpc error: code = NotFound desc = could not find container \"6abcac1d9b60b611a551194295030c089df5e7acff760afe8c02ecf36f9e6832\": container with ID starting with 6abcac1d9b60b611a551194295030c089df5e7acff760afe8c02ecf36f9e6832 not found: ID does not exist" Dec 01 14:49:17 crc kubenswrapper[4810]: I1201 14:49:17.937168 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w"] Dec 01 14:49:17 crc kubenswrapper[4810]: E1201 14:49:17.937376 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6e0ff27-df96-47ca-897a-53b8ad395f24" containerName="extract-content" Dec 01 14:49:17 crc kubenswrapper[4810]: I1201 14:49:17.937594 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6e0ff27-df96-47ca-897a-53b8ad395f24" containerName="extract-content" Dec 01 14:49:17 crc kubenswrapper[4810]: E1201 14:49:17.937612 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6e0ff27-df96-47ca-897a-53b8ad395f24" containerName="extract-utilities" Dec 01 14:49:17 crc kubenswrapper[4810]: I1201 14:49:17.937620 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6e0ff27-df96-47ca-897a-53b8ad395f24" containerName="extract-utilities" Dec 01 14:49:17 crc kubenswrapper[4810]: E1201 14:49:17.937639 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6e0ff27-df96-47ca-897a-53b8ad395f24" containerName="registry-server" Dec 01 14:49:17 crc kubenswrapper[4810]: I1201 14:49:17.937648 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6e0ff27-df96-47ca-897a-53b8ad395f24" containerName="registry-server" Dec 01 14:49:17 crc kubenswrapper[4810]: I1201 14:49:17.937753 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6e0ff27-df96-47ca-897a-53b8ad395f24" containerName="registry-server" Dec 01 14:49:17 crc kubenswrapper[4810]: I1201 14:49:17.938157 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w" Dec 01 14:49:17 crc kubenswrapper[4810]: I1201 14:49:17.939995 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 01 14:49:17 crc kubenswrapper[4810]: I1201 14:49:17.940379 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-zgpzw" Dec 01 14:49:17 crc kubenswrapper[4810]: I1201 14:49:17.940655 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 01 14:49:17 crc kubenswrapper[4810]: I1201 14:49:17.940773 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 01 14:49:17 crc kubenswrapper[4810]: I1201 14:49:17.941024 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 01 14:49:17 crc kubenswrapper[4810]: I1201 14:49:17.956370 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w"] Dec 01 14:49:17 crc kubenswrapper[4810]: I1201 14:49:17.996090 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qps9\" (UniqueName: \"kubernetes.io/projected/6e8d7459-6378-4fa0-90c2-0533b2407764-kube-api-access-7qps9\") pod \"metallb-operator-controller-manager-855bb66d65-ldd2w\" (UID: \"6e8d7459-6378-4fa0-90c2-0533b2407764\") " pod="metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w" Dec 01 14:49:17 crc kubenswrapper[4810]: I1201 14:49:17.996161 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6e8d7459-6378-4fa0-90c2-0533b2407764-webhook-cert\") pod \"metallb-operator-controller-manager-855bb66d65-ldd2w\" (UID: \"6e8d7459-6378-4fa0-90c2-0533b2407764\") " pod="metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w" Dec 01 14:49:17 crc kubenswrapper[4810]: I1201 14:49:17.996182 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6e8d7459-6378-4fa0-90c2-0533b2407764-apiservice-cert\") pod \"metallb-operator-controller-manager-855bb66d65-ldd2w\" (UID: \"6e8d7459-6378-4fa0-90c2-0533b2407764\") " pod="metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.097277 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6e8d7459-6378-4fa0-90c2-0533b2407764-webhook-cert\") pod \"metallb-operator-controller-manager-855bb66d65-ldd2w\" (UID: \"6e8d7459-6378-4fa0-90c2-0533b2407764\") " pod="metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.097354 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6e8d7459-6378-4fa0-90c2-0533b2407764-apiservice-cert\") pod \"metallb-operator-controller-manager-855bb66d65-ldd2w\" (UID: \"6e8d7459-6378-4fa0-90c2-0533b2407764\") " pod="metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.097404 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qps9\" (UniqueName: \"kubernetes.io/projected/6e8d7459-6378-4fa0-90c2-0533b2407764-kube-api-access-7qps9\") pod \"metallb-operator-controller-manager-855bb66d65-ldd2w\" (UID: \"6e8d7459-6378-4fa0-90c2-0533b2407764\") " pod="metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.102217 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6e8d7459-6378-4fa0-90c2-0533b2407764-webhook-cert\") pod \"metallb-operator-controller-manager-855bb66d65-ldd2w\" (UID: \"6e8d7459-6378-4fa0-90c2-0533b2407764\") " pod="metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.102217 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6e8d7459-6378-4fa0-90c2-0533b2407764-apiservice-cert\") pod \"metallb-operator-controller-manager-855bb66d65-ldd2w\" (UID: \"6e8d7459-6378-4fa0-90c2-0533b2407764\") " pod="metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.135293 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qps9\" (UniqueName: \"kubernetes.io/projected/6e8d7459-6378-4fa0-90c2-0533b2407764-kube-api-access-7qps9\") pod \"metallb-operator-controller-manager-855bb66d65-ldd2w\" (UID: \"6e8d7459-6378-4fa0-90c2-0533b2407764\") " pod="metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.182648 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt"] Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.183662 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.186161 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.186519 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.187990 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-fgwz8" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.198243 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms5kd\" (UniqueName: \"kubernetes.io/projected/55a90415-842f-4a4f-b9ee-7932c3baca33-kube-api-access-ms5kd\") pod \"metallb-operator-webhook-server-7c9798ff88-mt9tt\" (UID: \"55a90415-842f-4a4f-b9ee-7932c3baca33\") " pod="metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.198335 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/55a90415-842f-4a4f-b9ee-7932c3baca33-apiservice-cert\") pod \"metallb-operator-webhook-server-7c9798ff88-mt9tt\" (UID: \"55a90415-842f-4a4f-b9ee-7932c3baca33\") " pod="metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.198375 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/55a90415-842f-4a4f-b9ee-7932c3baca33-webhook-cert\") pod \"metallb-operator-webhook-server-7c9798ff88-mt9tt\" (UID: \"55a90415-842f-4a4f-b9ee-7932c3baca33\") " pod="metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.208998 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt"] Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.252281 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.300073 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms5kd\" (UniqueName: \"kubernetes.io/projected/55a90415-842f-4a4f-b9ee-7932c3baca33-kube-api-access-ms5kd\") pod \"metallb-operator-webhook-server-7c9798ff88-mt9tt\" (UID: \"55a90415-842f-4a4f-b9ee-7932c3baca33\") " pod="metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.300171 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/55a90415-842f-4a4f-b9ee-7932c3baca33-apiservice-cert\") pod \"metallb-operator-webhook-server-7c9798ff88-mt9tt\" (UID: \"55a90415-842f-4a4f-b9ee-7932c3baca33\") " pod="metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.300218 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/55a90415-842f-4a4f-b9ee-7932c3baca33-webhook-cert\") pod \"metallb-operator-webhook-server-7c9798ff88-mt9tt\" (UID: \"55a90415-842f-4a4f-b9ee-7932c3baca33\") " pod="metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.304213 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/55a90415-842f-4a4f-b9ee-7932c3baca33-webhook-cert\") pod \"metallb-operator-webhook-server-7c9798ff88-mt9tt\" (UID: \"55a90415-842f-4a4f-b9ee-7932c3baca33\") " pod="metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.304425 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/55a90415-842f-4a4f-b9ee-7932c3baca33-apiservice-cert\") pod \"metallb-operator-webhook-server-7c9798ff88-mt9tt\" (UID: \"55a90415-842f-4a4f-b9ee-7932c3baca33\") " pod="metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.337007 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms5kd\" (UniqueName: \"kubernetes.io/projected/55a90415-842f-4a4f-b9ee-7932c3baca33-kube-api-access-ms5kd\") pod \"metallb-operator-webhook-server-7c9798ff88-mt9tt\" (UID: \"55a90415-842f-4a4f-b9ee-7932c3baca33\") " pod="metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.499741 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt" Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.510183 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6e0ff27-df96-47ca-897a-53b8ad395f24" path="/var/lib/kubelet/pods/a6e0ff27-df96-47ca-897a-53b8ad395f24/volumes" Dec 01 14:49:18 crc kubenswrapper[4810]: W1201 14:49:18.767541 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e8d7459_6378_4fa0_90c2_0533b2407764.slice/crio-26d73f457f9356da950048af39eb31cd566b6e2cb296d724ae62e22db23795db WatchSource:0}: Error finding container 26d73f457f9356da950048af39eb31cd566b6e2cb296d724ae62e22db23795db: Status 404 returned error can't find the container with id 26d73f457f9356da950048af39eb31cd566b6e2cb296d724ae62e22db23795db Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.771507 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w"] Dec 01 14:49:18 crc kubenswrapper[4810]: I1201 14:49:18.946420 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w" event={"ID":"6e8d7459-6378-4fa0-90c2-0533b2407764","Type":"ContainerStarted","Data":"26d73f457f9356da950048af39eb31cd566b6e2cb296d724ae62e22db23795db"} Dec 01 14:49:19 crc kubenswrapper[4810]: I1201 14:49:19.020490 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt"] Dec 01 14:49:19 crc kubenswrapper[4810]: W1201 14:49:19.028776 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55a90415_842f_4a4f_b9ee_7932c3baca33.slice/crio-923a7d19a15eca77c0269bcabfad31138d731c442285cf79008b096ba5c752dd WatchSource:0}: Error finding container 923a7d19a15eca77c0269bcabfad31138d731c442285cf79008b096ba5c752dd: Status 404 returned error can't find the container with id 923a7d19a15eca77c0269bcabfad31138d731c442285cf79008b096ba5c752dd Dec 01 14:49:19 crc kubenswrapper[4810]: I1201 14:49:19.952741 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt" event={"ID":"55a90415-842f-4a4f-b9ee-7932c3baca33","Type":"ContainerStarted","Data":"923a7d19a15eca77c0269bcabfad31138d731c442285cf79008b096ba5c752dd"} Dec 01 14:49:21 crc kubenswrapper[4810]: I1201 14:49:21.277425 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8d79z" Dec 01 14:49:21 crc kubenswrapper[4810]: I1201 14:49:21.277491 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8d79z" Dec 01 14:49:21 crc kubenswrapper[4810]: I1201 14:49:21.318103 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8d79z" Dec 01 14:49:22 crc kubenswrapper[4810]: I1201 14:49:22.035302 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8d79z" Dec 01 14:49:22 crc kubenswrapper[4810]: I1201 14:49:22.973586 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w" event={"ID":"6e8d7459-6378-4fa0-90c2-0533b2407764","Type":"ContainerStarted","Data":"7bc3d00200bd1714cff82cd15716e2f7cb1cc521a72f2cb5c7a458787ddabd12"} Dec 01 14:49:22 crc kubenswrapper[4810]: I1201 14:49:22.973853 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w" Dec 01 14:49:22 crc kubenswrapper[4810]: I1201 14:49:22.992494 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w" podStartSLOduration=1.971908375 podStartE2EDuration="5.992452929s" podCreationTimestamp="2025-12-01 14:49:17 +0000 UTC" firstStartedPulling="2025-12-01 14:49:18.76975003 +0000 UTC m=+924.533259633" lastFinishedPulling="2025-12-01 14:49:22.790294584 +0000 UTC m=+928.553804187" observedRunningTime="2025-12-01 14:49:22.992105309 +0000 UTC m=+928.755614912" watchObservedRunningTime="2025-12-01 14:49:22.992452929 +0000 UTC m=+928.755962532" Dec 01 14:49:23 crc kubenswrapper[4810]: I1201 14:49:23.350249 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8d79z"] Dec 01 14:49:24 crc kubenswrapper[4810]: I1201 14:49:24.045868 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8d79z" podUID="76135880-43c5-43c3-bb4d-7591e82a0a29" containerName="registry-server" containerID="cri-o://ed4a69e43f531cafb475cd93c40be7672b67ede892d91e7c12789669e9c1ccb3" gracePeriod=2 Dec 01 14:49:24 crc kubenswrapper[4810]: I1201 14:49:24.464157 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8d79z" Dec 01 14:49:24 crc kubenswrapper[4810]: I1201 14:49:24.566406 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76135880-43c5-43c3-bb4d-7591e82a0a29-utilities\") pod \"76135880-43c5-43c3-bb4d-7591e82a0a29\" (UID: \"76135880-43c5-43c3-bb4d-7591e82a0a29\") " Dec 01 14:49:24 crc kubenswrapper[4810]: I1201 14:49:24.566507 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76135880-43c5-43c3-bb4d-7591e82a0a29-catalog-content\") pod \"76135880-43c5-43c3-bb4d-7591e82a0a29\" (UID: \"76135880-43c5-43c3-bb4d-7591e82a0a29\") " Dec 01 14:49:24 crc kubenswrapper[4810]: I1201 14:49:24.566560 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kzj5\" (UniqueName: \"kubernetes.io/projected/76135880-43c5-43c3-bb4d-7591e82a0a29-kube-api-access-8kzj5\") pod \"76135880-43c5-43c3-bb4d-7591e82a0a29\" (UID: \"76135880-43c5-43c3-bb4d-7591e82a0a29\") " Dec 01 14:49:24 crc kubenswrapper[4810]: I1201 14:49:24.568258 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76135880-43c5-43c3-bb4d-7591e82a0a29-utilities" (OuterVolumeSpecName: "utilities") pod "76135880-43c5-43c3-bb4d-7591e82a0a29" (UID: "76135880-43c5-43c3-bb4d-7591e82a0a29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:49:24 crc kubenswrapper[4810]: I1201 14:49:24.571943 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76135880-43c5-43c3-bb4d-7591e82a0a29-kube-api-access-8kzj5" (OuterVolumeSpecName: "kube-api-access-8kzj5") pod "76135880-43c5-43c3-bb4d-7591e82a0a29" (UID: "76135880-43c5-43c3-bb4d-7591e82a0a29"). InnerVolumeSpecName "kube-api-access-8kzj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:49:24 crc kubenswrapper[4810]: I1201 14:49:24.584502 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76135880-43c5-43c3-bb4d-7591e82a0a29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76135880-43c5-43c3-bb4d-7591e82a0a29" (UID: "76135880-43c5-43c3-bb4d-7591e82a0a29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:49:24 crc kubenswrapper[4810]: I1201 14:49:24.668332 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76135880-43c5-43c3-bb4d-7591e82a0a29-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 14:49:24 crc kubenswrapper[4810]: I1201 14:49:24.668368 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76135880-43c5-43c3-bb4d-7591e82a0a29-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 14:49:24 crc kubenswrapper[4810]: I1201 14:49:24.668379 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kzj5\" (UniqueName: \"kubernetes.io/projected/76135880-43c5-43c3-bb4d-7591e82a0a29-kube-api-access-8kzj5\") on node \"crc\" DevicePath \"\"" Dec 01 14:49:25 crc kubenswrapper[4810]: I1201 14:49:25.060795 4810 generic.go:334] "Generic (PLEG): container finished" podID="76135880-43c5-43c3-bb4d-7591e82a0a29" containerID="ed4a69e43f531cafb475cd93c40be7672b67ede892d91e7c12789669e9c1ccb3" exitCode=0 Dec 01 14:49:25 crc kubenswrapper[4810]: I1201 14:49:25.060842 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8d79z" event={"ID":"76135880-43c5-43c3-bb4d-7591e82a0a29","Type":"ContainerDied","Data":"ed4a69e43f531cafb475cd93c40be7672b67ede892d91e7c12789669e9c1ccb3"} Dec 01 14:49:25 crc kubenswrapper[4810]: I1201 14:49:25.060873 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8d79z" event={"ID":"76135880-43c5-43c3-bb4d-7591e82a0a29","Type":"ContainerDied","Data":"71d6dbd26523050c4fc3ee0bf0673741b80d3934b6e240a1f92f7588430d4ed4"} Dec 01 14:49:25 crc kubenswrapper[4810]: I1201 14:49:25.060849 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8d79z" Dec 01 14:49:25 crc kubenswrapper[4810]: I1201 14:49:25.060896 4810 scope.go:117] "RemoveContainer" containerID="ed4a69e43f531cafb475cd93c40be7672b67ede892d91e7c12789669e9c1ccb3" Dec 01 14:49:25 crc kubenswrapper[4810]: I1201 14:49:25.089690 4810 scope.go:117] "RemoveContainer" containerID="f832ce0f2b9efeb7bbb97a6637f468b889439fb27ad4979bc4b6aafc49e04680" Dec 01 14:49:25 crc kubenswrapper[4810]: I1201 14:49:25.090234 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8d79z"] Dec 01 14:49:25 crc kubenswrapper[4810]: I1201 14:49:25.095057 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8d79z"] Dec 01 14:49:25 crc kubenswrapper[4810]: I1201 14:49:25.122663 4810 scope.go:117] "RemoveContainer" containerID="4ebe0e45ef0dfc4e919fdea6b705b9d8ba3ab57e0841bd904ef357ccd91f1620" Dec 01 14:49:25 crc kubenswrapper[4810]: I1201 14:49:25.140274 4810 scope.go:117] "RemoveContainer" containerID="ed4a69e43f531cafb475cd93c40be7672b67ede892d91e7c12789669e9c1ccb3" Dec 01 14:49:25 crc kubenswrapper[4810]: E1201 14:49:25.140833 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed4a69e43f531cafb475cd93c40be7672b67ede892d91e7c12789669e9c1ccb3\": container with ID starting with ed4a69e43f531cafb475cd93c40be7672b67ede892d91e7c12789669e9c1ccb3 not found: ID does not exist" containerID="ed4a69e43f531cafb475cd93c40be7672b67ede892d91e7c12789669e9c1ccb3" Dec 01 14:49:25 crc kubenswrapper[4810]: I1201 14:49:25.140914 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed4a69e43f531cafb475cd93c40be7672b67ede892d91e7c12789669e9c1ccb3"} err="failed to get container status \"ed4a69e43f531cafb475cd93c40be7672b67ede892d91e7c12789669e9c1ccb3\": rpc error: code = NotFound desc = could not find container \"ed4a69e43f531cafb475cd93c40be7672b67ede892d91e7c12789669e9c1ccb3\": container with ID starting with ed4a69e43f531cafb475cd93c40be7672b67ede892d91e7c12789669e9c1ccb3 not found: ID does not exist" Dec 01 14:49:25 crc kubenswrapper[4810]: I1201 14:49:25.140942 4810 scope.go:117] "RemoveContainer" containerID="f832ce0f2b9efeb7bbb97a6637f468b889439fb27ad4979bc4b6aafc49e04680" Dec 01 14:49:25 crc kubenswrapper[4810]: E1201 14:49:25.141334 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f832ce0f2b9efeb7bbb97a6637f468b889439fb27ad4979bc4b6aafc49e04680\": container with ID starting with f832ce0f2b9efeb7bbb97a6637f468b889439fb27ad4979bc4b6aafc49e04680 not found: ID does not exist" containerID="f832ce0f2b9efeb7bbb97a6637f468b889439fb27ad4979bc4b6aafc49e04680" Dec 01 14:49:25 crc kubenswrapper[4810]: I1201 14:49:25.141360 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f832ce0f2b9efeb7bbb97a6637f468b889439fb27ad4979bc4b6aafc49e04680"} err="failed to get container status \"f832ce0f2b9efeb7bbb97a6637f468b889439fb27ad4979bc4b6aafc49e04680\": rpc error: code = NotFound desc = could not find container \"f832ce0f2b9efeb7bbb97a6637f468b889439fb27ad4979bc4b6aafc49e04680\": container with ID starting with f832ce0f2b9efeb7bbb97a6637f468b889439fb27ad4979bc4b6aafc49e04680 not found: ID does not exist" Dec 01 14:49:25 crc kubenswrapper[4810]: I1201 14:49:25.141379 4810 scope.go:117] "RemoveContainer" containerID="4ebe0e45ef0dfc4e919fdea6b705b9d8ba3ab57e0841bd904ef357ccd91f1620" Dec 01 14:49:25 crc kubenswrapper[4810]: E1201 14:49:25.141589 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ebe0e45ef0dfc4e919fdea6b705b9d8ba3ab57e0841bd904ef357ccd91f1620\": container with ID starting with 4ebe0e45ef0dfc4e919fdea6b705b9d8ba3ab57e0841bd904ef357ccd91f1620 not found: ID does not exist" containerID="4ebe0e45ef0dfc4e919fdea6b705b9d8ba3ab57e0841bd904ef357ccd91f1620" Dec 01 14:49:25 crc kubenswrapper[4810]: I1201 14:49:25.141616 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ebe0e45ef0dfc4e919fdea6b705b9d8ba3ab57e0841bd904ef357ccd91f1620"} err="failed to get container status \"4ebe0e45ef0dfc4e919fdea6b705b9d8ba3ab57e0841bd904ef357ccd91f1620\": rpc error: code = NotFound desc = could not find container \"4ebe0e45ef0dfc4e919fdea6b705b9d8ba3ab57e0841bd904ef357ccd91f1620\": container with ID starting with 4ebe0e45ef0dfc4e919fdea6b705b9d8ba3ab57e0841bd904ef357ccd91f1620 not found: ID does not exist" Dec 01 14:49:26 crc kubenswrapper[4810]: I1201 14:49:26.509838 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76135880-43c5-43c3-bb4d-7591e82a0a29" path="/var/lib/kubelet/pods/76135880-43c5-43c3-bb4d-7591e82a0a29/volumes" Dec 01 14:49:30 crc kubenswrapper[4810]: I1201 14:49:30.099330 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt" event={"ID":"55a90415-842f-4a4f-b9ee-7932c3baca33","Type":"ContainerStarted","Data":"5bc5266d694bfe2b710ea4ac58d0887ce45a77f5893a4b9a25d5b267c3daca3b"} Dec 01 14:49:30 crc kubenswrapper[4810]: I1201 14:49:30.101809 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt" Dec 01 14:49:30 crc kubenswrapper[4810]: I1201 14:49:30.121872 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt" podStartSLOduration=1.801108051 podStartE2EDuration="12.121850657s" podCreationTimestamp="2025-12-01 14:49:18 +0000 UTC" firstStartedPulling="2025-12-01 14:49:19.032347788 +0000 UTC m=+924.795857391" lastFinishedPulling="2025-12-01 14:49:29.353090394 +0000 UTC m=+935.116599997" observedRunningTime="2025-12-01 14:49:30.11866654 +0000 UTC m=+935.882176143" watchObservedRunningTime="2025-12-01 14:49:30.121850657 +0000 UTC m=+935.885360260" Dec 01 14:49:48 crc kubenswrapper[4810]: I1201 14:49:48.513327 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7c9798ff88-mt9tt" Dec 01 14:49:58 crc kubenswrapper[4810]: I1201 14:49:58.255039 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-855bb66d65-ldd2w" Dec 01 14:49:58 crc kubenswrapper[4810]: I1201 14:49:58.358338 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qcnz7"] Dec 01 14:49:58 crc kubenswrapper[4810]: E1201 14:49:58.358634 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76135880-43c5-43c3-bb4d-7591e82a0a29" containerName="extract-content" Dec 01 14:49:58 crc kubenswrapper[4810]: I1201 14:49:58.358654 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="76135880-43c5-43c3-bb4d-7591e82a0a29" containerName="extract-content" Dec 01 14:49:58 crc kubenswrapper[4810]: E1201 14:49:58.358674 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76135880-43c5-43c3-bb4d-7591e82a0a29" containerName="extract-utilities" Dec 01 14:49:58 crc kubenswrapper[4810]: I1201 14:49:58.358682 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="76135880-43c5-43c3-bb4d-7591e82a0a29" containerName="extract-utilities" Dec 01 14:49:58 crc kubenswrapper[4810]: E1201 14:49:58.358694 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76135880-43c5-43c3-bb4d-7591e82a0a29" containerName="registry-server" Dec 01 14:49:58 crc kubenswrapper[4810]: I1201 14:49:58.358703 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="76135880-43c5-43c3-bb4d-7591e82a0a29" containerName="registry-server" Dec 01 14:49:58 crc kubenswrapper[4810]: I1201 14:49:58.358835 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="76135880-43c5-43c3-bb4d-7591e82a0a29" containerName="registry-server" Dec 01 14:49:58 crc kubenswrapper[4810]: I1201 14:49:58.359856 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qcnz7" Dec 01 14:49:58 crc kubenswrapper[4810]: I1201 14:49:58.371503 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qcnz7"] Dec 01 14:49:58 crc kubenswrapper[4810]: I1201 14:49:58.513888 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11e4f839-23c0-4b27-a027-b6fc07a9c389-utilities\") pod \"certified-operators-qcnz7\" (UID: \"11e4f839-23c0-4b27-a027-b6fc07a9c389\") " pod="openshift-marketplace/certified-operators-qcnz7" Dec 01 14:49:58 crc kubenswrapper[4810]: I1201 14:49:58.514040 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11e4f839-23c0-4b27-a027-b6fc07a9c389-catalog-content\") pod \"certified-operators-qcnz7\" (UID: \"11e4f839-23c0-4b27-a027-b6fc07a9c389\") " pod="openshift-marketplace/certified-operators-qcnz7" Dec 01 14:49:58 crc kubenswrapper[4810]: I1201 14:49:58.514158 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gbb6\" (UniqueName: \"kubernetes.io/projected/11e4f839-23c0-4b27-a027-b6fc07a9c389-kube-api-access-2gbb6\") pod \"certified-operators-qcnz7\" (UID: \"11e4f839-23c0-4b27-a027-b6fc07a9c389\") " pod="openshift-marketplace/certified-operators-qcnz7" Dec 01 14:49:58 crc kubenswrapper[4810]: I1201 14:49:58.615033 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gbb6\" (UniqueName: \"kubernetes.io/projected/11e4f839-23c0-4b27-a027-b6fc07a9c389-kube-api-access-2gbb6\") pod \"certified-operators-qcnz7\" (UID: \"11e4f839-23c0-4b27-a027-b6fc07a9c389\") " pod="openshift-marketplace/certified-operators-qcnz7" Dec 01 14:49:58 crc kubenswrapper[4810]: I1201 14:49:58.615343 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11e4f839-23c0-4b27-a027-b6fc07a9c389-utilities\") pod \"certified-operators-qcnz7\" (UID: \"11e4f839-23c0-4b27-a027-b6fc07a9c389\") " pod="openshift-marketplace/certified-operators-qcnz7" Dec 01 14:49:58 crc kubenswrapper[4810]: I1201 14:49:58.615460 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11e4f839-23c0-4b27-a027-b6fc07a9c389-catalog-content\") pod \"certified-operators-qcnz7\" (UID: \"11e4f839-23c0-4b27-a027-b6fc07a9c389\") " pod="openshift-marketplace/certified-operators-qcnz7" Dec 01 14:49:58 crc kubenswrapper[4810]: I1201 14:49:58.615915 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11e4f839-23c0-4b27-a027-b6fc07a9c389-catalog-content\") pod \"certified-operators-qcnz7\" (UID: \"11e4f839-23c0-4b27-a027-b6fc07a9c389\") " pod="openshift-marketplace/certified-operators-qcnz7" Dec 01 14:49:58 crc kubenswrapper[4810]: I1201 14:49:58.616171 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11e4f839-23c0-4b27-a027-b6fc07a9c389-utilities\") pod \"certified-operators-qcnz7\" (UID: \"11e4f839-23c0-4b27-a027-b6fc07a9c389\") " pod="openshift-marketplace/certified-operators-qcnz7" Dec 01 14:49:58 crc kubenswrapper[4810]: I1201 14:49:58.634259 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gbb6\" (UniqueName: \"kubernetes.io/projected/11e4f839-23c0-4b27-a027-b6fc07a9c389-kube-api-access-2gbb6\") pod \"certified-operators-qcnz7\" (UID: \"11e4f839-23c0-4b27-a027-b6fc07a9c389\") " pod="openshift-marketplace/certified-operators-qcnz7" Dec 01 14:49:58 crc kubenswrapper[4810]: I1201 14:49:58.681910 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qcnz7" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.001029 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-76dwm"] Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.003104 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.004841 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.012416 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-sth2l"] Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.013066 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sth2l" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.013717 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.013922 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-7bmlv" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.015738 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.022968 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-sth2l"] Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.107268 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-4gcwv"] Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.108283 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-4gcwv" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.110161 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.110360 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.110464 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.110676 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-stcgs" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.122782 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c6232fb4-99db-4844-b364-4f1141abdbb4-frr-startup\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.122818 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c6232fb4-99db-4844-b364-4f1141abdbb4-metrics-certs\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.122842 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c6232fb4-99db-4844-b364-4f1141abdbb4-reloader\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.122859 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d774165e-4724-45f6-994e-979f8b8a5fdd-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-sth2l\" (UID: \"d774165e-4724-45f6-994e-979f8b8a5fdd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sth2l" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.122888 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twzjb\" (UniqueName: \"kubernetes.io/projected/c6232fb4-99db-4844-b364-4f1141abdbb4-kube-api-access-twzjb\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.122904 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c6232fb4-99db-4844-b364-4f1141abdbb4-frr-sockets\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.122967 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c6232fb4-99db-4844-b364-4f1141abdbb4-metrics\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.122981 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8f9x\" (UniqueName: \"kubernetes.io/projected/d774165e-4724-45f6-994e-979f8b8a5fdd-kube-api-access-q8f9x\") pod \"frr-k8s-webhook-server-7fcb986d4-sth2l\" (UID: \"d774165e-4724-45f6-994e-979f8b8a5fdd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sth2l" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.123113 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c6232fb4-99db-4844-b364-4f1141abdbb4-frr-conf\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.140489 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-54tc5"] Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.141431 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-54tc5" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.144141 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.157651 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-54tc5"] Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.235256 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c6232fb4-99db-4844-b364-4f1141abdbb4-frr-conf\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.235297 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49a2d6d2-847f-42bb-995d-54c4009c52ca-cert\") pod \"controller-f8648f98b-54tc5\" (UID: \"49a2d6d2-847f-42bb-995d-54c4009c52ca\") " pod="metallb-system/controller-f8648f98b-54tc5" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.235327 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m596s\" (UniqueName: \"kubernetes.io/projected/d7514ada-83b0-48a3-a03b-07b2f346f607-kube-api-access-m596s\") pod \"speaker-4gcwv\" (UID: \"d7514ada-83b0-48a3-a03b-07b2f346f607\") " pod="metallb-system/speaker-4gcwv" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.235351 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c6232fb4-99db-4844-b364-4f1141abdbb4-frr-startup\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.235372 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c6232fb4-99db-4844-b364-4f1141abdbb4-metrics-certs\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.235387 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d7514ada-83b0-48a3-a03b-07b2f346f607-memberlist\") pod \"speaker-4gcwv\" (UID: \"d7514ada-83b0-48a3-a03b-07b2f346f607\") " pod="metallb-system/speaker-4gcwv" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.235403 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c6232fb4-99db-4844-b364-4f1141abdbb4-reloader\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.235422 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drq8f\" (UniqueName: \"kubernetes.io/projected/49a2d6d2-847f-42bb-995d-54c4009c52ca-kube-api-access-drq8f\") pod \"controller-f8648f98b-54tc5\" (UID: \"49a2d6d2-847f-42bb-995d-54c4009c52ca\") " pod="metallb-system/controller-f8648f98b-54tc5" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.235438 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d774165e-4724-45f6-994e-979f8b8a5fdd-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-sth2l\" (UID: \"d774165e-4724-45f6-994e-979f8b8a5fdd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sth2l" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.235452 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d7514ada-83b0-48a3-a03b-07b2f346f607-metrics-certs\") pod \"speaker-4gcwv\" (UID: \"d7514ada-83b0-48a3-a03b-07b2f346f607\") " pod="metallb-system/speaker-4gcwv" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.239030 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d7514ada-83b0-48a3-a03b-07b2f346f607-metallb-excludel2\") pod \"speaker-4gcwv\" (UID: \"d7514ada-83b0-48a3-a03b-07b2f346f607\") " pod="metallb-system/speaker-4gcwv" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.239102 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twzjb\" (UniqueName: \"kubernetes.io/projected/c6232fb4-99db-4844-b364-4f1141abdbb4-kube-api-access-twzjb\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.239125 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c6232fb4-99db-4844-b364-4f1141abdbb4-frr-sockets\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.239309 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qcnz7"] Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.239682 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c6232fb4-99db-4844-b364-4f1141abdbb4-frr-conf\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: E1201 14:49:59.239780 4810 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 01 14:49:59 crc kubenswrapper[4810]: E1201 14:49:59.239853 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d774165e-4724-45f6-994e-979f8b8a5fdd-cert podName:d774165e-4724-45f6-994e-979f8b8a5fdd nodeName:}" failed. No retries permitted until 2025-12-01 14:49:59.739821769 +0000 UTC m=+965.503331372 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d774165e-4724-45f6-994e-979f8b8a5fdd-cert") pod "frr-k8s-webhook-server-7fcb986d4-sth2l" (UID: "d774165e-4724-45f6-994e-979f8b8a5fdd") : secret "frr-k8s-webhook-server-cert" not found Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.241605 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/49a2d6d2-847f-42bb-995d-54c4009c52ca-metrics-certs\") pod \"controller-f8648f98b-54tc5\" (UID: \"49a2d6d2-847f-42bb-995d-54c4009c52ca\") " pod="metallb-system/controller-f8648f98b-54tc5" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.242137 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c6232fb4-99db-4844-b364-4f1141abdbb4-metrics\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.242176 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8f9x\" (UniqueName: \"kubernetes.io/projected/d774165e-4724-45f6-994e-979f8b8a5fdd-kube-api-access-q8f9x\") pod \"frr-k8s-webhook-server-7fcb986d4-sth2l\" (UID: \"d774165e-4724-45f6-994e-979f8b8a5fdd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sth2l" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.244615 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c6232fb4-99db-4844-b364-4f1141abdbb4-reloader\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.244856 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c6232fb4-99db-4844-b364-4f1141abdbb4-frr-sockets\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.245000 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c6232fb4-99db-4844-b364-4f1141abdbb4-metrics\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.245223 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c6232fb4-99db-4844-b364-4f1141abdbb4-metrics-certs\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.245456 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c6232fb4-99db-4844-b364-4f1141abdbb4-frr-startup\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.269412 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8f9x\" (UniqueName: \"kubernetes.io/projected/d774165e-4724-45f6-994e-979f8b8a5fdd-kube-api-access-q8f9x\") pod \"frr-k8s-webhook-server-7fcb986d4-sth2l\" (UID: \"d774165e-4724-45f6-994e-979f8b8a5fdd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sth2l" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.271964 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twzjb\" (UniqueName: \"kubernetes.io/projected/c6232fb4-99db-4844-b364-4f1141abdbb4-kube-api-access-twzjb\") pod \"frr-k8s-76dwm\" (UID: \"c6232fb4-99db-4844-b364-4f1141abdbb4\") " pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.290728 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qcnz7" event={"ID":"11e4f839-23c0-4b27-a027-b6fc07a9c389","Type":"ContainerStarted","Data":"9bb46f584000856b1492517c1df2b7861c29297f1bcf5a889de482b696d2e3ec"} Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.323784 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-76dwm" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.343057 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d7514ada-83b0-48a3-a03b-07b2f346f607-memberlist\") pod \"speaker-4gcwv\" (UID: \"d7514ada-83b0-48a3-a03b-07b2f346f607\") " pod="metallb-system/speaker-4gcwv" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.343243 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drq8f\" (UniqueName: \"kubernetes.io/projected/49a2d6d2-847f-42bb-995d-54c4009c52ca-kube-api-access-drq8f\") pod \"controller-f8648f98b-54tc5\" (UID: \"49a2d6d2-847f-42bb-995d-54c4009c52ca\") " pod="metallb-system/controller-f8648f98b-54tc5" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.343360 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d7514ada-83b0-48a3-a03b-07b2f346f607-metrics-certs\") pod \"speaker-4gcwv\" (UID: \"d7514ada-83b0-48a3-a03b-07b2f346f607\") " pod="metallb-system/speaker-4gcwv" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.343488 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d7514ada-83b0-48a3-a03b-07b2f346f607-metallb-excludel2\") pod \"speaker-4gcwv\" (UID: \"d7514ada-83b0-48a3-a03b-07b2f346f607\") " pod="metallb-system/speaker-4gcwv" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.343576 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/49a2d6d2-847f-42bb-995d-54c4009c52ca-metrics-certs\") pod \"controller-f8648f98b-54tc5\" (UID: \"49a2d6d2-847f-42bb-995d-54c4009c52ca\") " pod="metallb-system/controller-f8648f98b-54tc5" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.343714 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49a2d6d2-847f-42bb-995d-54c4009c52ca-cert\") pod \"controller-f8648f98b-54tc5\" (UID: \"49a2d6d2-847f-42bb-995d-54c4009c52ca\") " pod="metallb-system/controller-f8648f98b-54tc5" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.343796 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m596s\" (UniqueName: \"kubernetes.io/projected/d7514ada-83b0-48a3-a03b-07b2f346f607-kube-api-access-m596s\") pod \"speaker-4gcwv\" (UID: \"d7514ada-83b0-48a3-a03b-07b2f346f607\") " pod="metallb-system/speaker-4gcwv" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.344611 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d7514ada-83b0-48a3-a03b-07b2f346f607-metallb-excludel2\") pod \"speaker-4gcwv\" (UID: \"d7514ada-83b0-48a3-a03b-07b2f346f607\") " pod="metallb-system/speaker-4gcwv" Dec 01 14:49:59 crc kubenswrapper[4810]: E1201 14:49:59.344924 4810 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 01 14:49:59 crc kubenswrapper[4810]: E1201 14:49:59.344967 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d7514ada-83b0-48a3-a03b-07b2f346f607-memberlist podName:d7514ada-83b0-48a3-a03b-07b2f346f607 nodeName:}" failed. No retries permitted until 2025-12-01 14:49:59.844952822 +0000 UTC m=+965.608462425 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d7514ada-83b0-48a3-a03b-07b2f346f607-memberlist") pod "speaker-4gcwv" (UID: "d7514ada-83b0-48a3-a03b-07b2f346f607") : secret "metallb-memberlist" not found Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.349160 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/49a2d6d2-847f-42bb-995d-54c4009c52ca-metrics-certs\") pod \"controller-f8648f98b-54tc5\" (UID: \"49a2d6d2-847f-42bb-995d-54c4009c52ca\") " pod="metallb-system/controller-f8648f98b-54tc5" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.350841 4810 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.351048 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d7514ada-83b0-48a3-a03b-07b2f346f607-metrics-certs\") pod \"speaker-4gcwv\" (UID: \"d7514ada-83b0-48a3-a03b-07b2f346f607\") " pod="metallb-system/speaker-4gcwv" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.360845 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49a2d6d2-847f-42bb-995d-54c4009c52ca-cert\") pod \"controller-f8648f98b-54tc5\" (UID: \"49a2d6d2-847f-42bb-995d-54c4009c52ca\") " pod="metallb-system/controller-f8648f98b-54tc5" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.367942 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m596s\" (UniqueName: \"kubernetes.io/projected/d7514ada-83b0-48a3-a03b-07b2f346f607-kube-api-access-m596s\") pod \"speaker-4gcwv\" (UID: \"d7514ada-83b0-48a3-a03b-07b2f346f607\") " pod="metallb-system/speaker-4gcwv" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.371127 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drq8f\" (UniqueName: \"kubernetes.io/projected/49a2d6d2-847f-42bb-995d-54c4009c52ca-kube-api-access-drq8f\") pod \"controller-f8648f98b-54tc5\" (UID: \"49a2d6d2-847f-42bb-995d-54c4009c52ca\") " pod="metallb-system/controller-f8648f98b-54tc5" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.466368 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-54tc5" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.696044 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-54tc5"] Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.749877 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d774165e-4724-45f6-994e-979f8b8a5fdd-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-sth2l\" (UID: \"d774165e-4724-45f6-994e-979f8b8a5fdd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sth2l" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.755305 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d774165e-4724-45f6-994e-979f8b8a5fdd-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-sth2l\" (UID: \"d774165e-4724-45f6-994e-979f8b8a5fdd\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sth2l" Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.850604 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d7514ada-83b0-48a3-a03b-07b2f346f607-memberlist\") pod \"speaker-4gcwv\" (UID: \"d7514ada-83b0-48a3-a03b-07b2f346f607\") " pod="metallb-system/speaker-4gcwv" Dec 01 14:49:59 crc kubenswrapper[4810]: E1201 14:49:59.851019 4810 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 01 14:49:59 crc kubenswrapper[4810]: E1201 14:49:59.851097 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d7514ada-83b0-48a3-a03b-07b2f346f607-memberlist podName:d7514ada-83b0-48a3-a03b-07b2f346f607 nodeName:}" failed. No retries permitted until 2025-12-01 14:50:00.851078656 +0000 UTC m=+966.614588259 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d7514ada-83b0-48a3-a03b-07b2f346f607-memberlist") pod "speaker-4gcwv" (UID: "d7514ada-83b0-48a3-a03b-07b2f346f607") : secret "metallb-memberlist" not found Dec 01 14:49:59 crc kubenswrapper[4810]: I1201 14:49:59.931174 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sth2l" Dec 01 14:50:00 crc kubenswrapper[4810]: I1201 14:50:00.297807 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-54tc5" event={"ID":"49a2d6d2-847f-42bb-995d-54c4009c52ca","Type":"ContainerStarted","Data":"a90725d515b71f618bf9a87c364dd874fa77cc0c6bbcda32aa472d0fa1a15c25"} Dec 01 14:50:00 crc kubenswrapper[4810]: I1201 14:50:00.298171 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-54tc5" Dec 01 14:50:00 crc kubenswrapper[4810]: I1201 14:50:00.298186 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-54tc5" event={"ID":"49a2d6d2-847f-42bb-995d-54c4009c52ca","Type":"ContainerStarted","Data":"d5ebe483643943a0808e0f49b113b2fae3edbcc713f9abfde8fa35466cbaf24d"} Dec 01 14:50:00 crc kubenswrapper[4810]: I1201 14:50:00.298200 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-54tc5" event={"ID":"49a2d6d2-847f-42bb-995d-54c4009c52ca","Type":"ContainerStarted","Data":"359bac2cb78964974d70cfb396a69aa0393f5cd88b9b1600153338c419db275c"} Dec 01 14:50:00 crc kubenswrapper[4810]: I1201 14:50:00.299493 4810 generic.go:334] "Generic (PLEG): container finished" podID="11e4f839-23c0-4b27-a027-b6fc07a9c389" containerID="ba14f22a4eae8e8a041615c6e905cbcf1308f332f70bf3c1403b1f46e8296c0c" exitCode=0 Dec 01 14:50:00 crc kubenswrapper[4810]: I1201 14:50:00.299579 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qcnz7" event={"ID":"11e4f839-23c0-4b27-a027-b6fc07a9c389","Type":"ContainerDied","Data":"ba14f22a4eae8e8a041615c6e905cbcf1308f332f70bf3c1403b1f46e8296c0c"} Dec 01 14:50:00 crc kubenswrapper[4810]: I1201 14:50:00.301225 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-76dwm" event={"ID":"c6232fb4-99db-4844-b364-4f1141abdbb4","Type":"ContainerStarted","Data":"cc18531184ce2822f610efc9fdb3812a16407bdee27d0ca9b786d261e9ebf75e"} Dec 01 14:50:00 crc kubenswrapper[4810]: I1201 14:50:00.320917 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-54tc5" podStartSLOduration=1.320899158 podStartE2EDuration="1.320899158s" podCreationTimestamp="2025-12-01 14:49:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:50:00.317764602 +0000 UTC m=+966.081274215" watchObservedRunningTime="2025-12-01 14:50:00.320899158 +0000 UTC m=+966.084408761" Dec 01 14:50:00 crc kubenswrapper[4810]: I1201 14:50:00.367227 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-sth2l"] Dec 01 14:50:00 crc kubenswrapper[4810]: W1201 14:50:00.380232 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd774165e_4724_45f6_994e_979f8b8a5fdd.slice/crio-3cc657cae53b392dca6b752b4c3f0da933227fb4df21ee6f1728f3dab6f6b0e7 WatchSource:0}: Error finding container 3cc657cae53b392dca6b752b4c3f0da933227fb4df21ee6f1728f3dab6f6b0e7: Status 404 returned error can't find the container with id 3cc657cae53b392dca6b752b4c3f0da933227fb4df21ee6f1728f3dab6f6b0e7 Dec 01 14:50:00 crc kubenswrapper[4810]: I1201 14:50:00.864788 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d7514ada-83b0-48a3-a03b-07b2f346f607-memberlist\") pod \"speaker-4gcwv\" (UID: \"d7514ada-83b0-48a3-a03b-07b2f346f607\") " pod="metallb-system/speaker-4gcwv" Dec 01 14:50:00 crc kubenswrapper[4810]: I1201 14:50:00.872732 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d7514ada-83b0-48a3-a03b-07b2f346f607-memberlist\") pod \"speaker-4gcwv\" (UID: \"d7514ada-83b0-48a3-a03b-07b2f346f607\") " pod="metallb-system/speaker-4gcwv" Dec 01 14:50:00 crc kubenswrapper[4810]: I1201 14:50:00.920463 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-4gcwv" Dec 01 14:50:00 crc kubenswrapper[4810]: W1201 14:50:00.948269 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7514ada_83b0_48a3_a03b_07b2f346f607.slice/crio-8a696b291704c80df8691cc1befb8083c8247536daaa8d1d0e8f4c18594ba522 WatchSource:0}: Error finding container 8a696b291704c80df8691cc1befb8083c8247536daaa8d1d0e8f4c18594ba522: Status 404 returned error can't find the container with id 8a696b291704c80df8691cc1befb8083c8247536daaa8d1d0e8f4c18594ba522 Dec 01 14:50:01 crc kubenswrapper[4810]: I1201 14:50:01.311350 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-4gcwv" event={"ID":"d7514ada-83b0-48a3-a03b-07b2f346f607","Type":"ContainerStarted","Data":"36d5a472039c6982bbcb35abc1421c66aab7e17289e859aac94e374d9800e543"} Dec 01 14:50:01 crc kubenswrapper[4810]: I1201 14:50:01.311721 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-4gcwv" event={"ID":"d7514ada-83b0-48a3-a03b-07b2f346f607","Type":"ContainerStarted","Data":"8a696b291704c80df8691cc1befb8083c8247536daaa8d1d0e8f4c18594ba522"} Dec 01 14:50:01 crc kubenswrapper[4810]: I1201 14:50:01.313654 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qcnz7" event={"ID":"11e4f839-23c0-4b27-a027-b6fc07a9c389","Type":"ContainerStarted","Data":"c5adf68c50cad6f9acc4c3556b244cb56c55032cab76cac807404820236b68d6"} Dec 01 14:50:01 crc kubenswrapper[4810]: I1201 14:50:01.315991 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sth2l" event={"ID":"d774165e-4724-45f6-994e-979f8b8a5fdd","Type":"ContainerStarted","Data":"3cc657cae53b392dca6b752b4c3f0da933227fb4df21ee6f1728f3dab6f6b0e7"} Dec 01 14:50:02 crc kubenswrapper[4810]: I1201 14:50:02.342608 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-4gcwv" event={"ID":"d7514ada-83b0-48a3-a03b-07b2f346f607","Type":"ContainerStarted","Data":"28a0ff03fe18565c9157f88124870dfb6014b90d4ad381e265e6f89cbb5778b5"} Dec 01 14:50:02 crc kubenswrapper[4810]: I1201 14:50:02.343416 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-4gcwv" Dec 01 14:50:02 crc kubenswrapper[4810]: I1201 14:50:02.345542 4810 generic.go:334] "Generic (PLEG): container finished" podID="11e4f839-23c0-4b27-a027-b6fc07a9c389" containerID="c5adf68c50cad6f9acc4c3556b244cb56c55032cab76cac807404820236b68d6" exitCode=0 Dec 01 14:50:02 crc kubenswrapper[4810]: I1201 14:50:02.345578 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qcnz7" event={"ID":"11e4f839-23c0-4b27-a027-b6fc07a9c389","Type":"ContainerDied","Data":"c5adf68c50cad6f9acc4c3556b244cb56c55032cab76cac807404820236b68d6"} Dec 01 14:50:02 crc kubenswrapper[4810]: I1201 14:50:02.447677 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-4gcwv" podStartSLOduration=3.447658609 podStartE2EDuration="3.447658609s" podCreationTimestamp="2025-12-01 14:49:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:50:02.443032942 +0000 UTC m=+968.206542545" watchObservedRunningTime="2025-12-01 14:50:02.447658609 +0000 UTC m=+968.211168212" Dec 01 14:50:03 crc kubenswrapper[4810]: I1201 14:50:03.358626 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qcnz7" event={"ID":"11e4f839-23c0-4b27-a027-b6fc07a9c389","Type":"ContainerStarted","Data":"989781a3262a4e2de058366a3f1db08476e4783f360c10712e6a38d541b62b81"} Dec 01 14:50:03 crc kubenswrapper[4810]: I1201 14:50:03.377280 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qcnz7" podStartSLOduration=2.746083129 podStartE2EDuration="5.377262638s" podCreationTimestamp="2025-12-01 14:49:58 +0000 UTC" firstStartedPulling="2025-12-01 14:50:00.302066903 +0000 UTC m=+966.065576526" lastFinishedPulling="2025-12-01 14:50:02.933246432 +0000 UTC m=+968.696756035" observedRunningTime="2025-12-01 14:50:03.375032166 +0000 UTC m=+969.138541789" watchObservedRunningTime="2025-12-01 14:50:03.377262638 +0000 UTC m=+969.140772241" Dec 01 14:50:08 crc kubenswrapper[4810]: I1201 14:50:08.683017 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qcnz7" Dec 01 14:50:08 crc kubenswrapper[4810]: I1201 14:50:08.683618 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qcnz7" Dec 01 14:50:08 crc kubenswrapper[4810]: I1201 14:50:08.756529 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qcnz7" Dec 01 14:50:09 crc kubenswrapper[4810]: I1201 14:50:09.451688 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qcnz7" Dec 01 14:50:09 crc kubenswrapper[4810]: I1201 14:50:09.473308 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-54tc5" Dec 01 14:50:09 crc kubenswrapper[4810]: I1201 14:50:09.507826 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qcnz7"] Dec 01 14:50:10 crc kubenswrapper[4810]: I1201 14:50:10.412860 4810 generic.go:334] "Generic (PLEG): container finished" podID="c6232fb4-99db-4844-b364-4f1141abdbb4" containerID="e41727a5e35eb03cfca7347ba05a8bb53faba6c515c173207ccbb014bfc50eb6" exitCode=0 Dec 01 14:50:10 crc kubenswrapper[4810]: I1201 14:50:10.412960 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-76dwm" event={"ID":"c6232fb4-99db-4844-b364-4f1141abdbb4","Type":"ContainerDied","Data":"e41727a5e35eb03cfca7347ba05a8bb53faba6c515c173207ccbb014bfc50eb6"} Dec 01 14:50:10 crc kubenswrapper[4810]: I1201 14:50:10.417578 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sth2l" event={"ID":"d774165e-4724-45f6-994e-979f8b8a5fdd","Type":"ContainerStarted","Data":"6f80804c546c5bfb6fa8a249f978e90bf7de9fa388257d73505f1a2b371003d5"} Dec 01 14:50:10 crc kubenswrapper[4810]: I1201 14:50:10.454587 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sth2l" podStartSLOduration=2.989325088 podStartE2EDuration="12.454556241s" podCreationTimestamp="2025-12-01 14:49:58 +0000 UTC" firstStartedPulling="2025-12-01 14:50:00.382167413 +0000 UTC m=+966.145677016" lastFinishedPulling="2025-12-01 14:50:09.847398566 +0000 UTC m=+975.610908169" observedRunningTime="2025-12-01 14:50:10.453121622 +0000 UTC m=+976.216631225" watchObservedRunningTime="2025-12-01 14:50:10.454556241 +0000 UTC m=+976.218065874" Dec 01 14:50:11 crc kubenswrapper[4810]: I1201 14:50:11.425354 4810 generic.go:334] "Generic (PLEG): container finished" podID="c6232fb4-99db-4844-b364-4f1141abdbb4" containerID="f5fb7759b257c2341824c11615471e180f7b890c1b9d08fb1c5c5f427ad1e34a" exitCode=0 Dec 01 14:50:11 crc kubenswrapper[4810]: I1201 14:50:11.425448 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-76dwm" event={"ID":"c6232fb4-99db-4844-b364-4f1141abdbb4","Type":"ContainerDied","Data":"f5fb7759b257c2341824c11615471e180f7b890c1b9d08fb1c5c5f427ad1e34a"} Dec 01 14:50:11 crc kubenswrapper[4810]: I1201 14:50:11.425613 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qcnz7" podUID="11e4f839-23c0-4b27-a027-b6fc07a9c389" containerName="registry-server" containerID="cri-o://989781a3262a4e2de058366a3f1db08476e4783f360c10712e6a38d541b62b81" gracePeriod=2 Dec 01 14:50:11 crc kubenswrapper[4810]: I1201 14:50:11.425799 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sth2l" Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.276244 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qcnz7" Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.433031 4810 generic.go:334] "Generic (PLEG): container finished" podID="11e4f839-23c0-4b27-a027-b6fc07a9c389" containerID="989781a3262a4e2de058366a3f1db08476e4783f360c10712e6a38d541b62b81" exitCode=0 Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.433065 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qcnz7" event={"ID":"11e4f839-23c0-4b27-a027-b6fc07a9c389","Type":"ContainerDied","Data":"989781a3262a4e2de058366a3f1db08476e4783f360c10712e6a38d541b62b81"} Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.433095 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qcnz7" Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.433104 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qcnz7" event={"ID":"11e4f839-23c0-4b27-a027-b6fc07a9c389","Type":"ContainerDied","Data":"9bb46f584000856b1492517c1df2b7861c29297f1bcf5a889de482b696d2e3ec"} Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.433131 4810 scope.go:117] "RemoveContainer" containerID="989781a3262a4e2de058366a3f1db08476e4783f360c10712e6a38d541b62b81" Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.436088 4810 generic.go:334] "Generic (PLEG): container finished" podID="c6232fb4-99db-4844-b364-4f1141abdbb4" containerID="db600890a5293e2ccb5b6b67ca47c298a2b970218201d0279a3287c080d19aa2" exitCode=0 Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.436116 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-76dwm" event={"ID":"c6232fb4-99db-4844-b364-4f1141abdbb4","Type":"ContainerDied","Data":"db600890a5293e2ccb5b6b67ca47c298a2b970218201d0279a3287c080d19aa2"} Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.449298 4810 scope.go:117] "RemoveContainer" containerID="c5adf68c50cad6f9acc4c3556b244cb56c55032cab76cac807404820236b68d6" Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.468745 4810 scope.go:117] "RemoveContainer" containerID="ba14f22a4eae8e8a041615c6e905cbcf1308f332f70bf3c1403b1f46e8296c0c" Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.473916 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11e4f839-23c0-4b27-a027-b6fc07a9c389-catalog-content\") pod \"11e4f839-23c0-4b27-a027-b6fc07a9c389\" (UID: \"11e4f839-23c0-4b27-a027-b6fc07a9c389\") " Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.473994 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11e4f839-23c0-4b27-a027-b6fc07a9c389-utilities\") pod \"11e4f839-23c0-4b27-a027-b6fc07a9c389\" (UID: \"11e4f839-23c0-4b27-a027-b6fc07a9c389\") " Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.474068 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gbb6\" (UniqueName: \"kubernetes.io/projected/11e4f839-23c0-4b27-a027-b6fc07a9c389-kube-api-access-2gbb6\") pod \"11e4f839-23c0-4b27-a027-b6fc07a9c389\" (UID: \"11e4f839-23c0-4b27-a027-b6fc07a9c389\") " Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.474877 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11e4f839-23c0-4b27-a027-b6fc07a9c389-utilities" (OuterVolumeSpecName: "utilities") pod "11e4f839-23c0-4b27-a027-b6fc07a9c389" (UID: "11e4f839-23c0-4b27-a027-b6fc07a9c389"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.482130 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11e4f839-23c0-4b27-a027-b6fc07a9c389-kube-api-access-2gbb6" (OuterVolumeSpecName: "kube-api-access-2gbb6") pod "11e4f839-23c0-4b27-a027-b6fc07a9c389" (UID: "11e4f839-23c0-4b27-a027-b6fc07a9c389"). InnerVolumeSpecName "kube-api-access-2gbb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.492604 4810 scope.go:117] "RemoveContainer" containerID="989781a3262a4e2de058366a3f1db08476e4783f360c10712e6a38d541b62b81" Dec 01 14:50:12 crc kubenswrapper[4810]: E1201 14:50:12.493283 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"989781a3262a4e2de058366a3f1db08476e4783f360c10712e6a38d541b62b81\": container with ID starting with 989781a3262a4e2de058366a3f1db08476e4783f360c10712e6a38d541b62b81 not found: ID does not exist" containerID="989781a3262a4e2de058366a3f1db08476e4783f360c10712e6a38d541b62b81" Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.493365 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"989781a3262a4e2de058366a3f1db08476e4783f360c10712e6a38d541b62b81"} err="failed to get container status \"989781a3262a4e2de058366a3f1db08476e4783f360c10712e6a38d541b62b81\": rpc error: code = NotFound desc = could not find container \"989781a3262a4e2de058366a3f1db08476e4783f360c10712e6a38d541b62b81\": container with ID starting with 989781a3262a4e2de058366a3f1db08476e4783f360c10712e6a38d541b62b81 not found: ID does not exist" Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.493422 4810 scope.go:117] "RemoveContainer" containerID="c5adf68c50cad6f9acc4c3556b244cb56c55032cab76cac807404820236b68d6" Dec 01 14:50:12 crc kubenswrapper[4810]: E1201 14:50:12.493971 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5adf68c50cad6f9acc4c3556b244cb56c55032cab76cac807404820236b68d6\": container with ID starting with c5adf68c50cad6f9acc4c3556b244cb56c55032cab76cac807404820236b68d6 not found: ID does not exist" containerID="c5adf68c50cad6f9acc4c3556b244cb56c55032cab76cac807404820236b68d6" Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.494018 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5adf68c50cad6f9acc4c3556b244cb56c55032cab76cac807404820236b68d6"} err="failed to get container status \"c5adf68c50cad6f9acc4c3556b244cb56c55032cab76cac807404820236b68d6\": rpc error: code = NotFound desc = could not find container \"c5adf68c50cad6f9acc4c3556b244cb56c55032cab76cac807404820236b68d6\": container with ID starting with c5adf68c50cad6f9acc4c3556b244cb56c55032cab76cac807404820236b68d6 not found: ID does not exist" Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.494047 4810 scope.go:117] "RemoveContainer" containerID="ba14f22a4eae8e8a041615c6e905cbcf1308f332f70bf3c1403b1f46e8296c0c" Dec 01 14:50:12 crc kubenswrapper[4810]: E1201 14:50:12.494534 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba14f22a4eae8e8a041615c6e905cbcf1308f332f70bf3c1403b1f46e8296c0c\": container with ID starting with ba14f22a4eae8e8a041615c6e905cbcf1308f332f70bf3c1403b1f46e8296c0c not found: ID does not exist" containerID="ba14f22a4eae8e8a041615c6e905cbcf1308f332f70bf3c1403b1f46e8296c0c" Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.494597 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba14f22a4eae8e8a041615c6e905cbcf1308f332f70bf3c1403b1f46e8296c0c"} err="failed to get container status \"ba14f22a4eae8e8a041615c6e905cbcf1308f332f70bf3c1403b1f46e8296c0c\": rpc error: code = NotFound desc = could not find container \"ba14f22a4eae8e8a041615c6e905cbcf1308f332f70bf3c1403b1f46e8296c0c\": container with ID starting with ba14f22a4eae8e8a041615c6e905cbcf1308f332f70bf3c1403b1f46e8296c0c not found: ID does not exist" Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.535127 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11e4f839-23c0-4b27-a027-b6fc07a9c389-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "11e4f839-23c0-4b27-a027-b6fc07a9c389" (UID: "11e4f839-23c0-4b27-a027-b6fc07a9c389"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.575163 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11e4f839-23c0-4b27-a027-b6fc07a9c389-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.575191 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11e4f839-23c0-4b27-a027-b6fc07a9c389-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.575202 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gbb6\" (UniqueName: \"kubernetes.io/projected/11e4f839-23c0-4b27-a027-b6fc07a9c389-kube-api-access-2gbb6\") on node \"crc\" DevicePath \"\"" Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.767434 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qcnz7"] Dec 01 14:50:12 crc kubenswrapper[4810]: I1201 14:50:12.771740 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qcnz7"] Dec 01 14:50:13 crc kubenswrapper[4810]: I1201 14:50:13.446734 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-76dwm" event={"ID":"c6232fb4-99db-4844-b364-4f1141abdbb4","Type":"ContainerStarted","Data":"147b571a4a3f9c3edbe7c70d84950c69b2e6b84b2cc722a3cb12a0d778cd5f5a"} Dec 01 14:50:13 crc kubenswrapper[4810]: I1201 14:50:13.447049 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-76dwm" event={"ID":"c6232fb4-99db-4844-b364-4f1141abdbb4","Type":"ContainerStarted","Data":"1e6426c24e05b75d6b166921ad033645b43700535bacd279356a20c0ccb3e747"} Dec 01 14:50:13 crc kubenswrapper[4810]: I1201 14:50:13.447063 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-76dwm" event={"ID":"c6232fb4-99db-4844-b364-4f1141abdbb4","Type":"ContainerStarted","Data":"c31b42e9091b974d4447d24db54f9f0692c9bd977c8e639e74320809b65df74b"} Dec 01 14:50:13 crc kubenswrapper[4810]: I1201 14:50:13.447074 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-76dwm" event={"ID":"c6232fb4-99db-4844-b364-4f1141abdbb4","Type":"ContainerStarted","Data":"841ba2f3a6393e234c749e129f088d1deb1b4746c2b9b5dfd819e497465045ea"} Dec 01 14:50:13 crc kubenswrapper[4810]: I1201 14:50:13.447085 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-76dwm" event={"ID":"c6232fb4-99db-4844-b364-4f1141abdbb4","Type":"ContainerStarted","Data":"59de11bc09538625df1ca1eca3735a3bd6b2b43a1e311614283c248291ac524a"} Dec 01 14:50:14 crc kubenswrapper[4810]: I1201 14:50:14.458331 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-76dwm" event={"ID":"c6232fb4-99db-4844-b364-4f1141abdbb4","Type":"ContainerStarted","Data":"5d8bf420b967c744ff1a1713f3c897e357be4d836345c6db5007cd51a2f7d3c8"} Dec 01 14:50:14 crc kubenswrapper[4810]: I1201 14:50:14.458639 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-76dwm" Dec 01 14:50:14 crc kubenswrapper[4810]: I1201 14:50:14.480715 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-76dwm" podStartSLOduration=6.304060452 podStartE2EDuration="16.48069817s" podCreationTimestamp="2025-12-01 14:49:58 +0000 UTC" firstStartedPulling="2025-12-01 14:49:59.651589704 +0000 UTC m=+965.415099307" lastFinishedPulling="2025-12-01 14:50:09.828227422 +0000 UTC m=+975.591737025" observedRunningTime="2025-12-01 14:50:14.477285186 +0000 UTC m=+980.240794789" watchObservedRunningTime="2025-12-01 14:50:14.48069817 +0000 UTC m=+980.244207773" Dec 01 14:50:14 crc kubenswrapper[4810]: I1201 14:50:14.498041 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11e4f839-23c0-4b27-a027-b6fc07a9c389" path="/var/lib/kubelet/pods/11e4f839-23c0-4b27-a027-b6fc07a9c389/volumes" Dec 01 14:50:19 crc kubenswrapper[4810]: I1201 14:50:19.323987 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-76dwm" Dec 01 14:50:19 crc kubenswrapper[4810]: I1201 14:50:19.360867 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-76dwm" Dec 01 14:50:20 crc kubenswrapper[4810]: I1201 14:50:20.925076 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-4gcwv" Dec 01 14:50:22 crc kubenswrapper[4810]: I1201 14:50:22.305097 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq"] Dec 01 14:50:22 crc kubenswrapper[4810]: E1201 14:50:22.306365 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11e4f839-23c0-4b27-a027-b6fc07a9c389" containerName="registry-server" Dec 01 14:50:22 crc kubenswrapper[4810]: I1201 14:50:22.306502 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="11e4f839-23c0-4b27-a027-b6fc07a9c389" containerName="registry-server" Dec 01 14:50:22 crc kubenswrapper[4810]: E1201 14:50:22.306623 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11e4f839-23c0-4b27-a027-b6fc07a9c389" containerName="extract-content" Dec 01 14:50:22 crc kubenswrapper[4810]: I1201 14:50:22.306724 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="11e4f839-23c0-4b27-a027-b6fc07a9c389" containerName="extract-content" Dec 01 14:50:22 crc kubenswrapper[4810]: E1201 14:50:22.306821 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11e4f839-23c0-4b27-a027-b6fc07a9c389" containerName="extract-utilities" Dec 01 14:50:22 crc kubenswrapper[4810]: I1201 14:50:22.306901 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="11e4f839-23c0-4b27-a027-b6fc07a9c389" containerName="extract-utilities" Dec 01 14:50:22 crc kubenswrapper[4810]: I1201 14:50:22.307189 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="11e4f839-23c0-4b27-a027-b6fc07a9c389" containerName="registry-server" Dec 01 14:50:22 crc kubenswrapper[4810]: I1201 14:50:22.308397 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq" Dec 01 14:50:22 crc kubenswrapper[4810]: I1201 14:50:22.310189 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 14:50:22 crc kubenswrapper[4810]: I1201 14:50:22.320507 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq"] Dec 01 14:50:22 crc kubenswrapper[4810]: I1201 14:50:22.500860 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2552822d-ff62-49fa-8ba0-321952bda208-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq\" (UID: \"2552822d-ff62-49fa-8ba0-321952bda208\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq" Dec 01 14:50:22 crc kubenswrapper[4810]: I1201 14:50:22.500946 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrjsc\" (UniqueName: \"kubernetes.io/projected/2552822d-ff62-49fa-8ba0-321952bda208-kube-api-access-vrjsc\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq\" (UID: \"2552822d-ff62-49fa-8ba0-321952bda208\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq" Dec 01 14:50:22 crc kubenswrapper[4810]: I1201 14:50:22.500989 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2552822d-ff62-49fa-8ba0-321952bda208-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq\" (UID: \"2552822d-ff62-49fa-8ba0-321952bda208\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq" Dec 01 14:50:22 crc kubenswrapper[4810]: I1201 14:50:22.602996 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2552822d-ff62-49fa-8ba0-321952bda208-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq\" (UID: \"2552822d-ff62-49fa-8ba0-321952bda208\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq" Dec 01 14:50:22 crc kubenswrapper[4810]: I1201 14:50:22.603135 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrjsc\" (UniqueName: \"kubernetes.io/projected/2552822d-ff62-49fa-8ba0-321952bda208-kube-api-access-vrjsc\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq\" (UID: \"2552822d-ff62-49fa-8ba0-321952bda208\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq" Dec 01 14:50:22 crc kubenswrapper[4810]: I1201 14:50:22.603180 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2552822d-ff62-49fa-8ba0-321952bda208-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq\" (UID: \"2552822d-ff62-49fa-8ba0-321952bda208\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq" Dec 01 14:50:22 crc kubenswrapper[4810]: I1201 14:50:22.604072 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2552822d-ff62-49fa-8ba0-321952bda208-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq\" (UID: \"2552822d-ff62-49fa-8ba0-321952bda208\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq" Dec 01 14:50:22 crc kubenswrapper[4810]: I1201 14:50:22.604147 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2552822d-ff62-49fa-8ba0-321952bda208-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq\" (UID: \"2552822d-ff62-49fa-8ba0-321952bda208\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq" Dec 01 14:50:22 crc kubenswrapper[4810]: I1201 14:50:22.631776 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrjsc\" (UniqueName: \"kubernetes.io/projected/2552822d-ff62-49fa-8ba0-321952bda208-kube-api-access-vrjsc\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq\" (UID: \"2552822d-ff62-49fa-8ba0-321952bda208\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq" Dec 01 14:50:22 crc kubenswrapper[4810]: I1201 14:50:22.932261 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq" Dec 01 14:50:23 crc kubenswrapper[4810]: I1201 14:50:23.413828 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq"] Dec 01 14:50:23 crc kubenswrapper[4810]: W1201 14:50:23.420586 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2552822d_ff62_49fa_8ba0_321952bda208.slice/crio-2694d490f68bec3dc0263f271363683e7c966ab6465a4055c58a33f055a00161 WatchSource:0}: Error finding container 2694d490f68bec3dc0263f271363683e7c966ab6465a4055c58a33f055a00161: Status 404 returned error can't find the container with id 2694d490f68bec3dc0263f271363683e7c966ab6465a4055c58a33f055a00161 Dec 01 14:50:23 crc kubenswrapper[4810]: I1201 14:50:23.528581 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq" event={"ID":"2552822d-ff62-49fa-8ba0-321952bda208","Type":"ContainerStarted","Data":"2694d490f68bec3dc0263f271363683e7c966ab6465a4055c58a33f055a00161"} Dec 01 14:50:24 crc kubenswrapper[4810]: I1201 14:50:24.540507 4810 generic.go:334] "Generic (PLEG): container finished" podID="2552822d-ff62-49fa-8ba0-321952bda208" containerID="bbeaa57ab29c4172120ea774037279abc46e699d2ba291e03465141f9c619dbb" exitCode=0 Dec 01 14:50:24 crc kubenswrapper[4810]: I1201 14:50:24.540645 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq" event={"ID":"2552822d-ff62-49fa-8ba0-321952bda208","Type":"ContainerDied","Data":"bbeaa57ab29c4172120ea774037279abc46e699d2ba291e03465141f9c619dbb"} Dec 01 14:50:28 crc kubenswrapper[4810]: I1201 14:50:28.564248 4810 generic.go:334] "Generic (PLEG): container finished" podID="2552822d-ff62-49fa-8ba0-321952bda208" containerID="ad4b9e7c530c0fc89f2866d52d9fa64ff5495363009dc30f0ed5db50bcab09fe" exitCode=0 Dec 01 14:50:28 crc kubenswrapper[4810]: I1201 14:50:28.564367 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq" event={"ID":"2552822d-ff62-49fa-8ba0-321952bda208","Type":"ContainerDied","Data":"ad4b9e7c530c0fc89f2866d52d9fa64ff5495363009dc30f0ed5db50bcab09fe"} Dec 01 14:50:29 crc kubenswrapper[4810]: I1201 14:50:29.327810 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-76dwm" Dec 01 14:50:29 crc kubenswrapper[4810]: I1201 14:50:29.572511 4810 generic.go:334] "Generic (PLEG): container finished" podID="2552822d-ff62-49fa-8ba0-321952bda208" containerID="fb562a93d60645a80d46b42f73b41db918410fc4828a8bdf3b3ba3bdda1ef753" exitCode=0 Dec 01 14:50:29 crc kubenswrapper[4810]: I1201 14:50:29.572565 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq" event={"ID":"2552822d-ff62-49fa-8ba0-321952bda208","Type":"ContainerDied","Data":"fb562a93d60645a80d46b42f73b41db918410fc4828a8bdf3b3ba3bdda1ef753"} Dec 01 14:50:29 crc kubenswrapper[4810]: I1201 14:50:29.937620 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sth2l" Dec 01 14:50:30 crc kubenswrapper[4810]: I1201 14:50:30.824528 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq" Dec 01 14:50:30 crc kubenswrapper[4810]: I1201 14:50:30.842633 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2552822d-ff62-49fa-8ba0-321952bda208-util\") pod \"2552822d-ff62-49fa-8ba0-321952bda208\" (UID: \"2552822d-ff62-49fa-8ba0-321952bda208\") " Dec 01 14:50:30 crc kubenswrapper[4810]: I1201 14:50:30.843077 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrjsc\" (UniqueName: \"kubernetes.io/projected/2552822d-ff62-49fa-8ba0-321952bda208-kube-api-access-vrjsc\") pod \"2552822d-ff62-49fa-8ba0-321952bda208\" (UID: \"2552822d-ff62-49fa-8ba0-321952bda208\") " Dec 01 14:50:30 crc kubenswrapper[4810]: I1201 14:50:30.843112 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2552822d-ff62-49fa-8ba0-321952bda208-bundle\") pod \"2552822d-ff62-49fa-8ba0-321952bda208\" (UID: \"2552822d-ff62-49fa-8ba0-321952bda208\") " Dec 01 14:50:30 crc kubenswrapper[4810]: I1201 14:50:30.843998 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2552822d-ff62-49fa-8ba0-321952bda208-bundle" (OuterVolumeSpecName: "bundle") pod "2552822d-ff62-49fa-8ba0-321952bda208" (UID: "2552822d-ff62-49fa-8ba0-321952bda208"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:50:30 crc kubenswrapper[4810]: I1201 14:50:30.853633 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2552822d-ff62-49fa-8ba0-321952bda208-util" (OuterVolumeSpecName: "util") pod "2552822d-ff62-49fa-8ba0-321952bda208" (UID: "2552822d-ff62-49fa-8ba0-321952bda208"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:50:30 crc kubenswrapper[4810]: I1201 14:50:30.854102 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2552822d-ff62-49fa-8ba0-321952bda208-kube-api-access-vrjsc" (OuterVolumeSpecName: "kube-api-access-vrjsc") pod "2552822d-ff62-49fa-8ba0-321952bda208" (UID: "2552822d-ff62-49fa-8ba0-321952bda208"). InnerVolumeSpecName "kube-api-access-vrjsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:50:30 crc kubenswrapper[4810]: I1201 14:50:30.944141 4810 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2552822d-ff62-49fa-8ba0-321952bda208-util\") on node \"crc\" DevicePath \"\"" Dec 01 14:50:30 crc kubenswrapper[4810]: I1201 14:50:30.944171 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrjsc\" (UniqueName: \"kubernetes.io/projected/2552822d-ff62-49fa-8ba0-321952bda208-kube-api-access-vrjsc\") on node \"crc\" DevicePath \"\"" Dec 01 14:50:30 crc kubenswrapper[4810]: I1201 14:50:30.944183 4810 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2552822d-ff62-49fa-8ba0-321952bda208-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:50:31 crc kubenswrapper[4810]: I1201 14:50:31.586728 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq" event={"ID":"2552822d-ff62-49fa-8ba0-321952bda208","Type":"ContainerDied","Data":"2694d490f68bec3dc0263f271363683e7c966ab6465a4055c58a33f055a00161"} Dec 01 14:50:31 crc kubenswrapper[4810]: I1201 14:50:31.586774 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2694d490f68bec3dc0263f271363683e7c966ab6465a4055c58a33f055a00161" Dec 01 14:50:31 crc kubenswrapper[4810]: I1201 14:50:31.586831 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq" Dec 01 14:50:32 crc kubenswrapper[4810]: I1201 14:50:32.971740 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:50:32 crc kubenswrapper[4810]: I1201 14:50:32.971811 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:50:36 crc kubenswrapper[4810]: I1201 14:50:36.042763 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-q5qv8"] Dec 01 14:50:36 crc kubenswrapper[4810]: E1201 14:50:36.043975 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2552822d-ff62-49fa-8ba0-321952bda208" containerName="pull" Dec 01 14:50:36 crc kubenswrapper[4810]: I1201 14:50:36.044048 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2552822d-ff62-49fa-8ba0-321952bda208" containerName="pull" Dec 01 14:50:36 crc kubenswrapper[4810]: E1201 14:50:36.044127 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2552822d-ff62-49fa-8ba0-321952bda208" containerName="util" Dec 01 14:50:36 crc kubenswrapper[4810]: I1201 14:50:36.044182 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2552822d-ff62-49fa-8ba0-321952bda208" containerName="util" Dec 01 14:50:36 crc kubenswrapper[4810]: E1201 14:50:36.044268 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2552822d-ff62-49fa-8ba0-321952bda208" containerName="extract" Dec 01 14:50:36 crc kubenswrapper[4810]: I1201 14:50:36.044330 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2552822d-ff62-49fa-8ba0-321952bda208" containerName="extract" Dec 01 14:50:36 crc kubenswrapper[4810]: I1201 14:50:36.044545 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="2552822d-ff62-49fa-8ba0-321952bda208" containerName="extract" Dec 01 14:50:36 crc kubenswrapper[4810]: I1201 14:50:36.045031 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-q5qv8" Dec 01 14:50:36 crc kubenswrapper[4810]: I1201 14:50:36.046773 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Dec 01 14:50:36 crc kubenswrapper[4810]: I1201 14:50:36.046779 4810 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-jsgmm" Dec 01 14:50:36 crc kubenswrapper[4810]: I1201 14:50:36.047193 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Dec 01 14:50:36 crc kubenswrapper[4810]: I1201 14:50:36.057682 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-q5qv8"] Dec 01 14:50:36 crc kubenswrapper[4810]: I1201 14:50:36.223489 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz8qp\" (UniqueName: \"kubernetes.io/projected/4149e943-952d-4ce1-bfe5-64d3caafe7b9-kube-api-access-sz8qp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-q5qv8\" (UID: \"4149e943-952d-4ce1-bfe5-64d3caafe7b9\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-q5qv8" Dec 01 14:50:36 crc kubenswrapper[4810]: I1201 14:50:36.223551 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/4149e943-952d-4ce1-bfe5-64d3caafe7b9-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-q5qv8\" (UID: \"4149e943-952d-4ce1-bfe5-64d3caafe7b9\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-q5qv8" Dec 01 14:50:36 crc kubenswrapper[4810]: I1201 14:50:36.325244 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz8qp\" (UniqueName: \"kubernetes.io/projected/4149e943-952d-4ce1-bfe5-64d3caafe7b9-kube-api-access-sz8qp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-q5qv8\" (UID: \"4149e943-952d-4ce1-bfe5-64d3caafe7b9\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-q5qv8" Dec 01 14:50:36 crc kubenswrapper[4810]: I1201 14:50:36.325399 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/4149e943-952d-4ce1-bfe5-64d3caafe7b9-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-q5qv8\" (UID: \"4149e943-952d-4ce1-bfe5-64d3caafe7b9\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-q5qv8" Dec 01 14:50:36 crc kubenswrapper[4810]: I1201 14:50:36.325862 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/4149e943-952d-4ce1-bfe5-64d3caafe7b9-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-q5qv8\" (UID: \"4149e943-952d-4ce1-bfe5-64d3caafe7b9\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-q5qv8" Dec 01 14:50:36 crc kubenswrapper[4810]: I1201 14:50:36.346243 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz8qp\" (UniqueName: \"kubernetes.io/projected/4149e943-952d-4ce1-bfe5-64d3caafe7b9-kube-api-access-sz8qp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-q5qv8\" (UID: \"4149e943-952d-4ce1-bfe5-64d3caafe7b9\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-q5qv8" Dec 01 14:50:36 crc kubenswrapper[4810]: I1201 14:50:36.364041 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-q5qv8" Dec 01 14:50:36 crc kubenswrapper[4810]: I1201 14:50:36.796070 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-q5qv8"] Dec 01 14:50:36 crc kubenswrapper[4810]: W1201 14:50:36.803057 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4149e943_952d_4ce1_bfe5_64d3caafe7b9.slice/crio-75cb1d8c1077f62c302496952c4d393dadcccaea211e4df6ce80d265c766e964 WatchSource:0}: Error finding container 75cb1d8c1077f62c302496952c4d393dadcccaea211e4df6ce80d265c766e964: Status 404 returned error can't find the container with id 75cb1d8c1077f62c302496952c4d393dadcccaea211e4df6ce80d265c766e964 Dec 01 14:50:37 crc kubenswrapper[4810]: I1201 14:50:37.633832 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-q5qv8" event={"ID":"4149e943-952d-4ce1-bfe5-64d3caafe7b9","Type":"ContainerStarted","Data":"75cb1d8c1077f62c302496952c4d393dadcccaea211e4df6ce80d265c766e964"} Dec 01 14:50:45 crc kubenswrapper[4810]: I1201 14:50:45.688519 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-q5qv8" event={"ID":"4149e943-952d-4ce1-bfe5-64d3caafe7b9","Type":"ContainerStarted","Data":"a3b386525b227a5bfc628e4875f8cc2e0dd36a0656186d003c10c432d39b3767"} Dec 01 14:50:45 crc kubenswrapper[4810]: I1201 14:50:45.709925 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-q5qv8" podStartSLOduration=1.118152391 podStartE2EDuration="9.709908099s" podCreationTimestamp="2025-12-01 14:50:36 +0000 UTC" firstStartedPulling="2025-12-01 14:50:36.805347241 +0000 UTC m=+1002.568856834" lastFinishedPulling="2025-12-01 14:50:45.397102939 +0000 UTC m=+1011.160612542" observedRunningTime="2025-12-01 14:50:45.707458522 +0000 UTC m=+1011.470968125" watchObservedRunningTime="2025-12-01 14:50:45.709908099 +0000 UTC m=+1011.473417702" Dec 01 14:50:50 crc kubenswrapper[4810]: I1201 14:50:50.089657 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-j8wjf"] Dec 01 14:50:50 crc kubenswrapper[4810]: I1201 14:50:50.092033 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-j8wjf" Dec 01 14:50:50 crc kubenswrapper[4810]: I1201 14:50:50.095908 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 01 14:50:50 crc kubenswrapper[4810]: I1201 14:50:50.096047 4810 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-bg8g4" Dec 01 14:50:50 crc kubenswrapper[4810]: I1201 14:50:50.096217 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 01 14:50:50 crc kubenswrapper[4810]: I1201 14:50:50.107011 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-j8wjf"] Dec 01 14:50:50 crc kubenswrapper[4810]: I1201 14:50:50.188729 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4cabe2df-a7db-4c52-bfc1-4e205d62e3f4-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-j8wjf\" (UID: \"4cabe2df-a7db-4c52-bfc1-4e205d62e3f4\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-j8wjf" Dec 01 14:50:50 crc kubenswrapper[4810]: I1201 14:50:50.189219 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqhzj\" (UniqueName: \"kubernetes.io/projected/4cabe2df-a7db-4c52-bfc1-4e205d62e3f4-kube-api-access-kqhzj\") pod \"cert-manager-webhook-f4fb5df64-j8wjf\" (UID: \"4cabe2df-a7db-4c52-bfc1-4e205d62e3f4\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-j8wjf" Dec 01 14:50:50 crc kubenswrapper[4810]: I1201 14:50:50.290714 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqhzj\" (UniqueName: \"kubernetes.io/projected/4cabe2df-a7db-4c52-bfc1-4e205d62e3f4-kube-api-access-kqhzj\") pod \"cert-manager-webhook-f4fb5df64-j8wjf\" (UID: \"4cabe2df-a7db-4c52-bfc1-4e205d62e3f4\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-j8wjf" Dec 01 14:50:50 crc kubenswrapper[4810]: I1201 14:50:50.290805 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4cabe2df-a7db-4c52-bfc1-4e205d62e3f4-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-j8wjf\" (UID: \"4cabe2df-a7db-4c52-bfc1-4e205d62e3f4\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-j8wjf" Dec 01 14:50:50 crc kubenswrapper[4810]: I1201 14:50:50.309864 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4cabe2df-a7db-4c52-bfc1-4e205d62e3f4-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-j8wjf\" (UID: \"4cabe2df-a7db-4c52-bfc1-4e205d62e3f4\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-j8wjf" Dec 01 14:50:50 crc kubenswrapper[4810]: I1201 14:50:50.312747 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqhzj\" (UniqueName: \"kubernetes.io/projected/4cabe2df-a7db-4c52-bfc1-4e205d62e3f4-kube-api-access-kqhzj\") pod \"cert-manager-webhook-f4fb5df64-j8wjf\" (UID: \"4cabe2df-a7db-4c52-bfc1-4e205d62e3f4\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-j8wjf" Dec 01 14:50:50 crc kubenswrapper[4810]: I1201 14:50:50.409321 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-j8wjf" Dec 01 14:50:51 crc kubenswrapper[4810]: I1201 14:50:51.010859 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-j8wjf"] Dec 01 14:50:51 crc kubenswrapper[4810]: I1201 14:50:51.045079 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-pzdxf"] Dec 01 14:50:51 crc kubenswrapper[4810]: I1201 14:50:51.046031 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-pzdxf" Dec 01 14:50:51 crc kubenswrapper[4810]: I1201 14:50:51.058806 4810 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-v5n85" Dec 01 14:50:51 crc kubenswrapper[4810]: I1201 14:50:51.061614 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-pzdxf"] Dec 01 14:50:51 crc kubenswrapper[4810]: I1201 14:50:51.099696 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/73c25462-0e75-40f9-82bb-4ab8ae29d8cf-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-pzdxf\" (UID: \"73c25462-0e75-40f9-82bb-4ab8ae29d8cf\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-pzdxf" Dec 01 14:50:51 crc kubenswrapper[4810]: I1201 14:50:51.099758 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ml9l\" (UniqueName: \"kubernetes.io/projected/73c25462-0e75-40f9-82bb-4ab8ae29d8cf-kube-api-access-9ml9l\") pod \"cert-manager-cainjector-855d9ccff4-pzdxf\" (UID: \"73c25462-0e75-40f9-82bb-4ab8ae29d8cf\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-pzdxf" Dec 01 14:50:51 crc kubenswrapper[4810]: I1201 14:50:51.202153 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/73c25462-0e75-40f9-82bb-4ab8ae29d8cf-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-pzdxf\" (UID: \"73c25462-0e75-40f9-82bb-4ab8ae29d8cf\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-pzdxf" Dec 01 14:50:51 crc kubenswrapper[4810]: I1201 14:50:51.202215 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ml9l\" (UniqueName: \"kubernetes.io/projected/73c25462-0e75-40f9-82bb-4ab8ae29d8cf-kube-api-access-9ml9l\") pod \"cert-manager-cainjector-855d9ccff4-pzdxf\" (UID: \"73c25462-0e75-40f9-82bb-4ab8ae29d8cf\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-pzdxf" Dec 01 14:50:51 crc kubenswrapper[4810]: I1201 14:50:51.222151 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ml9l\" (UniqueName: \"kubernetes.io/projected/73c25462-0e75-40f9-82bb-4ab8ae29d8cf-kube-api-access-9ml9l\") pod \"cert-manager-cainjector-855d9ccff4-pzdxf\" (UID: \"73c25462-0e75-40f9-82bb-4ab8ae29d8cf\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-pzdxf" Dec 01 14:50:51 crc kubenswrapper[4810]: I1201 14:50:51.222915 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/73c25462-0e75-40f9-82bb-4ab8ae29d8cf-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-pzdxf\" (UID: \"73c25462-0e75-40f9-82bb-4ab8ae29d8cf\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-pzdxf" Dec 01 14:50:51 crc kubenswrapper[4810]: I1201 14:50:51.361143 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-pzdxf" Dec 01 14:50:51 crc kubenswrapper[4810]: I1201 14:50:51.722669 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-j8wjf" event={"ID":"4cabe2df-a7db-4c52-bfc1-4e205d62e3f4","Type":"ContainerStarted","Data":"f3306a438c2f29ac05a2b5fca633dc146b2abe3d84b633dcf54bac4bdef3fc97"} Dec 01 14:50:51 crc kubenswrapper[4810]: I1201 14:50:51.819881 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-pzdxf"] Dec 01 14:50:51 crc kubenswrapper[4810]: W1201 14:50:51.824866 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73c25462_0e75_40f9_82bb_4ab8ae29d8cf.slice/crio-77a891a5049e52c07be8871d6f8590c5d2dfa01bad61501a7791fa44ad179d4a WatchSource:0}: Error finding container 77a891a5049e52c07be8871d6f8590c5d2dfa01bad61501a7791fa44ad179d4a: Status 404 returned error can't find the container with id 77a891a5049e52c07be8871d6f8590c5d2dfa01bad61501a7791fa44ad179d4a Dec 01 14:50:52 crc kubenswrapper[4810]: I1201 14:50:52.730106 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-pzdxf" event={"ID":"73c25462-0e75-40f9-82bb-4ab8ae29d8cf","Type":"ContainerStarted","Data":"77a891a5049e52c07be8871d6f8590c5d2dfa01bad61501a7791fa44ad179d4a"} Dec 01 14:51:02 crc kubenswrapper[4810]: I1201 14:51:02.122812 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-pzdxf" event={"ID":"73c25462-0e75-40f9-82bb-4ab8ae29d8cf","Type":"ContainerStarted","Data":"33ef2e599524964443b21826f78035a681120a814f6f36a4182de7760ede8d40"} Dec 01 14:51:02 crc kubenswrapper[4810]: I1201 14:51:02.124941 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-j8wjf" event={"ID":"4cabe2df-a7db-4c52-bfc1-4e205d62e3f4","Type":"ContainerStarted","Data":"b98d196ddcaefa85196a464af407db0fe9d227fb1e821c761f103510841685ba"} Dec 01 14:51:02 crc kubenswrapper[4810]: I1201 14:51:02.125052 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-j8wjf" Dec 01 14:51:02 crc kubenswrapper[4810]: I1201 14:51:02.138018 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-pzdxf" podStartSLOduration=1.357890748 podStartE2EDuration="11.137998239s" podCreationTimestamp="2025-12-01 14:50:51 +0000 UTC" firstStartedPulling="2025-12-01 14:50:51.826794272 +0000 UTC m=+1017.590303875" lastFinishedPulling="2025-12-01 14:51:01.606901763 +0000 UTC m=+1027.370411366" observedRunningTime="2025-12-01 14:51:02.135161572 +0000 UTC m=+1027.898671185" watchObservedRunningTime="2025-12-01 14:51:02.137998239 +0000 UTC m=+1027.901507842" Dec 01 14:51:02 crc kubenswrapper[4810]: I1201 14:51:02.162553 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-j8wjf" podStartSLOduration=1.594254006 podStartE2EDuration="12.16253247s" podCreationTimestamp="2025-12-01 14:50:50 +0000 UTC" firstStartedPulling="2025-12-01 14:50:51.017656976 +0000 UTC m=+1016.781166579" lastFinishedPulling="2025-12-01 14:51:01.58593544 +0000 UTC m=+1027.349445043" observedRunningTime="2025-12-01 14:51:02.162264453 +0000 UTC m=+1027.925774066" watchObservedRunningTime="2025-12-01 14:51:02.16253247 +0000 UTC m=+1027.926042083" Dec 01 14:51:02 crc kubenswrapper[4810]: I1201 14:51:02.541845 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-6smjs"] Dec 01 14:51:02 crc kubenswrapper[4810]: I1201 14:51:02.542814 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-6smjs" Dec 01 14:51:02 crc kubenswrapper[4810]: I1201 14:51:02.544823 4810 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-lvjfv" Dec 01 14:51:02 crc kubenswrapper[4810]: I1201 14:51:02.547913 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7b781210-3868-4d4b-9f15-890ac7f08766-bound-sa-token\") pod \"cert-manager-86cb77c54b-6smjs\" (UID: \"7b781210-3868-4d4b-9f15-890ac7f08766\") " pod="cert-manager/cert-manager-86cb77c54b-6smjs" Dec 01 14:51:02 crc kubenswrapper[4810]: I1201 14:51:02.548028 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txk8j\" (UniqueName: \"kubernetes.io/projected/7b781210-3868-4d4b-9f15-890ac7f08766-kube-api-access-txk8j\") pod \"cert-manager-86cb77c54b-6smjs\" (UID: \"7b781210-3868-4d4b-9f15-890ac7f08766\") " pod="cert-manager/cert-manager-86cb77c54b-6smjs" Dec 01 14:51:02 crc kubenswrapper[4810]: I1201 14:51:02.548943 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-6smjs"] Dec 01 14:51:02 crc kubenswrapper[4810]: I1201 14:51:02.649087 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txk8j\" (UniqueName: \"kubernetes.io/projected/7b781210-3868-4d4b-9f15-890ac7f08766-kube-api-access-txk8j\") pod \"cert-manager-86cb77c54b-6smjs\" (UID: \"7b781210-3868-4d4b-9f15-890ac7f08766\") " pod="cert-manager/cert-manager-86cb77c54b-6smjs" Dec 01 14:51:02 crc kubenswrapper[4810]: I1201 14:51:02.649802 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7b781210-3868-4d4b-9f15-890ac7f08766-bound-sa-token\") pod \"cert-manager-86cb77c54b-6smjs\" (UID: \"7b781210-3868-4d4b-9f15-890ac7f08766\") " pod="cert-manager/cert-manager-86cb77c54b-6smjs" Dec 01 14:51:02 crc kubenswrapper[4810]: I1201 14:51:02.673877 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7b781210-3868-4d4b-9f15-890ac7f08766-bound-sa-token\") pod \"cert-manager-86cb77c54b-6smjs\" (UID: \"7b781210-3868-4d4b-9f15-890ac7f08766\") " pod="cert-manager/cert-manager-86cb77c54b-6smjs" Dec 01 14:51:02 crc kubenswrapper[4810]: I1201 14:51:02.675105 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txk8j\" (UniqueName: \"kubernetes.io/projected/7b781210-3868-4d4b-9f15-890ac7f08766-kube-api-access-txk8j\") pod \"cert-manager-86cb77c54b-6smjs\" (UID: \"7b781210-3868-4d4b-9f15-890ac7f08766\") " pod="cert-manager/cert-manager-86cb77c54b-6smjs" Dec 01 14:51:02 crc kubenswrapper[4810]: I1201 14:51:02.859604 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-6smjs" Dec 01 14:51:02 crc kubenswrapper[4810]: I1201 14:51:02.972085 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:51:02 crc kubenswrapper[4810]: I1201 14:51:02.972748 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:51:03 crc kubenswrapper[4810]: I1201 14:51:03.255020 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-6smjs"] Dec 01 14:51:04 crc kubenswrapper[4810]: I1201 14:51:04.137096 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-6smjs" event={"ID":"7b781210-3868-4d4b-9f15-890ac7f08766","Type":"ContainerStarted","Data":"c22a0098ad44a2a73389e5b7b454355fbc6df96a060d6bf26a75c533a541eeb4"} Dec 01 14:51:04 crc kubenswrapper[4810]: I1201 14:51:04.137386 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-6smjs" event={"ID":"7b781210-3868-4d4b-9f15-890ac7f08766","Type":"ContainerStarted","Data":"57639304bcbc22a903a682536b0c9e0f07eaa9771de5745177d5bdb2de9e403a"} Dec 01 14:51:04 crc kubenswrapper[4810]: I1201 14:51:04.151806 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-6smjs" podStartSLOduration=2.151785613 podStartE2EDuration="2.151785613s" podCreationTimestamp="2025-12-01 14:51:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:51:04.151295339 +0000 UTC m=+1029.914804962" watchObservedRunningTime="2025-12-01 14:51:04.151785613 +0000 UTC m=+1029.915295216" Dec 01 14:51:10 crc kubenswrapper[4810]: I1201 14:51:10.413535 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-j8wjf" Dec 01 14:51:13 crc kubenswrapper[4810]: I1201 14:51:13.433727 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-rn2zk"] Dec 01 14:51:13 crc kubenswrapper[4810]: I1201 14:51:13.436170 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rn2zk" Dec 01 14:51:13 crc kubenswrapper[4810]: I1201 14:51:13.444130 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-xtkws" Dec 01 14:51:13 crc kubenswrapper[4810]: I1201 14:51:13.444400 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 01 14:51:13 crc kubenswrapper[4810]: I1201 14:51:13.444146 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 01 14:51:13 crc kubenswrapper[4810]: I1201 14:51:13.455344 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rn2zk"] Dec 01 14:51:13 crc kubenswrapper[4810]: I1201 14:51:13.529264 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4grk8\" (UniqueName: \"kubernetes.io/projected/b8307b6c-fba8-44d9-8ed4-1627cfa87006-kube-api-access-4grk8\") pod \"openstack-operator-index-rn2zk\" (UID: \"b8307b6c-fba8-44d9-8ed4-1627cfa87006\") " pod="openstack-operators/openstack-operator-index-rn2zk" Dec 01 14:51:13 crc kubenswrapper[4810]: I1201 14:51:13.630896 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4grk8\" (UniqueName: \"kubernetes.io/projected/b8307b6c-fba8-44d9-8ed4-1627cfa87006-kube-api-access-4grk8\") pod \"openstack-operator-index-rn2zk\" (UID: \"b8307b6c-fba8-44d9-8ed4-1627cfa87006\") " pod="openstack-operators/openstack-operator-index-rn2zk" Dec 01 14:51:13 crc kubenswrapper[4810]: I1201 14:51:13.658841 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4grk8\" (UniqueName: \"kubernetes.io/projected/b8307b6c-fba8-44d9-8ed4-1627cfa87006-kube-api-access-4grk8\") pod \"openstack-operator-index-rn2zk\" (UID: \"b8307b6c-fba8-44d9-8ed4-1627cfa87006\") " pod="openstack-operators/openstack-operator-index-rn2zk" Dec 01 14:51:13 crc kubenswrapper[4810]: I1201 14:51:13.755178 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rn2zk" Dec 01 14:51:14 crc kubenswrapper[4810]: I1201 14:51:14.270674 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rn2zk"] Dec 01 14:51:14 crc kubenswrapper[4810]: W1201 14:51:14.272533 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8307b6c_fba8_44d9_8ed4_1627cfa87006.slice/crio-ac657e3e80a04305a449cddc5b77e37a31d0adb060d4ebc33603c92ab961fc0a WatchSource:0}: Error finding container ac657e3e80a04305a449cddc5b77e37a31d0adb060d4ebc33603c92ab961fc0a: Status 404 returned error can't find the container with id ac657e3e80a04305a449cddc5b77e37a31d0adb060d4ebc33603c92ab961fc0a Dec 01 14:51:15 crc kubenswrapper[4810]: I1201 14:51:15.211532 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rn2zk" event={"ID":"b8307b6c-fba8-44d9-8ed4-1627cfa87006","Type":"ContainerStarted","Data":"ac657e3e80a04305a449cddc5b77e37a31d0adb060d4ebc33603c92ab961fc0a"} Dec 01 14:51:16 crc kubenswrapper[4810]: I1201 14:51:16.219023 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rn2zk" event={"ID":"b8307b6c-fba8-44d9-8ed4-1627cfa87006","Type":"ContainerStarted","Data":"8c97d12ccf5a5aaeefbaf0c10886d03e5a9aa491e7b07e93ff24481b1bced45d"} Dec 01 14:51:16 crc kubenswrapper[4810]: I1201 14:51:16.241027 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-rn2zk" podStartSLOduration=1.837001229 podStartE2EDuration="3.241010438s" podCreationTimestamp="2025-12-01 14:51:13 +0000 UTC" firstStartedPulling="2025-12-01 14:51:14.275130691 +0000 UTC m=+1040.038640334" lastFinishedPulling="2025-12-01 14:51:15.67913993 +0000 UTC m=+1041.442649543" observedRunningTime="2025-12-01 14:51:16.240197126 +0000 UTC m=+1042.003706769" watchObservedRunningTime="2025-12-01 14:51:16.241010438 +0000 UTC m=+1042.004520061" Dec 01 14:51:16 crc kubenswrapper[4810]: I1201 14:51:16.599121 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-rn2zk"] Dec 01 14:51:17 crc kubenswrapper[4810]: I1201 14:51:17.206020 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-4s9sl"] Dec 01 14:51:17 crc kubenswrapper[4810]: I1201 14:51:17.210121 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4s9sl" Dec 01 14:51:17 crc kubenswrapper[4810]: I1201 14:51:17.212223 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-4s9sl"] Dec 01 14:51:17 crc kubenswrapper[4810]: I1201 14:51:17.330724 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpm5w\" (UniqueName: \"kubernetes.io/projected/15fa679c-c2b6-4b41-a18c-cf0f42d1b4cf-kube-api-access-xpm5w\") pod \"openstack-operator-index-4s9sl\" (UID: \"15fa679c-c2b6-4b41-a18c-cf0f42d1b4cf\") " pod="openstack-operators/openstack-operator-index-4s9sl" Dec 01 14:51:17 crc kubenswrapper[4810]: I1201 14:51:17.433133 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpm5w\" (UniqueName: \"kubernetes.io/projected/15fa679c-c2b6-4b41-a18c-cf0f42d1b4cf-kube-api-access-xpm5w\") pod \"openstack-operator-index-4s9sl\" (UID: \"15fa679c-c2b6-4b41-a18c-cf0f42d1b4cf\") " pod="openstack-operators/openstack-operator-index-4s9sl" Dec 01 14:51:17 crc kubenswrapper[4810]: I1201 14:51:17.452442 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpm5w\" (UniqueName: \"kubernetes.io/projected/15fa679c-c2b6-4b41-a18c-cf0f42d1b4cf-kube-api-access-xpm5w\") pod \"openstack-operator-index-4s9sl\" (UID: \"15fa679c-c2b6-4b41-a18c-cf0f42d1b4cf\") " pod="openstack-operators/openstack-operator-index-4s9sl" Dec 01 14:51:17 crc kubenswrapper[4810]: I1201 14:51:17.526748 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4s9sl" Dec 01 14:51:18 crc kubenswrapper[4810]: I1201 14:51:18.100688 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-4s9sl"] Dec 01 14:51:18 crc kubenswrapper[4810]: W1201 14:51:18.107890 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15fa679c_c2b6_4b41_a18c_cf0f42d1b4cf.slice/crio-1dcc2d33b0b912250594f373953b60cfbcade8b94a4de521b5de7298b06de34f WatchSource:0}: Error finding container 1dcc2d33b0b912250594f373953b60cfbcade8b94a4de521b5de7298b06de34f: Status 404 returned error can't find the container with id 1dcc2d33b0b912250594f373953b60cfbcade8b94a4de521b5de7298b06de34f Dec 01 14:51:18 crc kubenswrapper[4810]: I1201 14:51:18.232118 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4s9sl" event={"ID":"15fa679c-c2b6-4b41-a18c-cf0f42d1b4cf","Type":"ContainerStarted","Data":"1dcc2d33b0b912250594f373953b60cfbcade8b94a4de521b5de7298b06de34f"} Dec 01 14:51:18 crc kubenswrapper[4810]: I1201 14:51:18.232254 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-rn2zk" podUID="b8307b6c-fba8-44d9-8ed4-1627cfa87006" containerName="registry-server" containerID="cri-o://8c97d12ccf5a5aaeefbaf0c10886d03e5a9aa491e7b07e93ff24481b1bced45d" gracePeriod=2 Dec 01 14:51:18 crc kubenswrapper[4810]: I1201 14:51:18.641347 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rn2zk" Dec 01 14:51:18 crc kubenswrapper[4810]: I1201 14:51:18.750767 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4grk8\" (UniqueName: \"kubernetes.io/projected/b8307b6c-fba8-44d9-8ed4-1627cfa87006-kube-api-access-4grk8\") pod \"b8307b6c-fba8-44d9-8ed4-1627cfa87006\" (UID: \"b8307b6c-fba8-44d9-8ed4-1627cfa87006\") " Dec 01 14:51:18 crc kubenswrapper[4810]: I1201 14:51:18.762681 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8307b6c-fba8-44d9-8ed4-1627cfa87006-kube-api-access-4grk8" (OuterVolumeSpecName: "kube-api-access-4grk8") pod "b8307b6c-fba8-44d9-8ed4-1627cfa87006" (UID: "b8307b6c-fba8-44d9-8ed4-1627cfa87006"). InnerVolumeSpecName "kube-api-access-4grk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:51:18 crc kubenswrapper[4810]: I1201 14:51:18.852817 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4grk8\" (UniqueName: \"kubernetes.io/projected/b8307b6c-fba8-44d9-8ed4-1627cfa87006-kube-api-access-4grk8\") on node \"crc\" DevicePath \"\"" Dec 01 14:51:19 crc kubenswrapper[4810]: I1201 14:51:19.239877 4810 generic.go:334] "Generic (PLEG): container finished" podID="b8307b6c-fba8-44d9-8ed4-1627cfa87006" containerID="8c97d12ccf5a5aaeefbaf0c10886d03e5a9aa491e7b07e93ff24481b1bced45d" exitCode=0 Dec 01 14:51:19 crc kubenswrapper[4810]: I1201 14:51:19.239918 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rn2zk" Dec 01 14:51:19 crc kubenswrapper[4810]: I1201 14:51:19.239977 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rn2zk" event={"ID":"b8307b6c-fba8-44d9-8ed4-1627cfa87006","Type":"ContainerDied","Data":"8c97d12ccf5a5aaeefbaf0c10886d03e5a9aa491e7b07e93ff24481b1bced45d"} Dec 01 14:51:19 crc kubenswrapper[4810]: I1201 14:51:19.240105 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rn2zk" event={"ID":"b8307b6c-fba8-44d9-8ed4-1627cfa87006","Type":"ContainerDied","Data":"ac657e3e80a04305a449cddc5b77e37a31d0adb060d4ebc33603c92ab961fc0a"} Dec 01 14:51:19 crc kubenswrapper[4810]: I1201 14:51:19.240123 4810 scope.go:117] "RemoveContainer" containerID="8c97d12ccf5a5aaeefbaf0c10886d03e5a9aa491e7b07e93ff24481b1bced45d" Dec 01 14:51:19 crc kubenswrapper[4810]: I1201 14:51:19.241838 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4s9sl" event={"ID":"15fa679c-c2b6-4b41-a18c-cf0f42d1b4cf","Type":"ContainerStarted","Data":"67115c9cd6bec93d7483ace2f1b4f4e95f46fc1a66c5e4a2ce47184e4908344f"} Dec 01 14:51:19 crc kubenswrapper[4810]: I1201 14:51:19.262264 4810 scope.go:117] "RemoveContainer" containerID="8c97d12ccf5a5aaeefbaf0c10886d03e5a9aa491e7b07e93ff24481b1bced45d" Dec 01 14:51:19 crc kubenswrapper[4810]: E1201 14:51:19.262759 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c97d12ccf5a5aaeefbaf0c10886d03e5a9aa491e7b07e93ff24481b1bced45d\": container with ID starting with 8c97d12ccf5a5aaeefbaf0c10886d03e5a9aa491e7b07e93ff24481b1bced45d not found: ID does not exist" containerID="8c97d12ccf5a5aaeefbaf0c10886d03e5a9aa491e7b07e93ff24481b1bced45d" Dec 01 14:51:19 crc kubenswrapper[4810]: I1201 14:51:19.262851 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c97d12ccf5a5aaeefbaf0c10886d03e5a9aa491e7b07e93ff24481b1bced45d"} err="failed to get container status \"8c97d12ccf5a5aaeefbaf0c10886d03e5a9aa491e7b07e93ff24481b1bced45d\": rpc error: code = NotFound desc = could not find container \"8c97d12ccf5a5aaeefbaf0c10886d03e5a9aa491e7b07e93ff24481b1bced45d\": container with ID starting with 8c97d12ccf5a5aaeefbaf0c10886d03e5a9aa491e7b07e93ff24481b1bced45d not found: ID does not exist" Dec 01 14:51:19 crc kubenswrapper[4810]: I1201 14:51:19.264160 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-4s9sl" podStartSLOduration=1.846729501 podStartE2EDuration="2.264137443s" podCreationTimestamp="2025-12-01 14:51:17 +0000 UTC" firstStartedPulling="2025-12-01 14:51:18.110927847 +0000 UTC m=+1043.874437450" lastFinishedPulling="2025-12-01 14:51:18.528335789 +0000 UTC m=+1044.291845392" observedRunningTime="2025-12-01 14:51:19.258462119 +0000 UTC m=+1045.021971722" watchObservedRunningTime="2025-12-01 14:51:19.264137443 +0000 UTC m=+1045.027647046" Dec 01 14:51:19 crc kubenswrapper[4810]: I1201 14:51:19.285021 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-rn2zk"] Dec 01 14:51:19 crc kubenswrapper[4810]: I1201 14:51:19.289210 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-rn2zk"] Dec 01 14:51:20 crc kubenswrapper[4810]: I1201 14:51:20.504623 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8307b6c-fba8-44d9-8ed4-1627cfa87006" path="/var/lib/kubelet/pods/b8307b6c-fba8-44d9-8ed4-1627cfa87006/volumes" Dec 01 14:51:27 crc kubenswrapper[4810]: I1201 14:51:27.527096 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-4s9sl" Dec 01 14:51:27 crc kubenswrapper[4810]: I1201 14:51:27.528619 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-4s9sl" Dec 01 14:51:27 crc kubenswrapper[4810]: I1201 14:51:27.558480 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-4s9sl" Dec 01 14:51:28 crc kubenswrapper[4810]: I1201 14:51:28.336224 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-4s9sl" Dec 01 14:51:32 crc kubenswrapper[4810]: I1201 14:51:32.972017 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:51:32 crc kubenswrapper[4810]: I1201 14:51:32.972685 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:51:32 crc kubenswrapper[4810]: I1201 14:51:32.972932 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:51:32 crc kubenswrapper[4810]: I1201 14:51:32.973605 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7d9c11d61c66e363668182137baa230beaee1296f3efd5bf440ccc7a13833e7c"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 14:51:32 crc kubenswrapper[4810]: I1201 14:51:32.973670 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://7d9c11d61c66e363668182137baa230beaee1296f3efd5bf440ccc7a13833e7c" gracePeriod=600 Dec 01 14:51:33 crc kubenswrapper[4810]: I1201 14:51:33.334275 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="7d9c11d61c66e363668182137baa230beaee1296f3efd5bf440ccc7a13833e7c" exitCode=0 Dec 01 14:51:33 crc kubenswrapper[4810]: I1201 14:51:33.334523 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"7d9c11d61c66e363668182137baa230beaee1296f3efd5bf440ccc7a13833e7c"} Dec 01 14:51:33 crc kubenswrapper[4810]: I1201 14:51:33.334575 4810 scope.go:117] "RemoveContainer" containerID="d9ccbffc9cbcd02b32617e3a533f267bf2f98615c6b47efb95676ef8e399d454" Dec 01 14:51:34 crc kubenswrapper[4810]: I1201 14:51:34.344212 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"8e1ade99d7746e14166be3755e17c0d1329bcaa748d059179439c1c0ba6fc843"} Dec 01 14:51:35 crc kubenswrapper[4810]: I1201 14:51:35.151141 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6"] Dec 01 14:51:35 crc kubenswrapper[4810]: E1201 14:51:35.151798 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8307b6c-fba8-44d9-8ed4-1627cfa87006" containerName="registry-server" Dec 01 14:51:35 crc kubenswrapper[4810]: I1201 14:51:35.151819 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8307b6c-fba8-44d9-8ed4-1627cfa87006" containerName="registry-server" Dec 01 14:51:35 crc kubenswrapper[4810]: I1201 14:51:35.151952 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8307b6c-fba8-44d9-8ed4-1627cfa87006" containerName="registry-server" Dec 01 14:51:35 crc kubenswrapper[4810]: I1201 14:51:35.152860 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6" Dec 01 14:51:35 crc kubenswrapper[4810]: I1201 14:51:35.154281 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-4dp9m" Dec 01 14:51:35 crc kubenswrapper[4810]: I1201 14:51:35.162783 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6"] Dec 01 14:51:35 crc kubenswrapper[4810]: I1201 14:51:35.271280 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d8a8743-6e6d-438c-bb7a-c48047c8c7ed-bundle\") pod \"34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6\" (UID: \"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed\") " pod="openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6" Dec 01 14:51:35 crc kubenswrapper[4810]: I1201 14:51:35.271345 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d8a8743-6e6d-438c-bb7a-c48047c8c7ed-util\") pod \"34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6\" (UID: \"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed\") " pod="openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6" Dec 01 14:51:35 crc kubenswrapper[4810]: I1201 14:51:35.271377 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzpqb\" (UniqueName: \"kubernetes.io/projected/2d8a8743-6e6d-438c-bb7a-c48047c8c7ed-kube-api-access-fzpqb\") pod \"34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6\" (UID: \"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed\") " pod="openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6" Dec 01 14:51:35 crc kubenswrapper[4810]: I1201 14:51:35.372976 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d8a8743-6e6d-438c-bb7a-c48047c8c7ed-bundle\") pod \"34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6\" (UID: \"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed\") " pod="openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6" Dec 01 14:51:35 crc kubenswrapper[4810]: I1201 14:51:35.373054 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d8a8743-6e6d-438c-bb7a-c48047c8c7ed-util\") pod \"34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6\" (UID: \"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed\") " pod="openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6" Dec 01 14:51:35 crc kubenswrapper[4810]: I1201 14:51:35.373094 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzpqb\" (UniqueName: \"kubernetes.io/projected/2d8a8743-6e6d-438c-bb7a-c48047c8c7ed-kube-api-access-fzpqb\") pod \"34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6\" (UID: \"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed\") " pod="openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6" Dec 01 14:51:35 crc kubenswrapper[4810]: I1201 14:51:35.373527 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d8a8743-6e6d-438c-bb7a-c48047c8c7ed-bundle\") pod \"34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6\" (UID: \"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed\") " pod="openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6" Dec 01 14:51:35 crc kubenswrapper[4810]: I1201 14:51:35.373555 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d8a8743-6e6d-438c-bb7a-c48047c8c7ed-util\") pod \"34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6\" (UID: \"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed\") " pod="openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6" Dec 01 14:51:35 crc kubenswrapper[4810]: I1201 14:51:35.394786 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzpqb\" (UniqueName: \"kubernetes.io/projected/2d8a8743-6e6d-438c-bb7a-c48047c8c7ed-kube-api-access-fzpqb\") pod \"34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6\" (UID: \"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed\") " pod="openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6" Dec 01 14:51:35 crc kubenswrapper[4810]: I1201 14:51:35.468687 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6" Dec 01 14:51:35 crc kubenswrapper[4810]: I1201 14:51:35.893325 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6"] Dec 01 14:51:36 crc kubenswrapper[4810]: I1201 14:51:36.359676 4810 generic.go:334] "Generic (PLEG): container finished" podID="2d8a8743-6e6d-438c-bb7a-c48047c8c7ed" containerID="4e950a28a4240fcaa2b6f4a78cb73b9a5f3423f58da03804f900d610fd772661" exitCode=0 Dec 01 14:51:36 crc kubenswrapper[4810]: I1201 14:51:36.359739 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6" event={"ID":"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed","Type":"ContainerDied","Data":"4e950a28a4240fcaa2b6f4a78cb73b9a5f3423f58da03804f900d610fd772661"} Dec 01 14:51:36 crc kubenswrapper[4810]: I1201 14:51:36.360113 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6" event={"ID":"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed","Type":"ContainerStarted","Data":"49e2497557c8fc1a91a3b78153981677be545c253e8cc346a7e13672039426fe"} Dec 01 14:51:38 crc kubenswrapper[4810]: I1201 14:51:38.374903 4810 generic.go:334] "Generic (PLEG): container finished" podID="2d8a8743-6e6d-438c-bb7a-c48047c8c7ed" containerID="acdce0b24e16520b9cb64145d5da6a8a512cc2f7823ff4b670393612784dd5dc" exitCode=0 Dec 01 14:51:38 crc kubenswrapper[4810]: I1201 14:51:38.375003 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6" event={"ID":"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed","Type":"ContainerDied","Data":"acdce0b24e16520b9cb64145d5da6a8a512cc2f7823ff4b670393612784dd5dc"} Dec 01 14:51:39 crc kubenswrapper[4810]: I1201 14:51:39.382370 4810 generic.go:334] "Generic (PLEG): container finished" podID="2d8a8743-6e6d-438c-bb7a-c48047c8c7ed" containerID="0ef20204dbacf22f848065595f90e3f28c6d3f99726077880dcec9b7048ee6ff" exitCode=0 Dec 01 14:51:39 crc kubenswrapper[4810]: I1201 14:51:39.382441 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6" event={"ID":"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed","Type":"ContainerDied","Data":"0ef20204dbacf22f848065595f90e3f28c6d3f99726077880dcec9b7048ee6ff"} Dec 01 14:51:40 crc kubenswrapper[4810]: I1201 14:51:40.718741 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6" Dec 01 14:51:40 crc kubenswrapper[4810]: I1201 14:51:40.848088 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d8a8743-6e6d-438c-bb7a-c48047c8c7ed-util\") pod \"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed\" (UID: \"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed\") " Dec 01 14:51:40 crc kubenswrapper[4810]: I1201 14:51:40.848154 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d8a8743-6e6d-438c-bb7a-c48047c8c7ed-bundle\") pod \"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed\" (UID: \"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed\") " Dec 01 14:51:40 crc kubenswrapper[4810]: I1201 14:51:40.848215 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzpqb\" (UniqueName: \"kubernetes.io/projected/2d8a8743-6e6d-438c-bb7a-c48047c8c7ed-kube-api-access-fzpqb\") pod \"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed\" (UID: \"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed\") " Dec 01 14:51:40 crc kubenswrapper[4810]: I1201 14:51:40.848877 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d8a8743-6e6d-438c-bb7a-c48047c8c7ed-bundle" (OuterVolumeSpecName: "bundle") pod "2d8a8743-6e6d-438c-bb7a-c48047c8c7ed" (UID: "2d8a8743-6e6d-438c-bb7a-c48047c8c7ed"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:51:40 crc kubenswrapper[4810]: I1201 14:51:40.853220 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d8a8743-6e6d-438c-bb7a-c48047c8c7ed-kube-api-access-fzpqb" (OuterVolumeSpecName: "kube-api-access-fzpqb") pod "2d8a8743-6e6d-438c-bb7a-c48047c8c7ed" (UID: "2d8a8743-6e6d-438c-bb7a-c48047c8c7ed"). InnerVolumeSpecName "kube-api-access-fzpqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:51:40 crc kubenswrapper[4810]: I1201 14:51:40.862159 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d8a8743-6e6d-438c-bb7a-c48047c8c7ed-util" (OuterVolumeSpecName: "util") pod "2d8a8743-6e6d-438c-bb7a-c48047c8c7ed" (UID: "2d8a8743-6e6d-438c-bb7a-c48047c8c7ed"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:51:40 crc kubenswrapper[4810]: I1201 14:51:40.949636 4810 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d8a8743-6e6d-438c-bb7a-c48047c8c7ed-util\") on node \"crc\" DevicePath \"\"" Dec 01 14:51:40 crc kubenswrapper[4810]: I1201 14:51:40.949675 4810 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d8a8743-6e6d-438c-bb7a-c48047c8c7ed-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:51:40 crc kubenswrapper[4810]: I1201 14:51:40.949685 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzpqb\" (UniqueName: \"kubernetes.io/projected/2d8a8743-6e6d-438c-bb7a-c48047c8c7ed-kube-api-access-fzpqb\") on node \"crc\" DevicePath \"\"" Dec 01 14:51:41 crc kubenswrapper[4810]: I1201 14:51:41.396114 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6" event={"ID":"2d8a8743-6e6d-438c-bb7a-c48047c8c7ed","Type":"ContainerDied","Data":"49e2497557c8fc1a91a3b78153981677be545c253e8cc346a7e13672039426fe"} Dec 01 14:51:41 crc kubenswrapper[4810]: I1201 14:51:41.396430 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49e2497557c8fc1a91a3b78153981677be545c253e8cc346a7e13672039426fe" Dec 01 14:51:41 crc kubenswrapper[4810]: I1201 14:51:41.396392 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6" Dec 01 14:51:47 crc kubenswrapper[4810]: I1201 14:51:47.189927 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6ddddd9d6f-rs8qj"] Dec 01 14:51:47 crc kubenswrapper[4810]: E1201 14:51:47.194501 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d8a8743-6e6d-438c-bb7a-c48047c8c7ed" containerName="extract" Dec 01 14:51:47 crc kubenswrapper[4810]: I1201 14:51:47.194521 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d8a8743-6e6d-438c-bb7a-c48047c8c7ed" containerName="extract" Dec 01 14:51:47 crc kubenswrapper[4810]: E1201 14:51:47.194536 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d8a8743-6e6d-438c-bb7a-c48047c8c7ed" containerName="pull" Dec 01 14:51:47 crc kubenswrapper[4810]: I1201 14:51:47.194544 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d8a8743-6e6d-438c-bb7a-c48047c8c7ed" containerName="pull" Dec 01 14:51:47 crc kubenswrapper[4810]: E1201 14:51:47.194562 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d8a8743-6e6d-438c-bb7a-c48047c8c7ed" containerName="util" Dec 01 14:51:47 crc kubenswrapper[4810]: I1201 14:51:47.194571 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d8a8743-6e6d-438c-bb7a-c48047c8c7ed" containerName="util" Dec 01 14:51:47 crc kubenswrapper[4810]: I1201 14:51:47.194708 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d8a8743-6e6d-438c-bb7a-c48047c8c7ed" containerName="extract" Dec 01 14:51:47 crc kubenswrapper[4810]: I1201 14:51:47.197771 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6ddddd9d6f-rs8qj" Dec 01 14:51:47 crc kubenswrapper[4810]: I1201 14:51:47.200039 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-v84wl" Dec 01 14:51:47 crc kubenswrapper[4810]: I1201 14:51:47.224004 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6ddddd9d6f-rs8qj"] Dec 01 14:51:47 crc kubenswrapper[4810]: I1201 14:51:47.342260 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d9r6\" (UniqueName: \"kubernetes.io/projected/666e10e7-f807-43b7-9ee4-5dabdeb0c8f7-kube-api-access-6d9r6\") pod \"openstack-operator-controller-operator-6ddddd9d6f-rs8qj\" (UID: \"666e10e7-f807-43b7-9ee4-5dabdeb0c8f7\") " pod="openstack-operators/openstack-operator-controller-operator-6ddddd9d6f-rs8qj" Dec 01 14:51:47 crc kubenswrapper[4810]: I1201 14:51:47.443105 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d9r6\" (UniqueName: \"kubernetes.io/projected/666e10e7-f807-43b7-9ee4-5dabdeb0c8f7-kube-api-access-6d9r6\") pod \"openstack-operator-controller-operator-6ddddd9d6f-rs8qj\" (UID: \"666e10e7-f807-43b7-9ee4-5dabdeb0c8f7\") " pod="openstack-operators/openstack-operator-controller-operator-6ddddd9d6f-rs8qj" Dec 01 14:51:47 crc kubenswrapper[4810]: I1201 14:51:47.462361 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d9r6\" (UniqueName: \"kubernetes.io/projected/666e10e7-f807-43b7-9ee4-5dabdeb0c8f7-kube-api-access-6d9r6\") pod \"openstack-operator-controller-operator-6ddddd9d6f-rs8qj\" (UID: \"666e10e7-f807-43b7-9ee4-5dabdeb0c8f7\") " pod="openstack-operators/openstack-operator-controller-operator-6ddddd9d6f-rs8qj" Dec 01 14:51:47 crc kubenswrapper[4810]: I1201 14:51:47.515802 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6ddddd9d6f-rs8qj" Dec 01 14:51:47 crc kubenswrapper[4810]: I1201 14:51:47.962361 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6ddddd9d6f-rs8qj"] Dec 01 14:51:48 crc kubenswrapper[4810]: I1201 14:51:48.439385 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6ddddd9d6f-rs8qj" event={"ID":"666e10e7-f807-43b7-9ee4-5dabdeb0c8f7","Type":"ContainerStarted","Data":"6e6349df9a8ca8d4fd38d446af8e0b00c901b98fb9629e1ae972ae25feb72b6c"} Dec 01 14:51:54 crc kubenswrapper[4810]: I1201 14:51:54.498574 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6ddddd9d6f-rs8qj" event={"ID":"666e10e7-f807-43b7-9ee4-5dabdeb0c8f7","Type":"ContainerStarted","Data":"5adec288ae09678ea986de222ea5cb860f8c8fd203b5be722d003a29489ae841"} Dec 01 14:51:54 crc kubenswrapper[4810]: I1201 14:51:54.499132 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6ddddd9d6f-rs8qj" Dec 01 14:51:54 crc kubenswrapper[4810]: I1201 14:51:54.547993 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6ddddd9d6f-rs8qj" podStartSLOduration=1.316985031 podStartE2EDuration="7.547976361s" podCreationTimestamp="2025-12-01 14:51:47 +0000 UTC" firstStartedPulling="2025-12-01 14:51:47.976642603 +0000 UTC m=+1073.740152216" lastFinishedPulling="2025-12-01 14:51:54.207633943 +0000 UTC m=+1079.971143546" observedRunningTime="2025-12-01 14:51:54.546893791 +0000 UTC m=+1080.310403404" watchObservedRunningTime="2025-12-01 14:51:54.547976361 +0000 UTC m=+1080.311485964" Dec 01 14:52:07 crc kubenswrapper[4810]: I1201 14:52:07.519209 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6ddddd9d6f-rs8qj" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.396666 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-4zjxs"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.398144 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4zjxs" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.399779 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-kj5jw" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.406352 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-ccwrk"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.407483 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ccwrk" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.409939 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-kgph5" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.428287 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-ccwrk"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.436176 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-7bcjp"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.437637 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-7bcjp" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.440365 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-twt95" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.459246 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-nnvmc"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.460994 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-nnvmc" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.461684 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-7bcjp"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.475217 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-f9gkq" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.490086 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lh6jm"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.490549 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgrjm\" (UniqueName: \"kubernetes.io/projected/308f7f89-2d44-42cb-8533-2b82306341c6-kube-api-access-pgrjm\") pod \"barbican-operator-controller-manager-7d9dfd778-4zjxs\" (UID: \"308f7f89-2d44-42cb-8533-2b82306341c6\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4zjxs" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.490711 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4gcj\" (UniqueName: \"kubernetes.io/projected/322588d9-4c11-4676-8eb4-15cce77ccd44-kube-api-access-l4gcj\") pod \"designate-operator-controller-manager-78b4bc895b-7bcjp\" (UID: \"322588d9-4c11-4676-8eb4-15cce77ccd44\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-7bcjp" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.490908 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vmxf\" (UniqueName: \"kubernetes.io/projected/ae4759b1-d998-49b4-8031-f638141177a5-kube-api-access-2vmxf\") pod \"glance-operator-controller-manager-668d9c48b9-nnvmc\" (UID: \"ae4759b1-d998-49b4-8031-f638141177a5\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-nnvmc" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.491041 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7x6z\" (UniqueName: \"kubernetes.io/projected/4b6ce85f-d265-40a2-aa92-2a2d76339acf-kube-api-access-w7x6z\") pod \"cinder-operator-controller-manager-859b6ccc6-ccwrk\" (UID: \"4b6ce85f-d265-40a2-aa92-2a2d76339acf\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ccwrk" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.493403 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lh6jm" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.496028 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-s4t47" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.557556 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-nnvmc"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.609821 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9zzgw"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.615292 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgrjm\" (UniqueName: \"kubernetes.io/projected/308f7f89-2d44-42cb-8533-2b82306341c6-kube-api-access-pgrjm\") pod \"barbican-operator-controller-manager-7d9dfd778-4zjxs\" (UID: \"308f7f89-2d44-42cb-8533-2b82306341c6\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4zjxs" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.615584 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4gcj\" (UniqueName: \"kubernetes.io/projected/322588d9-4c11-4676-8eb4-15cce77ccd44-kube-api-access-l4gcj\") pod \"designate-operator-controller-manager-78b4bc895b-7bcjp\" (UID: \"322588d9-4c11-4676-8eb4-15cce77ccd44\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-7bcjp" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.615744 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vmxf\" (UniqueName: \"kubernetes.io/projected/ae4759b1-d998-49b4-8031-f638141177a5-kube-api-access-2vmxf\") pod \"glance-operator-controller-manager-668d9c48b9-nnvmc\" (UID: \"ae4759b1-d998-49b4-8031-f638141177a5\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-nnvmc" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.623156 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7x6z\" (UniqueName: \"kubernetes.io/projected/4b6ce85f-d265-40a2-aa92-2a2d76339acf-kube-api-access-w7x6z\") pod \"cinder-operator-controller-manager-859b6ccc6-ccwrk\" (UID: \"4b6ce85f-d265-40a2-aa92-2a2d76339acf\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ccwrk" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.616690 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9zzgw" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.634984 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lh6jm"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.636966 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-r6hnf" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.663235 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4gcj\" (UniqueName: \"kubernetes.io/projected/322588d9-4c11-4676-8eb4-15cce77ccd44-kube-api-access-l4gcj\") pod \"designate-operator-controller-manager-78b4bc895b-7bcjp\" (UID: \"322588d9-4c11-4676-8eb4-15cce77ccd44\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-7bcjp" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.663895 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7x6z\" (UniqueName: \"kubernetes.io/projected/4b6ce85f-d265-40a2-aa92-2a2d76339acf-kube-api-access-w7x6z\") pod \"cinder-operator-controller-manager-859b6ccc6-ccwrk\" (UID: \"4b6ce85f-d265-40a2-aa92-2a2d76339acf\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ccwrk" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.664417 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgrjm\" (UniqueName: \"kubernetes.io/projected/308f7f89-2d44-42cb-8533-2b82306341c6-kube-api-access-pgrjm\") pod \"barbican-operator-controller-manager-7d9dfd778-4zjxs\" (UID: \"308f7f89-2d44-42cb-8533-2b82306341c6\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4zjxs" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.684982 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-2lp78"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.686184 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.688617 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vmxf\" (UniqueName: \"kubernetes.io/projected/ae4759b1-d998-49b4-8031-f638141177a5-kube-api-access-2vmxf\") pod \"glance-operator-controller-manager-668d9c48b9-nnvmc\" (UID: \"ae4759b1-d998-49b4-8031-f638141177a5\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-nnvmc" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.688773 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.688825 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-cd5br" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.694177 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-4zjxs"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.704862 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9zzgw"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.711055 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-2lp78"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.717783 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-gvx8g"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.719857 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-gvx8g" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.723423 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-b6lmg" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.724416 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h87lm\" (UniqueName: \"kubernetes.io/projected/fc17d065-8168-486a-ba45-1fd0cd31d9f0-kube-api-access-h87lm\") pod \"heat-operator-controller-manager-5f64f6f8bb-lh6jm\" (UID: \"fc17d065-8168-486a-ba45-1fd0cd31d9f0\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lh6jm" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.725210 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-hpln8"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.726460 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpln8" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.726573 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4zjxs" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.728056 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-q7j6j" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.742815 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ccwrk" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.747298 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-hpln8"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.756887 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-gvx8g"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.757228 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-7bcjp" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.765686 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-wss6v"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.766991 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-wss6v" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.769978 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-8lcnf" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.778310 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-nnvmc" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.787896 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-wss6v"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.801144 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lrsc2"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.802214 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lrsc2" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.805854 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-7zbzp" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.816440 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-72tnh"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.817768 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-72tnh" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.824832 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-2hn5b" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.826234 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h87lm\" (UniqueName: \"kubernetes.io/projected/fc17d065-8168-486a-ba45-1fd0cd31d9f0-kube-api-access-h87lm\") pod \"heat-operator-controller-manager-5f64f6f8bb-lh6jm\" (UID: \"fc17d065-8168-486a-ba45-1fd0cd31d9f0\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lh6jm" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.826383 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5ctk\" (UniqueName: \"kubernetes.io/projected/573876c9-548b-477b-8d92-b7b7d50c7bb5-kube-api-access-d5ctk\") pod \"ironic-operator-controller-manager-6c548fd776-hpln8\" (UID: \"573876c9-548b-477b-8d92-b7b7d50c7bb5\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpln8" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.826523 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7bcb\" (UniqueName: \"kubernetes.io/projected/50c4dc84-7965-4b16-8935-f557dc0c3a74-kube-api-access-r7bcb\") pod \"infra-operator-controller-manager-57548d458d-2lp78\" (UID: \"50c4dc84-7965-4b16-8935-f557dc0c3a74\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.826648 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwccn\" (UniqueName: \"kubernetes.io/projected/b6329d13-284d-48cc-866e-57744f0826bf-kube-api-access-lwccn\") pod \"horizon-operator-controller-manager-68c6d99b8f-9zzgw\" (UID: \"b6329d13-284d-48cc-866e-57744f0826bf\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9zzgw" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.826800 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k89qh\" (UniqueName: \"kubernetes.io/projected/e6700584-b120-491c-8920-caa3332ba444-kube-api-access-k89qh\") pod \"keystone-operator-controller-manager-546d4bdf48-gvx8g\" (UID: \"e6700584-b120-491c-8920-caa3332ba444\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-gvx8g" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.826915 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/50c4dc84-7965-4b16-8935-f557dc0c3a74-cert\") pod \"infra-operator-controller-manager-57548d458d-2lp78\" (UID: \"50c4dc84-7965-4b16-8935-f557dc0c3a74\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.840903 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lrsc2"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.860049 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h87lm\" (UniqueName: \"kubernetes.io/projected/fc17d065-8168-486a-ba45-1fd0cd31d9f0-kube-api-access-h87lm\") pod \"heat-operator-controller-manager-5f64f6f8bb-lh6jm\" (UID: \"fc17d065-8168-486a-ba45-1fd0cd31d9f0\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lh6jm" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.860390 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-72tnh"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.869358 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-x2png"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.871165 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x2png" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.883407 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-bzlwj" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.884653 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-x2png"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.892731 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-9pbbh"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.893968 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9pbbh" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.912366 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-gqgq2" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.926526 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-9pbbh"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.931501 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvwf4\" (UniqueName: \"kubernetes.io/projected/b56ca08d-173a-48a4-a537-8b9e84b4b2a2-kube-api-access-pvwf4\") pod \"manila-operator-controller-manager-6546668bfd-wss6v\" (UID: \"b56ca08d-173a-48a4-a537-8b9e84b4b2a2\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-wss6v" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.931586 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k89qh\" (UniqueName: \"kubernetes.io/projected/e6700584-b120-491c-8920-caa3332ba444-kube-api-access-k89qh\") pod \"keystone-operator-controller-manager-546d4bdf48-gvx8g\" (UID: \"e6700584-b120-491c-8920-caa3332ba444\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-gvx8g" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.931643 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/50c4dc84-7965-4b16-8935-f557dc0c3a74-cert\") pod \"infra-operator-controller-manager-57548d458d-2lp78\" (UID: \"50c4dc84-7965-4b16-8935-f557dc0c3a74\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.931698 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47g4f\" (UniqueName: \"kubernetes.io/projected/27e828db-c91c-49ff-88f2-eaab1ea075a0-kube-api-access-47g4f\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-lrsc2\" (UID: \"27e828db-c91c-49ff-88f2-eaab1ea075a0\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lrsc2" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.931800 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8wsc\" (UniqueName: \"kubernetes.io/projected/e67d0431-4b18-4743-842d-d0728b23727a-kube-api-access-k8wsc\") pod \"mariadb-operator-controller-manager-56bbcc9d85-72tnh\" (UID: \"e67d0431-4b18-4743-842d-d0728b23727a\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-72tnh" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.931829 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5ctk\" (UniqueName: \"kubernetes.io/projected/573876c9-548b-477b-8d92-b7b7d50c7bb5-kube-api-access-d5ctk\") pod \"ironic-operator-controller-manager-6c548fd776-hpln8\" (UID: \"573876c9-548b-477b-8d92-b7b7d50c7bb5\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpln8" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.931866 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwccn\" (UniqueName: \"kubernetes.io/projected/b6329d13-284d-48cc-866e-57744f0826bf-kube-api-access-lwccn\") pod \"horizon-operator-controller-manager-68c6d99b8f-9zzgw\" (UID: \"b6329d13-284d-48cc-866e-57744f0826bf\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9zzgw" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.931892 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7bcb\" (UniqueName: \"kubernetes.io/projected/50c4dc84-7965-4b16-8935-f557dc0c3a74-kube-api-access-r7bcb\") pod \"infra-operator-controller-manager-57548d458d-2lp78\" (UID: \"50c4dc84-7965-4b16-8935-f557dc0c3a74\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" Dec 01 14:52:34 crc kubenswrapper[4810]: E1201 14:52:34.936328 4810 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 14:52:34 crc kubenswrapper[4810]: E1201 14:52:34.936458 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50c4dc84-7965-4b16-8935-f557dc0c3a74-cert podName:50c4dc84-7965-4b16-8935-f557dc0c3a74 nodeName:}" failed. No retries permitted until 2025-12-01 14:52:35.436441181 +0000 UTC m=+1121.199950784 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/50c4dc84-7965-4b16-8935-f557dc0c3a74-cert") pod "infra-operator-controller-manager-57548d458d-2lp78" (UID: "50c4dc84-7965-4b16-8935-f557dc0c3a74") : secret "infra-operator-webhook-server-cert" not found Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.971385 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5ctk\" (UniqueName: \"kubernetes.io/projected/573876c9-548b-477b-8d92-b7b7d50c7bb5-kube-api-access-d5ctk\") pod \"ironic-operator-controller-manager-6c548fd776-hpln8\" (UID: \"573876c9-548b-477b-8d92-b7b7d50c7bb5\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpln8" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.971549 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k89qh\" (UniqueName: \"kubernetes.io/projected/e6700584-b120-491c-8920-caa3332ba444-kube-api-access-k89qh\") pod \"keystone-operator-controller-manager-546d4bdf48-gvx8g\" (UID: \"e6700584-b120-491c-8920-caa3332ba444\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-gvx8g" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.973510 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.981091 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwccn\" (UniqueName: \"kubernetes.io/projected/b6329d13-284d-48cc-866e-57744f0826bf-kube-api-access-lwccn\") pod \"horizon-operator-controller-manager-68c6d99b8f-9zzgw\" (UID: \"b6329d13-284d-48cc-866e-57744f0826bf\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9zzgw" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.984681 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-788fn"] Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.985905 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-788fn" Dec 01 14:52:34 crc kubenswrapper[4810]: I1201 14:52:34.986369 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.003269 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7bcb\" (UniqueName: \"kubernetes.io/projected/50c4dc84-7965-4b16-8935-f557dc0c3a74-kube-api-access-r7bcb\") pod \"infra-operator-controller-manager-57548d458d-2lp78\" (UID: \"50c4dc84-7965-4b16-8935-f557dc0c3a74\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.014739 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-788fn"] Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.024307 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9zzgw" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.121436 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99"] Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.152900 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lh6jm" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.121533 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-gvx8g" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.140367 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8wsc\" (UniqueName: \"kubernetes.io/projected/e67d0431-4b18-4743-842d-d0728b23727a-kube-api-access-k8wsc\") pod \"mariadb-operator-controller-manager-56bbcc9d85-72tnh\" (UID: \"e67d0431-4b18-4743-842d-d0728b23727a\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-72tnh" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.155849 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cszs\" (UniqueName: \"kubernetes.io/projected/7798189d-af43-43aa-afd1-4257194bcb12-kube-api-access-2cszs\") pod \"nova-operator-controller-manager-697bc559fc-x2png\" (UID: \"7798189d-af43-43aa-afd1-4257194bcb12\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x2png" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.155911 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdrrx\" (UniqueName: \"kubernetes.io/projected/c88f9be0-7e6f-4021-86c2-a43b34012a2c-kube-api-access-gdrrx\") pod \"octavia-operator-controller-manager-998648c74-9pbbh\" (UID: \"c88f9be0-7e6f-4021-86c2-a43b34012a2c\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-9pbbh" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.155942 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvwf4\" (UniqueName: \"kubernetes.io/projected/b56ca08d-173a-48a4-a537-8b9e84b4b2a2-kube-api-access-pvwf4\") pod \"manila-operator-controller-manager-6546668bfd-wss6v\" (UID: \"b56ca08d-173a-48a4-a537-8b9e84b4b2a2\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-wss6v" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.156006 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47g4f\" (UniqueName: \"kubernetes.io/projected/27e828db-c91c-49ff-88f2-eaab1ea075a0-kube-api-access-47g4f\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-lrsc2\" (UID: \"27e828db-c91c-49ff-88f2-eaab1ea075a0\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lrsc2" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.158243 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpln8" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.183644 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-9ft8n" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.238645 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8wsc\" (UniqueName: \"kubernetes.io/projected/e67d0431-4b18-4743-842d-d0728b23727a-kube-api-access-k8wsc\") pod \"mariadb-operator-controller-manager-56bbcc9d85-72tnh\" (UID: \"e67d0431-4b18-4743-842d-d0728b23727a\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-72tnh" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.262192 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47g4f\" (UniqueName: \"kubernetes.io/projected/27e828db-c91c-49ff-88f2-eaab1ea075a0-kube-api-access-47g4f\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-lrsc2\" (UID: \"27e828db-c91c-49ff-88f2-eaab1ea075a0\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lrsc2" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.263827 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdrrx\" (UniqueName: \"kubernetes.io/projected/c88f9be0-7e6f-4021-86c2-a43b34012a2c-kube-api-access-gdrrx\") pod \"octavia-operator-controller-manager-998648c74-9pbbh\" (UID: \"c88f9be0-7e6f-4021-86c2-a43b34012a2c\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-9pbbh" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.183692 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.296203 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgd9x\" (UniqueName: \"kubernetes.io/projected/816177c4-0c9b-4826-8844-84db106ec51d-kube-api-access-pgd9x\") pod \"ovn-operator-controller-manager-b6456fdb6-788fn\" (UID: \"816177c4-0c9b-4826-8844-84db106ec51d\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-788fn" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.183738 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-w4msl" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.297632 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68116291-3829-4a92-aba9-c1fbb755c988-cert\") pod \"openstack-baremetal-operator-controller-manager-6698bcb44667k99\" (UID: \"68116291-3829-4a92-aba9-c1fbb755c988\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.297826 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj8jm\" (UniqueName: \"kubernetes.io/projected/68116291-3829-4a92-aba9-c1fbb755c988-kube-api-access-gj8jm\") pod \"openstack-baremetal-operator-controller-manager-6698bcb44667k99\" (UID: \"68116291-3829-4a92-aba9-c1fbb755c988\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.297905 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cszs\" (UniqueName: \"kubernetes.io/projected/7798189d-af43-43aa-afd1-4257194bcb12-kube-api-access-2cszs\") pod \"nova-operator-controller-manager-697bc559fc-x2png\" (UID: \"7798189d-af43-43aa-afd1-4257194bcb12\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x2png" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.298574 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-c82g8"] Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.306789 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvwf4\" (UniqueName: \"kubernetes.io/projected/b56ca08d-173a-48a4-a537-8b9e84b4b2a2-kube-api-access-pvwf4\") pod \"manila-operator-controller-manager-6546668bfd-wss6v\" (UID: \"b56ca08d-173a-48a4-a537-8b9e84b4b2a2\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-wss6v" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.316951 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdrrx\" (UniqueName: \"kubernetes.io/projected/c88f9be0-7e6f-4021-86c2-a43b34012a2c-kube-api-access-gdrrx\") pod \"octavia-operator-controller-manager-998648c74-9pbbh\" (UID: \"c88f9be0-7e6f-4021-86c2-a43b34012a2c\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-9pbbh" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.340699 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c82g8" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.351017 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cszs\" (UniqueName: \"kubernetes.io/projected/7798189d-af43-43aa-afd1-4257194bcb12-kube-api-access-2cszs\") pod \"nova-operator-controller-manager-697bc559fc-x2png\" (UID: \"7798189d-af43-43aa-afd1-4257194bcb12\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x2png" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.411460 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgd9x\" (UniqueName: \"kubernetes.io/projected/816177c4-0c9b-4826-8844-84db106ec51d-kube-api-access-pgd9x\") pod \"ovn-operator-controller-manager-b6456fdb6-788fn\" (UID: \"816177c4-0c9b-4826-8844-84db106ec51d\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-788fn" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.455691 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68116291-3829-4a92-aba9-c1fbb755c988-cert\") pod \"openstack-baremetal-operator-controller-manager-6698bcb44667k99\" (UID: \"68116291-3829-4a92-aba9-c1fbb755c988\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.455932 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj8jm\" (UniqueName: \"kubernetes.io/projected/68116291-3829-4a92-aba9-c1fbb755c988-kube-api-access-gj8jm\") pod \"openstack-baremetal-operator-controller-manager-6698bcb44667k99\" (UID: \"68116291-3829-4a92-aba9-c1fbb755c988\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.456364 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-2cnvz" Dec 01 14:52:35 crc kubenswrapper[4810]: E1201 14:52:35.456611 4810 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 14:52:35 crc kubenswrapper[4810]: E1201 14:52:35.456698 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/68116291-3829-4a92-aba9-c1fbb755c988-cert podName:68116291-3829-4a92-aba9-c1fbb755c988 nodeName:}" failed. No retries permitted until 2025-12-01 14:52:35.956649026 +0000 UTC m=+1121.720158629 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/68116291-3829-4a92-aba9-c1fbb755c988-cert") pod "openstack-baremetal-operator-controller-manager-6698bcb44667k99" (UID: "68116291-3829-4a92-aba9-c1fbb755c988") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.457634 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-z85kg"] Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.459114 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-z85kg" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.463027 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-smvl6" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.528833 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-wss6v" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.551133 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lrsc2" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.555406 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-72tnh" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.557504 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlmvv\" (UniqueName: \"kubernetes.io/projected/7007dd52-d7df-4f27-81d0-95ee762c87c1-kube-api-access-wlmvv\") pod \"placement-operator-controller-manager-78f8948974-c82g8\" (UID: \"7007dd52-d7df-4f27-81d0-95ee762c87c1\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-c82g8" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.557584 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/50c4dc84-7965-4b16-8935-f557dc0c3a74-cert\") pod \"infra-operator-controller-manager-57548d458d-2lp78\" (UID: \"50c4dc84-7965-4b16-8935-f557dc0c3a74\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" Dec 01 14:52:35 crc kubenswrapper[4810]: E1201 14:52:35.557717 4810 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 14:52:35 crc kubenswrapper[4810]: E1201 14:52:35.557761 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50c4dc84-7965-4b16-8935-f557dc0c3a74-cert podName:50c4dc84-7965-4b16-8935-f557dc0c3a74 nodeName:}" failed. No retries permitted until 2025-12-01 14:52:36.557746823 +0000 UTC m=+1122.321256416 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/50c4dc84-7965-4b16-8935-f557dc0c3a74-cert") pod "infra-operator-controller-manager-57548d458d-2lp78" (UID: "50c4dc84-7965-4b16-8935-f557dc0c3a74") : secret "infra-operator-webhook-server-cert" not found Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.580443 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9pbbh" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.592715 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-c82g8"] Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.597044 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x2png" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.599974 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj8jm\" (UniqueName: \"kubernetes.io/projected/68116291-3829-4a92-aba9-c1fbb755c988-kube-api-access-gj8jm\") pod \"openstack-baremetal-operator-controller-manager-6698bcb44667k99\" (UID: \"68116291-3829-4a92-aba9-c1fbb755c988\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.603909 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-z85kg"] Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.604785 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgd9x\" (UniqueName: \"kubernetes.io/projected/816177c4-0c9b-4826-8844-84db106ec51d-kube-api-access-pgd9x\") pod \"ovn-operator-controller-manager-b6456fdb6-788fn\" (UID: \"816177c4-0c9b-4826-8844-84db106ec51d\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-788fn" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.609269 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-92m6z"] Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.610150 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-788fn" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.610320 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-92m6z" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.634862 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-ts5fg" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.662150 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-92m6z"] Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.663563 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n44jh\" (UniqueName: \"kubernetes.io/projected/58be13b4-ea8d-4ec2-8421-18d3fb2f744d-kube-api-access-n44jh\") pod \"telemetry-operator-controller-manager-76cc84c6bb-92m6z\" (UID: \"58be13b4-ea8d-4ec2-8421-18d3fb2f744d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-92m6z" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.663645 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpdtv\" (UniqueName: \"kubernetes.io/projected/19093fac-7ee4-43aa-ad57-7561d75de41f-kube-api-access-wpdtv\") pod \"swift-operator-controller-manager-5f8c65bbfc-z85kg\" (UID: \"19093fac-7ee4-43aa-ad57-7561d75de41f\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-z85kg" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.663687 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlmvv\" (UniqueName: \"kubernetes.io/projected/7007dd52-d7df-4f27-81d0-95ee762c87c1-kube-api-access-wlmvv\") pod \"placement-operator-controller-manager-78f8948974-c82g8\" (UID: \"7007dd52-d7df-4f27-81d0-95ee762c87c1\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-c82g8" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.711099 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-fs6gw"] Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.712172 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fs6gw" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.745267 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-vf79b" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.767563 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlmvv\" (UniqueName: \"kubernetes.io/projected/7007dd52-d7df-4f27-81d0-95ee762c87c1-kube-api-access-wlmvv\") pod \"placement-operator-controller-manager-78f8948974-c82g8\" (UID: \"7007dd52-d7df-4f27-81d0-95ee762c87c1\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-c82g8" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.768205 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n44jh\" (UniqueName: \"kubernetes.io/projected/58be13b4-ea8d-4ec2-8421-18d3fb2f744d-kube-api-access-n44jh\") pod \"telemetry-operator-controller-manager-76cc84c6bb-92m6z\" (UID: \"58be13b4-ea8d-4ec2-8421-18d3fb2f744d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-92m6z" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.768247 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpdtv\" (UniqueName: \"kubernetes.io/projected/19093fac-7ee4-43aa-ad57-7561d75de41f-kube-api-access-wpdtv\") pod \"swift-operator-controller-manager-5f8c65bbfc-z85kg\" (UID: \"19093fac-7ee4-43aa-ad57-7561d75de41f\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-z85kg" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.768293 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66cdv\" (UniqueName: \"kubernetes.io/projected/36f3dbad-9483-4695-95fe-7ce6520b4db3-kube-api-access-66cdv\") pod \"test-operator-controller-manager-5854674fcc-fs6gw\" (UID: \"36f3dbad-9483-4695-95fe-7ce6520b4db3\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-fs6gw" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.789621 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-fs6gw"] Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.811538 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c82g8" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.821595 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-8p7kt"] Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.822806 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8p7kt" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.832752 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-8p7kt"] Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.855152 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-ctxw4" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.869421 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhzgr\" (UniqueName: \"kubernetes.io/projected/dc03c0f9-e5d2-4fa3-98c6-d9a1d46638e0-kube-api-access-bhzgr\") pod \"watcher-operator-controller-manager-769dc69bc-8p7kt\" (UID: \"dc03c0f9-e5d2-4fa3-98c6-d9a1d46638e0\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8p7kt" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.869491 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66cdv\" (UniqueName: \"kubernetes.io/projected/36f3dbad-9483-4695-95fe-7ce6520b4db3-kube-api-access-66cdv\") pod \"test-operator-controller-manager-5854674fcc-fs6gw\" (UID: \"36f3dbad-9483-4695-95fe-7ce6520b4db3\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-fs6gw" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.935102 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2"] Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.935923 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.957233 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-4sf2w" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.957438 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.957581 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.970234 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhzgr\" (UniqueName: \"kubernetes.io/projected/dc03c0f9-e5d2-4fa3-98c6-d9a1d46638e0-kube-api-access-bhzgr\") pod \"watcher-operator-controller-manager-769dc69bc-8p7kt\" (UID: \"dc03c0f9-e5d2-4fa3-98c6-d9a1d46638e0\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8p7kt" Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.970322 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68116291-3829-4a92-aba9-c1fbb755c988-cert\") pod \"openstack-baremetal-operator-controller-manager-6698bcb44667k99\" (UID: \"68116291-3829-4a92-aba9-c1fbb755c988\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" Dec 01 14:52:35 crc kubenswrapper[4810]: E1201 14:52:35.970446 4810 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 14:52:35 crc kubenswrapper[4810]: E1201 14:52:35.970508 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/68116291-3829-4a92-aba9-c1fbb755c988-cert podName:68116291-3829-4a92-aba9-c1fbb755c988 nodeName:}" failed. No retries permitted until 2025-12-01 14:52:36.970493569 +0000 UTC m=+1122.734003172 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/68116291-3829-4a92-aba9-c1fbb755c988-cert") pod "openstack-baremetal-operator-controller-manager-6698bcb44667k99" (UID: "68116291-3829-4a92-aba9-c1fbb755c988") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 14:52:35 crc kubenswrapper[4810]: I1201 14:52:35.979070 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n44jh\" (UniqueName: \"kubernetes.io/projected/58be13b4-ea8d-4ec2-8421-18d3fb2f744d-kube-api-access-n44jh\") pod \"telemetry-operator-controller-manager-76cc84c6bb-92m6z\" (UID: \"58be13b4-ea8d-4ec2-8421-18d3fb2f744d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-92m6z" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.002776 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhzgr\" (UniqueName: \"kubernetes.io/projected/dc03c0f9-e5d2-4fa3-98c6-d9a1d46638e0-kube-api-access-bhzgr\") pod \"watcher-operator-controller-manager-769dc69bc-8p7kt\" (UID: \"dc03c0f9-e5d2-4fa3-98c6-d9a1d46638e0\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8p7kt" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.011309 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2"] Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.011989 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpdtv\" (UniqueName: \"kubernetes.io/projected/19093fac-7ee4-43aa-ad57-7561d75de41f-kube-api-access-wpdtv\") pod \"swift-operator-controller-manager-5f8c65bbfc-z85kg\" (UID: \"19093fac-7ee4-43aa-ad57-7561d75de41f\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-z85kg" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.017636 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66cdv\" (UniqueName: \"kubernetes.io/projected/36f3dbad-9483-4695-95fe-7ce6520b4db3-kube-api-access-66cdv\") pod \"test-operator-controller-manager-5854674fcc-fs6gw\" (UID: \"36f3dbad-9483-4695-95fe-7ce6520b4db3\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-fs6gw" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.069914 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8p7kt" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.071268 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-metrics-certs\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.071291 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-webhook-certs\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.071318 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbfk6\" (UniqueName: \"kubernetes.io/projected/2840e582-6c19-4473-9392-37a693d800dc-kube-api-access-hbfk6\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.122013 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xvf22"] Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.125041 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xvf22" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.127785 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-z85kg" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.129692 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-547qc" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.148038 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xvf22"] Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.160069 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-ccwrk"] Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.185392 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-metrics-certs\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.185460 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-webhook-certs\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.185538 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbfk6\" (UniqueName: \"kubernetes.io/projected/2840e582-6c19-4473-9392-37a693d800dc-kube-api-access-hbfk6\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:36 crc kubenswrapper[4810]: E1201 14:52:36.186585 4810 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 14:52:36 crc kubenswrapper[4810]: E1201 14:52:36.186666 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-metrics-certs podName:2840e582-6c19-4473-9392-37a693d800dc nodeName:}" failed. No retries permitted until 2025-12-01 14:52:36.686641271 +0000 UTC m=+1122.450150874 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-metrics-certs") pod "openstack-operator-controller-manager-656fd97d56-9ghs2" (UID: "2840e582-6c19-4473-9392-37a693d800dc") : secret "metrics-server-cert" not found Dec 01 14:52:36 crc kubenswrapper[4810]: E1201 14:52:36.187069 4810 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 14:52:36 crc kubenswrapper[4810]: E1201 14:52:36.187122 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-webhook-certs podName:2840e582-6c19-4473-9392-37a693d800dc nodeName:}" failed. No retries permitted until 2025-12-01 14:52:36.687107764 +0000 UTC m=+1122.450617367 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-webhook-certs") pod "openstack-operator-controller-manager-656fd97d56-9ghs2" (UID: "2840e582-6c19-4473-9392-37a693d800dc") : secret "webhook-server-cert" not found Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.187339 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-92m6z" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.249423 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbfk6\" (UniqueName: \"kubernetes.io/projected/2840e582-6c19-4473-9392-37a693d800dc-kube-api-access-hbfk6\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.283921 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fs6gw" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.289349 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xj4s\" (UniqueName: \"kubernetes.io/projected/380f8fee-59c6-445e-9a1e-47abe0eda51d-kube-api-access-7xj4s\") pod \"rabbitmq-cluster-operator-manager-668c99d594-xvf22\" (UID: \"380f8fee-59c6-445e-9a1e-47abe0eda51d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xvf22" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.330762 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-4zjxs"] Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.340096 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-nnvmc"] Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.392358 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xj4s\" (UniqueName: \"kubernetes.io/projected/380f8fee-59c6-445e-9a1e-47abe0eda51d-kube-api-access-7xj4s\") pod \"rabbitmq-cluster-operator-manager-668c99d594-xvf22\" (UID: \"380f8fee-59c6-445e-9a1e-47abe0eda51d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xvf22" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.437407 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xj4s\" (UniqueName: \"kubernetes.io/projected/380f8fee-59c6-445e-9a1e-47abe0eda51d-kube-api-access-7xj4s\") pod \"rabbitmq-cluster-operator-manager-668c99d594-xvf22\" (UID: \"380f8fee-59c6-445e-9a1e-47abe0eda51d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xvf22" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.595521 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xvf22" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.691802 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/50c4dc84-7965-4b16-8935-f557dc0c3a74-cert\") pod \"infra-operator-controller-manager-57548d458d-2lp78\" (UID: \"50c4dc84-7965-4b16-8935-f557dc0c3a74\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.691931 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-metrics-certs\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:36 crc kubenswrapper[4810]: I1201 14:52:36.691960 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-webhook-certs\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:36 crc kubenswrapper[4810]: E1201 14:52:36.696884 4810 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 14:52:36 crc kubenswrapper[4810]: E1201 14:52:36.696978 4810 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 14:52:36 crc kubenswrapper[4810]: E1201 14:52:36.697041 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-metrics-certs podName:2840e582-6c19-4473-9392-37a693d800dc nodeName:}" failed. No retries permitted until 2025-12-01 14:52:37.697014099 +0000 UTC m=+1123.460523702 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-metrics-certs") pod "openstack-operator-controller-manager-656fd97d56-9ghs2" (UID: "2840e582-6c19-4473-9392-37a693d800dc") : secret "metrics-server-cert" not found Dec 01 14:52:36 crc kubenswrapper[4810]: E1201 14:52:36.697154 4810 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 14:52:36 crc kubenswrapper[4810]: E1201 14:52:36.697180 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-webhook-certs podName:2840e582-6c19-4473-9392-37a693d800dc nodeName:}" failed. No retries permitted until 2025-12-01 14:52:37.697169024 +0000 UTC m=+1123.460678627 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-webhook-certs") pod "openstack-operator-controller-manager-656fd97d56-9ghs2" (UID: "2840e582-6c19-4473-9392-37a693d800dc") : secret "webhook-server-cert" not found Dec 01 14:52:36 crc kubenswrapper[4810]: E1201 14:52:36.697208 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50c4dc84-7965-4b16-8935-f557dc0c3a74-cert podName:50c4dc84-7965-4b16-8935-f557dc0c3a74 nodeName:}" failed. No retries permitted until 2025-12-01 14:52:38.697201005 +0000 UTC m=+1124.460710608 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/50c4dc84-7965-4b16-8935-f557dc0c3a74-cert") pod "infra-operator-controller-manager-57548d458d-2lp78" (UID: "50c4dc84-7965-4b16-8935-f557dc0c3a74") : secret "infra-operator-webhook-server-cert" not found Dec 01 14:52:37 crc kubenswrapper[4810]: I1201 14:52:37.066962 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68116291-3829-4a92-aba9-c1fbb755c988-cert\") pod \"openstack-baremetal-operator-controller-manager-6698bcb44667k99\" (UID: \"68116291-3829-4a92-aba9-c1fbb755c988\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" Dec 01 14:52:37 crc kubenswrapper[4810]: E1201 14:52:37.067179 4810 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 14:52:37 crc kubenswrapper[4810]: E1201 14:52:37.067228 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/68116291-3829-4a92-aba9-c1fbb755c988-cert podName:68116291-3829-4a92-aba9-c1fbb755c988 nodeName:}" failed. No retries permitted until 2025-12-01 14:52:39.067212598 +0000 UTC m=+1124.830722211 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/68116291-3829-4a92-aba9-c1fbb755c988-cert") pod "openstack-baremetal-operator-controller-manager-6698bcb44667k99" (UID: "68116291-3829-4a92-aba9-c1fbb755c988") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 14:52:37 crc kubenswrapper[4810]: I1201 14:52:37.078745 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4zjxs" event={"ID":"308f7f89-2d44-42cb-8533-2b82306341c6","Type":"ContainerStarted","Data":"3ad60dd70adcbb0ad669ef0f50114fa4d048ce31d3aed56f09365875fd5fa538"} Dec 01 14:52:37 crc kubenswrapper[4810]: I1201 14:52:37.086594 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ccwrk" event={"ID":"4b6ce85f-d265-40a2-aa92-2a2d76339acf","Type":"ContainerStarted","Data":"4c70a382c147ce8f1059f89a6a52007c997b0e2e09ab162f00b8862b771606e8"} Dec 01 14:52:37 crc kubenswrapper[4810]: I1201 14:52:37.088169 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-nnvmc" event={"ID":"ae4759b1-d998-49b4-8031-f638141177a5","Type":"ContainerStarted","Data":"8eb1816356e2248e220269ea7e4698654101038105f2b1c9de5faf0d763d2710"} Dec 01 14:52:37 crc kubenswrapper[4810]: I1201 14:52:37.784254 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-metrics-certs\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:37 crc kubenswrapper[4810]: I1201 14:52:37.785411 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-webhook-certs\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:37 crc kubenswrapper[4810]: E1201 14:52:37.784498 4810 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 14:52:37 crc kubenswrapper[4810]: E1201 14:52:37.785591 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-metrics-certs podName:2840e582-6c19-4473-9392-37a693d800dc nodeName:}" failed. No retries permitted until 2025-12-01 14:52:39.785557797 +0000 UTC m=+1125.549067400 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-metrics-certs") pod "openstack-operator-controller-manager-656fd97d56-9ghs2" (UID: "2840e582-6c19-4473-9392-37a693d800dc") : secret "metrics-server-cert" not found Dec 01 14:52:37 crc kubenswrapper[4810]: E1201 14:52:37.785803 4810 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 14:52:37 crc kubenswrapper[4810]: E1201 14:52:37.785837 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-webhook-certs podName:2840e582-6c19-4473-9392-37a693d800dc nodeName:}" failed. No retries permitted until 2025-12-01 14:52:39.785828855 +0000 UTC m=+1125.549338458 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-webhook-certs") pod "openstack-operator-controller-manager-656fd97d56-9ghs2" (UID: "2840e582-6c19-4473-9392-37a693d800dc") : secret "webhook-server-cert" not found Dec 01 14:52:38 crc kubenswrapper[4810]: I1201 14:52:38.188379 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-gvx8g"] Dec 01 14:52:38 crc kubenswrapper[4810]: I1201 14:52:38.367612 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9zzgw"] Dec 01 14:52:38 crc kubenswrapper[4810]: I1201 14:52:38.384762 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-9pbbh"] Dec 01 14:52:38 crc kubenswrapper[4810]: I1201 14:52:38.384820 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-7bcjp"] Dec 01 14:52:38 crc kubenswrapper[4810]: W1201 14:52:38.399232 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod322588d9_4c11_4676_8eb4_15cce77ccd44.slice/crio-c5f7025c6827e3afa619842d01cc1bc9e2f006dc041657aea587ac146054cbdf WatchSource:0}: Error finding container c5f7025c6827e3afa619842d01cc1bc9e2f006dc041657aea587ac146054cbdf: Status 404 returned error can't find the container with id c5f7025c6827e3afa619842d01cc1bc9e2f006dc041657aea587ac146054cbdf Dec 01 14:52:38 crc kubenswrapper[4810]: W1201 14:52:38.473907 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc88f9be0_7e6f_4021_86c2_a43b34012a2c.slice/crio-14f8c7e9e4d334bfe98d851127a4e19b8c9c85614adfeda6222c13994b5deda2 WatchSource:0}: Error finding container 14f8c7e9e4d334bfe98d851127a4e19b8c9c85614adfeda6222c13994b5deda2: Status 404 returned error can't find the container with id 14f8c7e9e4d334bfe98d851127a4e19b8c9c85614adfeda6222c13994b5deda2 Dec 01 14:52:38 crc kubenswrapper[4810]: I1201 14:52:38.550910 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-92m6z"] Dec 01 14:52:38 crc kubenswrapper[4810]: I1201 14:52:38.577761 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-hpln8"] Dec 01 14:52:38 crc kubenswrapper[4810]: I1201 14:52:38.584046 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-c82g8"] Dec 01 14:52:38 crc kubenswrapper[4810]: I1201 14:52:38.587463 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-788fn"] Dec 01 14:52:38 crc kubenswrapper[4810]: I1201 14:52:38.655203 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lh6jm"] Dec 01 14:52:38 crc kubenswrapper[4810]: W1201 14:52:38.665599 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod816177c4_0c9b_4826_8844_84db106ec51d.slice/crio-acc106540789037d65981c74a7d71722cecf88827810daecbc82bb48c6d3faab WatchSource:0}: Error finding container acc106540789037d65981c74a7d71722cecf88827810daecbc82bb48c6d3faab: Status 404 returned error can't find the container with id acc106540789037d65981c74a7d71722cecf88827810daecbc82bb48c6d3faab Dec 01 14:52:38 crc kubenswrapper[4810]: I1201 14:52:38.779606 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/50c4dc84-7965-4b16-8935-f557dc0c3a74-cert\") pod \"infra-operator-controller-manager-57548d458d-2lp78\" (UID: \"50c4dc84-7965-4b16-8935-f557dc0c3a74\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" Dec 01 14:52:38 crc kubenswrapper[4810]: E1201 14:52:38.779769 4810 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 14:52:38 crc kubenswrapper[4810]: E1201 14:52:38.779828 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50c4dc84-7965-4b16-8935-f557dc0c3a74-cert podName:50c4dc84-7965-4b16-8935-f557dc0c3a74 nodeName:}" failed. No retries permitted until 2025-12-01 14:52:42.779808954 +0000 UTC m=+1128.543318557 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/50c4dc84-7965-4b16-8935-f557dc0c3a74-cert") pod "infra-operator-controller-manager-57548d458d-2lp78" (UID: "50c4dc84-7965-4b16-8935-f557dc0c3a74") : secret "infra-operator-webhook-server-cert" not found Dec 01 14:52:38 crc kubenswrapper[4810]: I1201 14:52:38.937749 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-wss6v"] Dec 01 14:52:38 crc kubenswrapper[4810]: W1201 14:52:38.939051 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb56ca08d_173a_48a4_a537_8b9e84b4b2a2.slice/crio-58b4383a3918ef45efe825f58826e7145d391685e3e4880110df1c29194daf96 WatchSource:0}: Error finding container 58b4383a3918ef45efe825f58826e7145d391685e3e4880110df1c29194daf96: Status 404 returned error can't find the container with id 58b4383a3918ef45efe825f58826e7145d391685e3e4880110df1c29194daf96 Dec 01 14:52:38 crc kubenswrapper[4810]: I1201 14:52:38.952010 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-z85kg"] Dec 01 14:52:38 crc kubenswrapper[4810]: I1201 14:52:38.962770 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-72tnh"] Dec 01 14:52:38 crc kubenswrapper[4810]: I1201 14:52:38.984400 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-8p7kt"] Dec 01 14:52:38 crc kubenswrapper[4810]: W1201 14:52:38.985372 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode67d0431_4b18_4743_842d_d0728b23727a.slice/crio-11ce5a6483b775476c659c59e8051bb84c79f2a26b5396ecb31f90b80f5b892f WatchSource:0}: Error finding container 11ce5a6483b775476c659c59e8051bb84c79f2a26b5396ecb31f90b80f5b892f: Status 404 returned error can't find the container with id 11ce5a6483b775476c659c59e8051bb84c79f2a26b5396ecb31f90b80f5b892f Dec 01 14:52:38 crc kubenswrapper[4810]: W1201 14:52:38.999171 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc03c0f9_e5d2_4fa3_98c6_d9a1d46638e0.slice/crio-7ddefeab73956559f79305469ac828a1117a708babfcbd3208b210fd5f2fc71d WatchSource:0}: Error finding container 7ddefeab73956559f79305469ac828a1117a708babfcbd3208b210fd5f2fc71d: Status 404 returned error can't find the container with id 7ddefeab73956559f79305469ac828a1117a708babfcbd3208b210fd5f2fc71d Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.010297 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lrsc2"] Dec 01 14:52:39 crc kubenswrapper[4810]: W1201 14:52:39.022893 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27e828db_c91c_49ff_88f2_eaab1ea075a0.slice/crio-02a3d2b87762714bcee4c48ee24318b7f4b993310841954a4a29663a1011422c WatchSource:0}: Error finding container 02a3d2b87762714bcee4c48ee24318b7f4b993310841954a4a29663a1011422c: Status 404 returned error can't find the container with id 02a3d2b87762714bcee4c48ee24318b7f4b993310841954a4a29663a1011422c Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.031974 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-47g4f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-lrsc2_openstack-operators(27e828db-c91c-49ff-88f2-eaab1ea075a0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.034038 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-47g4f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-lrsc2_openstack-operators(27e828db-c91c-49ff-88f2-eaab1ea075a0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.035267 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lrsc2" podUID="27e828db-c91c-49ff-88f2-eaab1ea075a0" Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.058022 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xvf22"] Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.071388 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7xj4s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-xvf22_openstack-operators(380f8fee-59c6-445e-9a1e-47abe0eda51d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.072580 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xvf22" podUID="380f8fee-59c6-445e-9a1e-47abe0eda51d" Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.084041 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68116291-3829-4a92-aba9-c1fbb755c988-cert\") pod \"openstack-baremetal-operator-controller-manager-6698bcb44667k99\" (UID: \"68116291-3829-4a92-aba9-c1fbb755c988\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.084160 4810 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.084200 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/68116291-3829-4a92-aba9-c1fbb755c988-cert podName:68116291-3829-4a92-aba9-c1fbb755c988 nodeName:}" failed. No retries permitted until 2025-12-01 14:52:43.084188494 +0000 UTC m=+1128.847698097 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/68116291-3829-4a92-aba9-c1fbb755c988-cert") pod "openstack-baremetal-operator-controller-manager-6698bcb44667k99" (UID: "68116291-3829-4a92-aba9-c1fbb755c988") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.118800 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-fs6gw"] Dec 01 14:52:39 crc kubenswrapper[4810]: W1201 14:52:39.126673 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36f3dbad_9483_4695_95fe_7ce6520b4db3.slice/crio-28c738196b8a63bd8670c34133ff42668cdd0155542cf791aababb3d15a2e935 WatchSource:0}: Error finding container 28c738196b8a63bd8670c34133ff42668cdd0155542cf791aababb3d15a2e935: Status 404 returned error can't find the container with id 28c738196b8a63bd8670c34133ff42668cdd0155542cf791aababb3d15a2e935 Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.127019 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-x2png"] Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.130131 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-66cdv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-fs6gw_openstack-operators(36f3dbad-9483-4695-95fe-7ce6520b4db3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.132580 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-66cdv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-fs6gw_openstack-operators(36f3dbad-9483-4695-95fe-7ce6520b4db3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.133697 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fs6gw" podUID="36f3dbad-9483-4695-95fe-7ce6520b4db3" Dec 01 14:52:39 crc kubenswrapper[4810]: W1201 14:52:39.135229 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7798189d_af43_43aa_afd1_4257194bcb12.slice/crio-34e9f5f9a5395acaf64f700feaeac5843c2ff6c198850429973ff829ab29994c WatchSource:0}: Error finding container 34e9f5f9a5395acaf64f700feaeac5843c2ff6c198850429973ff829ab29994c: Status 404 returned error can't find the container with id 34e9f5f9a5395acaf64f700feaeac5843c2ff6c198850429973ff829ab29994c Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.140294 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2cszs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-x2png_openstack-operators(7798189d-af43-43aa-afd1-4257194bcb12): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.142285 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2cszs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-x2png_openstack-operators(7798189d-af43-43aa-afd1-4257194bcb12): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.143538 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x2png" podUID="7798189d-af43-43aa-afd1-4257194bcb12" Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.213632 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-7bcjp" event={"ID":"322588d9-4c11-4676-8eb4-15cce77ccd44","Type":"ContainerStarted","Data":"c5f7025c6827e3afa619842d01cc1bc9e2f006dc041657aea587ac146054cbdf"} Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.217811 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c82g8" event={"ID":"7007dd52-d7df-4f27-81d0-95ee762c87c1","Type":"ContainerStarted","Data":"fb9f2f645ac119260d971657ebc4ef7d3435d8f807a030e292991ec4c406adee"} Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.219755 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9zzgw" event={"ID":"b6329d13-284d-48cc-866e-57744f0826bf","Type":"ContainerStarted","Data":"17c7e5cb68108656edcd836ed44a94ced440ba333b5c3b50163f29165f1f7e1b"} Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.224258 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-z85kg" event={"ID":"19093fac-7ee4-43aa-ad57-7561d75de41f","Type":"ContainerStarted","Data":"e33ab258c7797a1400a917cab2bb200cc3c7b2da486128d649c0ed62b4f9e4e6"} Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.225198 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8p7kt" event={"ID":"dc03c0f9-e5d2-4fa3-98c6-d9a1d46638e0","Type":"ContainerStarted","Data":"7ddefeab73956559f79305469ac828a1117a708babfcbd3208b210fd5f2fc71d"} Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.226358 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lrsc2" event={"ID":"27e828db-c91c-49ff-88f2-eaab1ea075a0","Type":"ContainerStarted","Data":"02a3d2b87762714bcee4c48ee24318b7f4b993310841954a4a29663a1011422c"} Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.228053 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lrsc2" podUID="27e828db-c91c-49ff-88f2-eaab1ea075a0" Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.228936 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9pbbh" event={"ID":"c88f9be0-7e6f-4021-86c2-a43b34012a2c","Type":"ContainerStarted","Data":"14f8c7e9e4d334bfe98d851127a4e19b8c9c85614adfeda6222c13994b5deda2"} Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.232684 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-788fn" event={"ID":"816177c4-0c9b-4826-8844-84db106ec51d","Type":"ContainerStarted","Data":"acc106540789037d65981c74a7d71722cecf88827810daecbc82bb48c6d3faab"} Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.234871 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fs6gw" event={"ID":"36f3dbad-9483-4695-95fe-7ce6520b4db3","Type":"ContainerStarted","Data":"28c738196b8a63bd8670c34133ff42668cdd0155542cf791aababb3d15a2e935"} Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.236747 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xvf22" event={"ID":"380f8fee-59c6-445e-9a1e-47abe0eda51d","Type":"ContainerStarted","Data":"5c96644ce0905088d8862174a9b0d6de2214b66cc1bb504c2db3882ad8d8a72f"} Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.237174 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fs6gw" podUID="36f3dbad-9483-4695-95fe-7ce6520b4db3" Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.238488 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-gvx8g" event={"ID":"e6700584-b120-491c-8920-caa3332ba444","Type":"ContainerStarted","Data":"dc5aeee5d2b6e4205aa10e16954140f353577d8a4eedd5f80ac4aad20ecb32a8"} Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.238958 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xvf22" podUID="380f8fee-59c6-445e-9a1e-47abe0eda51d" Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.242676 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-wss6v" event={"ID":"b56ca08d-173a-48a4-a537-8b9e84b4b2a2","Type":"ContainerStarted","Data":"58b4383a3918ef45efe825f58826e7145d391685e3e4880110df1c29194daf96"} Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.248788 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpln8" event={"ID":"573876c9-548b-477b-8d92-b7b7d50c7bb5","Type":"ContainerStarted","Data":"011159a23fcccfaf114b07af80573400ddf41af981e865c85975927e85e876b8"} Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.251069 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-92m6z" event={"ID":"58be13b4-ea8d-4ec2-8421-18d3fb2f744d","Type":"ContainerStarted","Data":"1b0c68c0d3ccfe31d125ac2a00d94827319ab84649194ecdd0ed288693e4b20c"} Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.253023 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x2png" event={"ID":"7798189d-af43-43aa-afd1-4257194bcb12","Type":"ContainerStarted","Data":"34e9f5f9a5395acaf64f700feaeac5843c2ff6c198850429973ff829ab29994c"} Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.254254 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-72tnh" event={"ID":"e67d0431-4b18-4743-842d-d0728b23727a","Type":"ContainerStarted","Data":"11ce5a6483b775476c659c59e8051bb84c79f2a26b5396ecb31f90b80f5b892f"} Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.255881 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lh6jm" event={"ID":"fc17d065-8168-486a-ba45-1fd0cd31d9f0","Type":"ContainerStarted","Data":"50f0546eefab4c1e85b6ee12bb651bd92e40a0936acb50d82289dac0f0aa638b"} Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.257896 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x2png" podUID="7798189d-af43-43aa-afd1-4257194bcb12" Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.800757 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-metrics-certs\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:39 crc kubenswrapper[4810]: I1201 14:52:39.800823 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-webhook-certs\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.800968 4810 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.801062 4810 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.801095 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-metrics-certs podName:2840e582-6c19-4473-9392-37a693d800dc nodeName:}" failed. No retries permitted until 2025-12-01 14:52:43.801070523 +0000 UTC m=+1129.564580176 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-metrics-certs") pod "openstack-operator-controller-manager-656fd97d56-9ghs2" (UID: "2840e582-6c19-4473-9392-37a693d800dc") : secret "metrics-server-cert" not found Dec 01 14:52:39 crc kubenswrapper[4810]: E1201 14:52:39.801152 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-webhook-certs podName:2840e582-6c19-4473-9392-37a693d800dc nodeName:}" failed. No retries permitted until 2025-12-01 14:52:43.801127854 +0000 UTC m=+1129.564637527 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-webhook-certs") pod "openstack-operator-controller-manager-656fd97d56-9ghs2" (UID: "2840e582-6c19-4473-9392-37a693d800dc") : secret "webhook-server-cert" not found Dec 01 14:52:40 crc kubenswrapper[4810]: E1201 14:52:40.268300 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xvf22" podUID="380f8fee-59c6-445e-9a1e-47abe0eda51d" Dec 01 14:52:40 crc kubenswrapper[4810]: E1201 14:52:40.269023 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lrsc2" podUID="27e828db-c91c-49ff-88f2-eaab1ea075a0" Dec 01 14:52:40 crc kubenswrapper[4810]: E1201 14:52:40.269160 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x2png" podUID="7798189d-af43-43aa-afd1-4257194bcb12" Dec 01 14:52:40 crc kubenswrapper[4810]: E1201 14:52:40.269233 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fs6gw" podUID="36f3dbad-9483-4695-95fe-7ce6520b4db3" Dec 01 14:52:42 crc kubenswrapper[4810]: I1201 14:52:42.786967 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/50c4dc84-7965-4b16-8935-f557dc0c3a74-cert\") pod \"infra-operator-controller-manager-57548d458d-2lp78\" (UID: \"50c4dc84-7965-4b16-8935-f557dc0c3a74\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" Dec 01 14:52:42 crc kubenswrapper[4810]: E1201 14:52:42.787177 4810 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 14:52:42 crc kubenswrapper[4810]: E1201 14:52:42.787499 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50c4dc84-7965-4b16-8935-f557dc0c3a74-cert podName:50c4dc84-7965-4b16-8935-f557dc0c3a74 nodeName:}" failed. No retries permitted until 2025-12-01 14:52:50.78746252 +0000 UTC m=+1136.550972123 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/50c4dc84-7965-4b16-8935-f557dc0c3a74-cert") pod "infra-operator-controller-manager-57548d458d-2lp78" (UID: "50c4dc84-7965-4b16-8935-f557dc0c3a74") : secret "infra-operator-webhook-server-cert" not found Dec 01 14:52:43 crc kubenswrapper[4810]: I1201 14:52:43.101259 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68116291-3829-4a92-aba9-c1fbb755c988-cert\") pod \"openstack-baremetal-operator-controller-manager-6698bcb44667k99\" (UID: \"68116291-3829-4a92-aba9-c1fbb755c988\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" Dec 01 14:52:43 crc kubenswrapper[4810]: E1201 14:52:43.101435 4810 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 14:52:43 crc kubenswrapper[4810]: E1201 14:52:43.101537 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/68116291-3829-4a92-aba9-c1fbb755c988-cert podName:68116291-3829-4a92-aba9-c1fbb755c988 nodeName:}" failed. No retries permitted until 2025-12-01 14:52:51.101514344 +0000 UTC m=+1136.865023947 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/68116291-3829-4a92-aba9-c1fbb755c988-cert") pod "openstack-baremetal-operator-controller-manager-6698bcb44667k99" (UID: "68116291-3829-4a92-aba9-c1fbb755c988") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 14:52:43 crc kubenswrapper[4810]: I1201 14:52:43.864593 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-metrics-certs\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:43 crc kubenswrapper[4810]: I1201 14:52:43.864639 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-webhook-certs\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:43 crc kubenswrapper[4810]: E1201 14:52:43.864774 4810 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 14:52:43 crc kubenswrapper[4810]: E1201 14:52:43.864901 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-metrics-certs podName:2840e582-6c19-4473-9392-37a693d800dc nodeName:}" failed. No retries permitted until 2025-12-01 14:52:51.864872136 +0000 UTC m=+1137.628381749 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-metrics-certs") pod "openstack-operator-controller-manager-656fd97d56-9ghs2" (UID: "2840e582-6c19-4473-9392-37a693d800dc") : secret "metrics-server-cert" not found Dec 01 14:52:43 crc kubenswrapper[4810]: E1201 14:52:43.864788 4810 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 14:52:43 crc kubenswrapper[4810]: E1201 14:52:43.864966 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-webhook-certs podName:2840e582-6c19-4473-9392-37a693d800dc nodeName:}" failed. No retries permitted until 2025-12-01 14:52:51.864947688 +0000 UTC m=+1137.628457291 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-webhook-certs") pod "openstack-operator-controller-manager-656fd97d56-9ghs2" (UID: "2840e582-6c19-4473-9392-37a693d800dc") : secret "webhook-server-cert" not found Dec 01 14:52:50 crc kubenswrapper[4810]: I1201 14:52:50.801463 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/50c4dc84-7965-4b16-8935-f557dc0c3a74-cert\") pod \"infra-operator-controller-manager-57548d458d-2lp78\" (UID: \"50c4dc84-7965-4b16-8935-f557dc0c3a74\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" Dec 01 14:52:50 crc kubenswrapper[4810]: I1201 14:52:50.806889 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/50c4dc84-7965-4b16-8935-f557dc0c3a74-cert\") pod \"infra-operator-controller-manager-57548d458d-2lp78\" (UID: \"50c4dc84-7965-4b16-8935-f557dc0c3a74\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" Dec 01 14:52:50 crc kubenswrapper[4810]: I1201 14:52:50.945310 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" Dec 01 14:52:51 crc kubenswrapper[4810]: I1201 14:52:51.107089 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68116291-3829-4a92-aba9-c1fbb755c988-cert\") pod \"openstack-baremetal-operator-controller-manager-6698bcb44667k99\" (UID: \"68116291-3829-4a92-aba9-c1fbb755c988\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" Dec 01 14:52:51 crc kubenswrapper[4810]: I1201 14:52:51.110526 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/68116291-3829-4a92-aba9-c1fbb755c988-cert\") pod \"openstack-baremetal-operator-controller-manager-6698bcb44667k99\" (UID: \"68116291-3829-4a92-aba9-c1fbb755c988\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" Dec 01 14:52:51 crc kubenswrapper[4810]: I1201 14:52:51.225506 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" Dec 01 14:52:51 crc kubenswrapper[4810]: I1201 14:52:51.918269 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-metrics-certs\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:51 crc kubenswrapper[4810]: I1201 14:52:51.918693 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-webhook-certs\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:51 crc kubenswrapper[4810]: I1201 14:52:51.924251 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-metrics-certs\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:51 crc kubenswrapper[4810]: I1201 14:52:51.924693 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2840e582-6c19-4473-9392-37a693d800dc-webhook-certs\") pod \"openstack-operator-controller-manager-656fd97d56-9ghs2\" (UID: \"2840e582-6c19-4473-9392-37a693d800dc\") " pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:52 crc kubenswrapper[4810]: I1201 14:52:52.006382 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:52:53 crc kubenswrapper[4810]: E1201 14:52:53.346716 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801" Dec 01 14:52:53 crc kubenswrapper[4810]: E1201 14:52:53.346929 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w7x6z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-ccwrk_openstack-operators(4b6ce85f-d265-40a2-aa92-2a2d76339acf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:52:53 crc kubenswrapper[4810]: E1201 14:52:53.925414 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530" Dec 01 14:52:53 crc kubenswrapper[4810]: E1201 14:52:53.925622 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d5ctk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-hpln8_openstack-operators(573876c9-548b-477b-8d92-b7b7d50c7bb5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:52:54 crc kubenswrapper[4810]: E1201 14:52:54.597687 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621" Dec 01 14:52:54 crc kubenswrapper[4810]: E1201 14:52:54.598047 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bhzgr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-8p7kt_openstack-operators(dc03c0f9-e5d2-4fa3-98c6-d9a1d46638e0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:52:55 crc kubenswrapper[4810]: E1201 14:52:55.215986 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 01 14:52:55 crc kubenswrapper[4810]: E1201 14:52:55.216276 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lwccn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-9zzgw_openstack-operators(b6329d13-284d-48cc-866e-57744f0826bf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:52:55 crc kubenswrapper[4810]: E1201 14:52:55.770637 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 01 14:52:55 crc kubenswrapper[4810]: E1201 14:52:55.770848 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wlmvv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-c82g8_openstack-operators(7007dd52-d7df-4f27-81d0-95ee762c87c1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:52:56 crc kubenswrapper[4810]: E1201 14:52:56.663197 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 01 14:52:56 crc kubenswrapper[4810]: E1201 14:52:56.663374 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gdrrx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-9pbbh_openstack-operators(c88f9be0-7e6f-4021-86c2-a43b34012a2c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:52:57 crc kubenswrapper[4810]: E1201 14:52:57.185372 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d" Dec 01 14:52:57 crc kubenswrapper[4810]: E1201 14:52:57.185577 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wpdtv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-z85kg_openstack-operators(19093fac-7ee4-43aa-ad57-7561d75de41f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:52:57 crc kubenswrapper[4810]: E1201 14:52:57.711401 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385" Dec 01 14:52:57 crc kubenswrapper[4810]: E1201 14:52:57.711612 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n44jh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-92m6z_openstack-operators(58be13b4-ea8d-4ec2-8421-18d3fb2f744d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:52:57 crc kubenswrapper[4810]: I1201 14:52:57.714206 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 14:52:58 crc kubenswrapper[4810]: E1201 14:52:58.300960 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 01 14:52:58 crc kubenswrapper[4810]: E1201 14:52:58.301922 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pgd9x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-788fn_openstack-operators(816177c4-0c9b-4826-8844-84db106ec51d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:52:58 crc kubenswrapper[4810]: E1201 14:52:58.811616 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7" Dec 01 14:52:58 crc kubenswrapper[4810]: E1201 14:52:58.812022 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k8wsc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-72tnh_openstack-operators(e67d0431-4b18-4743-842d-d0728b23727a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:52:59 crc kubenswrapper[4810]: E1201 14:52:59.477802 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:ecf7be921850bdc04697ed1b332bab39ad2a64e4e45c2a445c04f9bae6ac61b5" Dec 01 14:52:59 crc kubenswrapper[4810]: E1201 14:52:59.477969 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:ecf7be921850bdc04697ed1b332bab39ad2a64e4e45c2a445c04f9bae6ac61b5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pvwf4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-6546668bfd-wss6v_openstack-operators(b56ca08d-173a-48a4-a537-8b9e84b4b2a2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:52:59 crc kubenswrapper[4810]: E1201 14:52:59.969332 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85" Dec 01 14:52:59 crc kubenswrapper[4810]: E1201 14:52:59.969577 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l4gcj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-7bcjp_openstack-operators(322588d9-4c11-4676-8eb4-15cce77ccd44): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:53:00 crc kubenswrapper[4810]: E1201 14:53:00.404359 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:986861e5a0a9954f63581d9d55a30f8057883cefea489415d76257774526eea3" Dec 01 14:53:00 crc kubenswrapper[4810]: E1201 14:53:00.404643 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:986861e5a0a9954f63581d9d55a30f8057883cefea489415d76257774526eea3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k89qh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-546d4bdf48-gvx8g_openstack-operators(e6700584-b120-491c-8920-caa3332ba444): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:53:00 crc kubenswrapper[4810]: E1201 14:53:00.874332 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 01 14:53:00 crc kubenswrapper[4810]: E1201 14:53:00.874518 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-h87lm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-lh6jm_openstack-operators(fc17d065-8168-486a-ba45-1fd0cd31d9f0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:53:03 crc kubenswrapper[4810]: I1201 14:53:03.591549 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-2lp78"] Dec 01 14:53:04 crc kubenswrapper[4810]: I1201 14:53:04.157230 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99"] Dec 01 14:53:04 crc kubenswrapper[4810]: I1201 14:53:04.230208 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2"] Dec 01 14:53:04 crc kubenswrapper[4810]: I1201 14:53:04.604761 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4zjxs" event={"ID":"308f7f89-2d44-42cb-8533-2b82306341c6","Type":"ContainerStarted","Data":"db3b357a70e890c52b3c769280e8f5d406807c5a7e34cfaae67670be2b2cf9e1"} Dec 01 14:53:04 crc kubenswrapper[4810]: I1201 14:53:04.606040 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x2png" event={"ID":"7798189d-af43-43aa-afd1-4257194bcb12","Type":"ContainerStarted","Data":"9de78ae53702865a61cf79ada47fd97da5f18d02ad57bfd3fb5d585a61471538"} Dec 01 14:53:04 crc kubenswrapper[4810]: I1201 14:53:04.607273 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lrsc2" event={"ID":"27e828db-c91c-49ff-88f2-eaab1ea075a0","Type":"ContainerStarted","Data":"2dba00abddcbfaf09fe7fa0290ae9b44554a70bf92b7df3d3cbc11bc4376fc02"} Dec 01 14:53:04 crc kubenswrapper[4810]: I1201 14:53:04.608301 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" event={"ID":"50c4dc84-7965-4b16-8935-f557dc0c3a74","Type":"ContainerStarted","Data":"28b5d1c737511503aa4eaf936b45bc0e25df5b9f7487eb33b028d1b03f930baa"} Dec 01 14:53:04 crc kubenswrapper[4810]: I1201 14:53:04.609673 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-nnvmc" event={"ID":"ae4759b1-d998-49b4-8031-f638141177a5","Type":"ContainerStarted","Data":"a8e0a5bfdb0fd7fc2143e8d4c75df87faa8f265a610f486187e5d08aeeab22bd"} Dec 01 14:53:04 crc kubenswrapper[4810]: I1201 14:53:04.610937 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fs6gw" event={"ID":"36f3dbad-9483-4695-95fe-7ce6520b4db3","Type":"ContainerStarted","Data":"eeab3c8e5e2392396f98c60dbd757ee1050c2e7775348581d0f8aa61bf269994"} Dec 01 14:53:04 crc kubenswrapper[4810]: W1201 14:53:04.651659 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68116291_3829_4a92_aba9_c1fbb755c988.slice/crio-18b1ef0813ba564ffcdb52241b932bbe270ce4d06a2c7890480d9c6c48a5b161 WatchSource:0}: Error finding container 18b1ef0813ba564ffcdb52241b932bbe270ce4d06a2c7890480d9c6c48a5b161: Status 404 returned error can't find the container with id 18b1ef0813ba564ffcdb52241b932bbe270ce4d06a2c7890480d9c6c48a5b161 Dec 01 14:53:05 crc kubenswrapper[4810]: I1201 14:53:05.627068 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" event={"ID":"2840e582-6c19-4473-9392-37a693d800dc","Type":"ContainerStarted","Data":"e66a7c0859268bef72b5155955d6b665fe77c5d76411112a387434e3d5564074"} Dec 01 14:53:05 crc kubenswrapper[4810]: I1201 14:53:05.627485 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:53:05 crc kubenswrapper[4810]: I1201 14:53:05.627502 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" event={"ID":"2840e582-6c19-4473-9392-37a693d800dc","Type":"ContainerStarted","Data":"e5140b02f63ed369d449eed01417d0a2dec8edb62905331811d3a3b6f5f7e28e"} Dec 01 14:53:05 crc kubenswrapper[4810]: I1201 14:53:05.632301 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xvf22" event={"ID":"380f8fee-59c6-445e-9a1e-47abe0eda51d","Type":"ContainerStarted","Data":"4361ff15aa37f26ca1e4cb2c32f841945a0242224bfbb4d79824d91b52e4b00a"} Dec 01 14:53:05 crc kubenswrapper[4810]: I1201 14:53:05.634012 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" event={"ID":"68116291-3829-4a92-aba9-c1fbb755c988","Type":"ContainerStarted","Data":"18b1ef0813ba564ffcdb52241b932bbe270ce4d06a2c7890480d9c6c48a5b161"} Dec 01 14:53:05 crc kubenswrapper[4810]: I1201 14:53:05.674238 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" podStartSLOduration=30.674219582 podStartE2EDuration="30.674219582s" podCreationTimestamp="2025-12-01 14:52:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:53:05.666921244 +0000 UTC m=+1151.430430847" watchObservedRunningTime="2025-12-01 14:53:05.674219582 +0000 UTC m=+1151.437729185" Dec 01 14:53:05 crc kubenswrapper[4810]: I1201 14:53:05.688352 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xvf22" podStartSLOduration=3.755517045 podStartE2EDuration="29.688331825s" podCreationTimestamp="2025-12-01 14:52:36 +0000 UTC" firstStartedPulling="2025-12-01 14:52:39.071187191 +0000 UTC m=+1124.834696814" lastFinishedPulling="2025-12-01 14:53:05.004002001 +0000 UTC m=+1150.767511594" observedRunningTime="2025-12-01 14:53:05.684499492 +0000 UTC m=+1151.448009115" watchObservedRunningTime="2025-12-01 14:53:05.688331825 +0000 UTC m=+1151.451841428" Dec 01 14:53:08 crc kubenswrapper[4810]: E1201 14:53:08.126183 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-72tnh" podUID="e67d0431-4b18-4743-842d-d0728b23727a" Dec 01 14:53:08 crc kubenswrapper[4810]: E1201 14:53:08.137942 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8p7kt" podUID="dc03c0f9-e5d2-4fa3-98c6-d9a1d46638e0" Dec 01 14:53:08 crc kubenswrapper[4810]: E1201 14:53:08.141940 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-92m6z" podUID="58be13b4-ea8d-4ec2-8421-18d3fb2f744d" Dec 01 14:53:08 crc kubenswrapper[4810]: I1201 14:53:08.658278 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-92m6z" event={"ID":"58be13b4-ea8d-4ec2-8421-18d3fb2f744d","Type":"ContainerStarted","Data":"96d51c5f0164f2975c5fcab523d6308d088fac6c415245da9c7d9138de481479"} Dec 01 14:53:08 crc kubenswrapper[4810]: I1201 14:53:08.661278 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8p7kt" event={"ID":"dc03c0f9-e5d2-4fa3-98c6-d9a1d46638e0","Type":"ContainerStarted","Data":"6c12ee8d517fa1b061322a4c68a7ab1b31b0af4e78dbc3bc25b8dcedca0265e9"} Dec 01 14:53:08 crc kubenswrapper[4810]: I1201 14:53:08.670887 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lrsc2" event={"ID":"27e828db-c91c-49ff-88f2-eaab1ea075a0","Type":"ContainerStarted","Data":"6ea12a986b557ee8fac21179dd2dd8e32b2dfc1d2eced873423e63e56cd38464"} Dec 01 14:53:08 crc kubenswrapper[4810]: I1201 14:53:08.671095 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lrsc2" Dec 01 14:53:08 crc kubenswrapper[4810]: I1201 14:53:08.675188 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-72tnh" event={"ID":"e67d0431-4b18-4743-842d-d0728b23727a","Type":"ContainerStarted","Data":"0253bd7fec1a63ac8490b0f34ace42b5da7ea3423ce958a9ed8457b806cd0262"} Dec 01 14:53:08 crc kubenswrapper[4810]: E1201 14:53:08.677315 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-72tnh" podUID="e67d0431-4b18-4743-842d-d0728b23727a" Dec 01 14:53:08 crc kubenswrapper[4810]: I1201 14:53:08.717162 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lrsc2" podStartSLOduration=5.823742829 podStartE2EDuration="34.717143075s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:39.031671467 +0000 UTC m=+1124.795181070" lastFinishedPulling="2025-12-01 14:53:07.925071713 +0000 UTC m=+1153.688581316" observedRunningTime="2025-12-01 14:53:08.715591592 +0000 UTC m=+1154.479101205" watchObservedRunningTime="2025-12-01 14:53:08.717143075 +0000 UTC m=+1154.480652678" Dec 01 14:53:08 crc kubenswrapper[4810]: E1201 14:53:08.855434 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9zzgw" podUID="b6329d13-284d-48cc-866e-57744f0826bf" Dec 01 14:53:09 crc kubenswrapper[4810]: E1201 14:53:09.145655 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-wss6v" podUID="b56ca08d-173a-48a4-a537-8b9e84b4b2a2" Dec 01 14:53:09 crc kubenswrapper[4810]: E1201 14:53:09.145724 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lh6jm" podUID="fc17d065-8168-486a-ba45-1fd0cd31d9f0" Dec 01 14:53:09 crc kubenswrapper[4810]: E1201 14:53:09.166641 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-gvx8g" podUID="e6700584-b120-491c-8920-caa3332ba444" Dec 01 14:53:09 crc kubenswrapper[4810]: I1201 14:53:09.691601 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" event={"ID":"68116291-3829-4a92-aba9-c1fbb755c988","Type":"ContainerStarted","Data":"4c388c9030e830c15a3447a71741fd566cfd67f246ad6471162e25da53b199d5"} Dec 01 14:53:09 crc kubenswrapper[4810]: I1201 14:53:09.693448 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-gvx8g" event={"ID":"e6700584-b120-491c-8920-caa3332ba444","Type":"ContainerStarted","Data":"70f91c70beb11109f9c6b802edf0f9d4f713ab009c0568aac20afa43bf0d49bb"} Dec 01 14:53:09 crc kubenswrapper[4810]: E1201 14:53:09.695407 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:986861e5a0a9954f63581d9d55a30f8057883cefea489415d76257774526eea3\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-gvx8g" podUID="e6700584-b120-491c-8920-caa3332ba444" Dec 01 14:53:09 crc kubenswrapper[4810]: I1201 14:53:09.696136 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-wss6v" event={"ID":"b56ca08d-173a-48a4-a537-8b9e84b4b2a2","Type":"ContainerStarted","Data":"2ef1ec9bc8209a19c1b6f171dbe567c1068e98a1e6c38dd574785d60cd2ca88f"} Dec 01 14:53:09 crc kubenswrapper[4810]: I1201 14:53:09.697960 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9zzgw" event={"ID":"b6329d13-284d-48cc-866e-57744f0826bf","Type":"ContainerStarted","Data":"59f0efe9f508d41f4d7d97efcd244be49e3ddb65a5f494c310a58ac771c6428f"} Dec 01 14:53:09 crc kubenswrapper[4810]: I1201 14:53:09.702755 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fs6gw" event={"ID":"36f3dbad-9483-4695-95fe-7ce6520b4db3","Type":"ContainerStarted","Data":"f15b6548ee3cd1e37878338bcd57703f5db34ea496b2e9899143cfd238c76a62"} Dec 01 14:53:09 crc kubenswrapper[4810]: I1201 14:53:09.702804 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fs6gw" Dec 01 14:53:09 crc kubenswrapper[4810]: I1201 14:53:09.705680 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lh6jm" event={"ID":"fc17d065-8168-486a-ba45-1fd0cd31d9f0","Type":"ContainerStarted","Data":"5334196ae61ed4e46008b8e1241322768d621b963e6c5539af059091b6339daf"} Dec 01 14:53:09 crc kubenswrapper[4810]: I1201 14:53:09.705896 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fs6gw" Dec 01 14:53:09 crc kubenswrapper[4810]: I1201 14:53:09.707671 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-lrsc2" Dec 01 14:53:09 crc kubenswrapper[4810]: E1201 14:53:09.707793 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429\\\"\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lh6jm" podUID="fc17d065-8168-486a-ba45-1fd0cd31d9f0" Dec 01 14:53:09 crc kubenswrapper[4810]: I1201 14:53:09.798557 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-fs6gw" podStartSLOduration=6.895637646 podStartE2EDuration="35.798533019s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:39.129990639 +0000 UTC m=+1124.893500242" lastFinishedPulling="2025-12-01 14:53:08.032886012 +0000 UTC m=+1153.796395615" observedRunningTime="2025-12-01 14:53:09.78680808 +0000 UTC m=+1155.550317703" watchObservedRunningTime="2025-12-01 14:53:09.798533019 +0000 UTC m=+1155.562042632" Dec 01 14:53:10 crc kubenswrapper[4810]: E1201 14:53:10.616610 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9pbbh" podUID="c88f9be0-7e6f-4021-86c2-a43b34012a2c" Dec 01 14:53:10 crc kubenswrapper[4810]: E1201 14:53:10.708188 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-788fn" podUID="816177c4-0c9b-4826-8844-84db106ec51d" Dec 01 14:53:10 crc kubenswrapper[4810]: I1201 14:53:10.719871 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-nnvmc" event={"ID":"ae4759b1-d998-49b4-8031-f638141177a5","Type":"ContainerStarted","Data":"2eaf39562fdb8660532d07860f735df202db4bfa150953ee5b80727632c32880"} Dec 01 14:53:10 crc kubenswrapper[4810]: I1201 14:53:10.720862 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-nnvmc" Dec 01 14:53:10 crc kubenswrapper[4810]: I1201 14:53:10.730157 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4zjxs" event={"ID":"308f7f89-2d44-42cb-8533-2b82306341c6","Type":"ContainerStarted","Data":"58e381d72bc739d8dec29b00c7ea5707476ce15f9942709f5de7a86cc5ed7c4f"} Dec 01 14:53:10 crc kubenswrapper[4810]: I1201 14:53:10.730569 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4zjxs" Dec 01 14:53:10 crc kubenswrapper[4810]: I1201 14:53:10.732005 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8p7kt" event={"ID":"dc03c0f9-e5d2-4fa3-98c6-d9a1d46638e0","Type":"ContainerStarted","Data":"ba24a4f9accf59aab98cc4a4a42b9e76fb73c4203971f07d368c805ec473342a"} Dec 01 14:53:10 crc kubenswrapper[4810]: I1201 14:53:10.732041 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8p7kt" Dec 01 14:53:10 crc kubenswrapper[4810]: I1201 14:53:10.739690 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-nnvmc" Dec 01 14:53:10 crc kubenswrapper[4810]: I1201 14:53:10.740693 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9pbbh" event={"ID":"c88f9be0-7e6f-4021-86c2-a43b34012a2c","Type":"ContainerStarted","Data":"fc2700a3baaaf0722df24123c9467f9082828f188affeaa4d1799624f24f2c20"} Dec 01 14:53:10 crc kubenswrapper[4810]: I1201 14:53:10.743046 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4zjxs" Dec 01 14:53:10 crc kubenswrapper[4810]: I1201 14:53:10.745226 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-nnvmc" podStartSLOduration=3.035298501 podStartE2EDuration="36.745210932s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:36.582008414 +0000 UTC m=+1122.345518027" lastFinishedPulling="2025-12-01 14:53:10.291920855 +0000 UTC m=+1156.055430458" observedRunningTime="2025-12-01 14:53:10.736892076 +0000 UTC m=+1156.500401679" watchObservedRunningTime="2025-12-01 14:53:10.745210932 +0000 UTC m=+1156.508720535" Dec 01 14:53:10 crc kubenswrapper[4810]: I1201 14:53:10.762978 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" event={"ID":"50c4dc84-7965-4b16-8935-f557dc0c3a74","Type":"ContainerStarted","Data":"06722e929ee8ab9771611523409f54c856df32eca5a720e37a0ca3a85505e943"} Dec 01 14:53:10 crc kubenswrapper[4810]: I1201 14:53:10.769912 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" event={"ID":"68116291-3829-4a92-aba9-c1fbb755c988","Type":"ContainerStarted","Data":"5b8e41af3ef78fb6d9dd269382b809c557793c2779d06413cce9bde98d5409ba"} Dec 01 14:53:10 crc kubenswrapper[4810]: I1201 14:53:10.770136 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" Dec 01 14:53:10 crc kubenswrapper[4810]: I1201 14:53:10.772446 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8p7kt" podStartSLOduration=5.499039777 podStartE2EDuration="36.772435321s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:39.003261476 +0000 UTC m=+1124.766771069" lastFinishedPulling="2025-12-01 14:53:10.27665701 +0000 UTC m=+1156.040166613" observedRunningTime="2025-12-01 14:53:10.770048017 +0000 UTC m=+1156.533557620" watchObservedRunningTime="2025-12-01 14:53:10.772435321 +0000 UTC m=+1156.535944924" Dec 01 14:53:10 crc kubenswrapper[4810]: I1201 14:53:10.793554 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-788fn" event={"ID":"816177c4-0c9b-4826-8844-84db106ec51d","Type":"ContainerStarted","Data":"2bae2acf5250d00661b491d1a0f05b9f4fdd3ba9d00d954d5b98dc1af864e7c2"} Dec 01 14:53:10 crc kubenswrapper[4810]: I1201 14:53:10.795374 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4zjxs" podStartSLOduration=3.118908824 podStartE2EDuration="36.795361805s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:36.588869501 +0000 UTC m=+1122.352379104" lastFinishedPulling="2025-12-01 14:53:10.265322482 +0000 UTC m=+1156.028832085" observedRunningTime="2025-12-01 14:53:10.795145999 +0000 UTC m=+1156.558655602" watchObservedRunningTime="2025-12-01 14:53:10.795361805 +0000 UTC m=+1156.558871408" Dec 01 14:53:10 crc kubenswrapper[4810]: E1201 14:53:10.800854 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429\\\"\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lh6jm" podUID="fc17d065-8168-486a-ba45-1fd0cd31d9f0" Dec 01 14:53:10 crc kubenswrapper[4810]: I1201 14:53:10.897169 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" podStartSLOduration=32.674138942 podStartE2EDuration="36.897151531s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:53:04.657616288 +0000 UTC m=+1150.421125891" lastFinishedPulling="2025-12-01 14:53:08.880628877 +0000 UTC m=+1154.644138480" observedRunningTime="2025-12-01 14:53:10.881754752 +0000 UTC m=+1156.645264365" watchObservedRunningTime="2025-12-01 14:53:10.897151531 +0000 UTC m=+1156.660661134" Dec 01 14:53:11 crc kubenswrapper[4810]: E1201 14:53:11.198666 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-7bcjp" podUID="322588d9-4c11-4676-8eb4-15cce77ccd44" Dec 01 14:53:11 crc kubenswrapper[4810]: E1201 14:53:11.225936 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpln8" podUID="573876c9-548b-477b-8d92-b7b7d50c7bb5" Dec 01 14:53:11 crc kubenswrapper[4810]: E1201 14:53:11.239016 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c82g8" podUID="7007dd52-d7df-4f27-81d0-95ee762c87c1" Dec 01 14:53:11 crc kubenswrapper[4810]: E1201 14:53:11.255265 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ccwrk" podUID="4b6ce85f-d265-40a2-aa92-2a2d76339acf" Dec 01 14:53:11 crc kubenswrapper[4810]: E1201 14:53:11.260786 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-z85kg" podUID="19093fac-7ee4-43aa-ad57-7561d75de41f" Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.813187 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-gvx8g" event={"ID":"e6700584-b120-491c-8920-caa3332ba444","Type":"ContainerStarted","Data":"5efe40c5d41beee6d1d7344a0d4d2485dfa00e03406472e5d09ea02fcb9e89c2"} Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.814147 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-gvx8g" Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.815532 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-7bcjp" event={"ID":"322588d9-4c11-4676-8eb4-15cce77ccd44","Type":"ContainerStarted","Data":"53222b54b5ce5899efaf2aa487bc73000d44b1d8a50f270c5787d8aac2984bf7"} Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.820897 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c82g8" event={"ID":"7007dd52-d7df-4f27-81d0-95ee762c87c1","Type":"ContainerStarted","Data":"b61b41d2d351f2947b3622ab678197b4fd2f63089f0d8bc0a21ec7184b17cd17"} Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.827354 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-92m6z" event={"ID":"58be13b4-ea8d-4ec2-8421-18d3fb2f744d","Type":"ContainerStarted","Data":"a8196f1cfacdb9060e119fa988ede6ed63d778414d3053f2becb3af2ec63aff5"} Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.827521 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-92m6z" Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.829823 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x2png" event={"ID":"7798189d-af43-43aa-afd1-4257194bcb12","Type":"ContainerStarted","Data":"581b256702620b1ff669e654ecb792d793dd46bc0f8649b50f6aa9a3b8e4dd71"} Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.830416 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x2png" Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.832057 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9pbbh" event={"ID":"c88f9be0-7e6f-4021-86c2-a43b34012a2c","Type":"ContainerStarted","Data":"d1d2d9c1d7d593f3149b66212f1a8ded7493de53505b5a031c046ca88c8bfe8b"} Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.832428 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9pbbh" Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.832986 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x2png" Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.834788 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-72tnh" event={"ID":"e67d0431-4b18-4743-842d-d0728b23727a","Type":"ContainerStarted","Data":"b5dd0ec1129809364f49a6130f7d28377971776743b08273abc1ca41962cca07"} Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.841198 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" event={"ID":"50c4dc84-7965-4b16-8935-f557dc0c3a74","Type":"ContainerStarted","Data":"bda36f8f109a0da280ef0783eb8984cdc46a0b3331d096819bfa94791cdb24ae"} Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.841322 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.844229 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-788fn" event={"ID":"816177c4-0c9b-4826-8844-84db106ec51d","Type":"ContainerStarted","Data":"c627539182510525eeec1a385f01a98e523a003c6a2ad1342ad0e773d2ad4a60"} Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.844345 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-788fn" Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.849723 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-gvx8g" podStartSLOduration=4.6840781719999995 podStartE2EDuration="37.849705433s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:38.362383701 +0000 UTC m=+1124.125893304" lastFinishedPulling="2025-12-01 14:53:11.528010962 +0000 UTC m=+1157.291520565" observedRunningTime="2025-12-01 14:53:11.846143396 +0000 UTC m=+1157.609652999" watchObservedRunningTime="2025-12-01 14:53:11.849705433 +0000 UTC m=+1157.613215056" Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.850779 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ccwrk" event={"ID":"4b6ce85f-d265-40a2-aa92-2a2d76339acf","Type":"ContainerStarted","Data":"7bf9240da7f517964864f746f83fd26fc19846f2183e04cc4cf456ae2410b096"} Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.858708 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-wss6v" event={"ID":"b56ca08d-173a-48a4-a537-8b9e84b4b2a2","Type":"ContainerStarted","Data":"73290b9eecd01a28082a8dd617800d03a15d64389397bf73883146c006f2b793"} Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.859416 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-wss6v" Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.862785 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-z85kg" event={"ID":"19093fac-7ee4-43aa-ad57-7561d75de41f","Type":"ContainerStarted","Data":"647dcb235df47cf8c2f67f980dd98e14076dffb746a4eebf738d5ba54f37c877"} Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.869172 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpln8" event={"ID":"573876c9-548b-477b-8d92-b7b7d50c7bb5","Type":"ContainerStarted","Data":"9bbbb957cf6ba2aa70efdbbf15e75bf514ca98c81d13992affed0ea75538be33"} Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.877827 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-788fn" podStartSLOduration=5.152332196 podStartE2EDuration="37.877810377s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:38.679543489 +0000 UTC m=+1124.443053092" lastFinishedPulling="2025-12-01 14:53:11.40502167 +0000 UTC m=+1157.168531273" observedRunningTime="2025-12-01 14:53:11.87310751 +0000 UTC m=+1157.636617113" watchObservedRunningTime="2025-12-01 14:53:11.877810377 +0000 UTC m=+1157.641319980" Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.892537 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9zzgw" event={"ID":"b6329d13-284d-48cc-866e-57744f0826bf","Type":"ContainerStarted","Data":"4c784ced0e78f5c0a64f2c15992b87b0b41748f08196a346aff34e2b433df47d"} Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.892578 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9zzgw" Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.934104 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-x2png" podStartSLOduration=6.7598532460000005 podStartE2EDuration="37.934085386s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:39.140156445 +0000 UTC m=+1124.903666048" lastFinishedPulling="2025-12-01 14:53:10.314388585 +0000 UTC m=+1156.077898188" observedRunningTime="2025-12-01 14:53:11.929173843 +0000 UTC m=+1157.692683456" watchObservedRunningTime="2025-12-01 14:53:11.934085386 +0000 UTC m=+1157.697594989" Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.964260 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-92m6z" podStartSLOduration=6.310461414 podStartE2EDuration="37.964244496s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:38.610585045 +0000 UTC m=+1124.374094648" lastFinishedPulling="2025-12-01 14:53:10.264368127 +0000 UTC m=+1156.027877730" observedRunningTime="2025-12-01 14:53:11.957899943 +0000 UTC m=+1157.721409556" watchObservedRunningTime="2025-12-01 14:53:11.964244496 +0000 UTC m=+1157.727754099" Dec 01 14:53:11 crc kubenswrapper[4810]: I1201 14:53:11.987256 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" podStartSLOduration=31.498720703 podStartE2EDuration="37.98723717s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:53:03.776133017 +0000 UTC m=+1149.539642620" lastFinishedPulling="2025-12-01 14:53:10.264649484 +0000 UTC m=+1156.028159087" observedRunningTime="2025-12-01 14:53:11.986310775 +0000 UTC m=+1157.749820378" watchObservedRunningTime="2025-12-01 14:53:11.98723717 +0000 UTC m=+1157.750746763" Dec 01 14:53:12 crc kubenswrapper[4810]: I1201 14:53:12.030432 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-656fd97d56-9ghs2" Dec 01 14:53:12 crc kubenswrapper[4810]: I1201 14:53:12.045754 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-72tnh" podStartSLOduration=6.517522191 podStartE2EDuration="38.0457346s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:38.98944802 +0000 UTC m=+1124.752957623" lastFinishedPulling="2025-12-01 14:53:10.517660439 +0000 UTC m=+1156.281170032" observedRunningTime="2025-12-01 14:53:12.044415544 +0000 UTC m=+1157.807925147" watchObservedRunningTime="2025-12-01 14:53:12.0457346 +0000 UTC m=+1157.809244203" Dec 01 14:53:12 crc kubenswrapper[4810]: I1201 14:53:12.078362 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9pbbh" podStartSLOduration=5.26763524 podStartE2EDuration="38.078338506s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:38.479603017 +0000 UTC m=+1124.243112620" lastFinishedPulling="2025-12-01 14:53:11.290306293 +0000 UTC m=+1157.053815886" observedRunningTime="2025-12-01 14:53:12.071286114 +0000 UTC m=+1157.834795737" watchObservedRunningTime="2025-12-01 14:53:12.078338506 +0000 UTC m=+1157.841848109" Dec 01 14:53:12 crc kubenswrapper[4810]: I1201 14:53:12.102825 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9zzgw" podStartSLOduration=5.969897031 podStartE2EDuration="38.102800811s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:38.43261836 +0000 UTC m=+1124.196127963" lastFinishedPulling="2025-12-01 14:53:10.56552214 +0000 UTC m=+1156.329031743" observedRunningTime="2025-12-01 14:53:12.09614483 +0000 UTC m=+1157.859654433" watchObservedRunningTime="2025-12-01 14:53:12.102800811 +0000 UTC m=+1157.866310414" Dec 01 14:53:12 crc kubenswrapper[4810]: I1201 14:53:12.223064 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-wss6v" podStartSLOduration=6.563929062 podStartE2EDuration="38.223039928s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:38.945912257 +0000 UTC m=+1124.709421860" lastFinishedPulling="2025-12-01 14:53:10.605023123 +0000 UTC m=+1156.368532726" observedRunningTime="2025-12-01 14:53:12.220201151 +0000 UTC m=+1157.983710754" watchObservedRunningTime="2025-12-01 14:53:12.223039928 +0000 UTC m=+1157.986549531" Dec 01 14:53:12 crc kubenswrapper[4810]: I1201 14:53:12.900618 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-7bcjp" event={"ID":"322588d9-4c11-4676-8eb4-15cce77ccd44","Type":"ContainerStarted","Data":"1c1b10ff80f04ae4791d2d11ceae63cc86e3106933915112cdaedce71edbb939"} Dec 01 14:53:12 crc kubenswrapper[4810]: I1201 14:53:12.901796 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-7bcjp" Dec 01 14:53:12 crc kubenswrapper[4810]: I1201 14:53:12.903687 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c82g8" event={"ID":"7007dd52-d7df-4f27-81d0-95ee762c87c1","Type":"ContainerStarted","Data":"d8c297120906b064d6f5b5b52d4c504b966f42809576e4149891bd72b21a812d"} Dec 01 14:53:12 crc kubenswrapper[4810]: I1201 14:53:12.929960 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-7bcjp" podStartSLOduration=5.089294003 podStartE2EDuration="38.929916855s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:38.417303433 +0000 UTC m=+1124.180813036" lastFinishedPulling="2025-12-01 14:53:12.257926285 +0000 UTC m=+1158.021435888" observedRunningTime="2025-12-01 14:53:12.924069266 +0000 UTC m=+1158.687578869" watchObservedRunningTime="2025-12-01 14:53:12.929916855 +0000 UTC m=+1158.693426458" Dec 01 14:53:13 crc kubenswrapper[4810]: I1201 14:53:13.053656 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c82g8" podStartSLOduration=5.419013242 podStartE2EDuration="39.053623947s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:38.625370927 +0000 UTC m=+1124.388880530" lastFinishedPulling="2025-12-01 14:53:12.259981632 +0000 UTC m=+1158.023491235" observedRunningTime="2025-12-01 14:53:13.045868165 +0000 UTC m=+1158.809377768" watchObservedRunningTime="2025-12-01 14:53:13.053623947 +0000 UTC m=+1158.817133560" Dec 01 14:53:13 crc kubenswrapper[4810]: I1201 14:53:13.909822 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c82g8" Dec 01 14:53:15 crc kubenswrapper[4810]: I1201 14:53:15.556300 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-72tnh" Dec 01 14:53:16 crc kubenswrapper[4810]: I1201 14:53:16.073075 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-8p7kt" Dec 01 14:53:16 crc kubenswrapper[4810]: I1201 14:53:16.190077 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-92m6z" Dec 01 14:53:20 crc kubenswrapper[4810]: I1201 14:53:20.951856 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2lp78" Dec 01 14:53:21 crc kubenswrapper[4810]: I1201 14:53:21.230339 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6698bcb44667k99" Dec 01 14:53:24 crc kubenswrapper[4810]: I1201 14:53:24.761635 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-7bcjp" Dec 01 14:53:24 crc kubenswrapper[4810]: I1201 14:53:24.983220 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpln8" event={"ID":"573876c9-548b-477b-8d92-b7b7d50c7bb5","Type":"ContainerStarted","Data":"a614dec398d61d4f5351a6ed46b4974aa0c9d712ec53f7f2e3ba11eb06b0c848"} Dec 01 14:53:24 crc kubenswrapper[4810]: I1201 14:53:24.983586 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpln8" Dec 01 14:53:24 crc kubenswrapper[4810]: I1201 14:53:24.985722 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-z85kg" event={"ID":"19093fac-7ee4-43aa-ad57-7561d75de41f","Type":"ContainerStarted","Data":"19d2222170b1c0509b82b0e912bacaec43e812fc5345ef1c09462a3f2ef43c0d"} Dec 01 14:53:24 crc kubenswrapper[4810]: I1201 14:53:24.985850 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-z85kg" Dec 01 14:53:24 crc kubenswrapper[4810]: I1201 14:53:24.987299 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ccwrk" event={"ID":"4b6ce85f-d265-40a2-aa92-2a2d76339acf","Type":"ContainerStarted","Data":"1ee5f0533e0639abf26677af337532eac0c30fb45a981f35ae6539fe331dc276"} Dec 01 14:53:24 crc kubenswrapper[4810]: I1201 14:53:24.987453 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ccwrk" Dec 01 14:53:25 crc kubenswrapper[4810]: I1201 14:53:25.005276 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpln8" podStartSLOduration=6.065756847 podStartE2EDuration="51.005256731s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:38.624829682 +0000 UTC m=+1124.388339285" lastFinishedPulling="2025-12-01 14:53:23.564329566 +0000 UTC m=+1169.327839169" observedRunningTime="2025-12-01 14:53:24.998877947 +0000 UTC m=+1170.762387570" watchObservedRunningTime="2025-12-01 14:53:25.005256731 +0000 UTC m=+1170.768766334" Dec 01 14:53:25 crc kubenswrapper[4810]: I1201 14:53:25.021401 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-z85kg" podStartSLOduration=5.274174167 podStartE2EDuration="51.021386299s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:38.979963502 +0000 UTC m=+1124.743473105" lastFinishedPulling="2025-12-01 14:53:24.727175634 +0000 UTC m=+1170.490685237" observedRunningTime="2025-12-01 14:53:25.015093148 +0000 UTC m=+1170.778602751" watchObservedRunningTime="2025-12-01 14:53:25.021386299 +0000 UTC m=+1170.784895892" Dec 01 14:53:25 crc kubenswrapper[4810]: I1201 14:53:25.028012 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9zzgw" Dec 01 14:53:25 crc kubenswrapper[4810]: I1201 14:53:25.035952 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ccwrk" podStartSLOduration=3.723043822 podStartE2EDuration="51.035930785s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:36.252494051 +0000 UTC m=+1122.016003654" lastFinishedPulling="2025-12-01 14:53:23.565381004 +0000 UTC m=+1169.328890617" observedRunningTime="2025-12-01 14:53:25.034021842 +0000 UTC m=+1170.797531445" watchObservedRunningTime="2025-12-01 14:53:25.035930785 +0000 UTC m=+1170.799440388" Dec 01 14:53:25 crc kubenswrapper[4810]: I1201 14:53:25.123674 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-gvx8g" Dec 01 14:53:25 crc kubenswrapper[4810]: I1201 14:53:25.532525 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-wss6v" Dec 01 14:53:25 crc kubenswrapper[4810]: I1201 14:53:25.568042 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-72tnh" Dec 01 14:53:25 crc kubenswrapper[4810]: I1201 14:53:25.592934 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9pbbh" Dec 01 14:53:25 crc kubenswrapper[4810]: I1201 14:53:25.612882 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-788fn" Dec 01 14:53:25 crc kubenswrapper[4810]: I1201 14:53:25.813958 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-c82g8" Dec 01 14:53:27 crc kubenswrapper[4810]: I1201 14:53:27.000736 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lh6jm" event={"ID":"fc17d065-8168-486a-ba45-1fd0cd31d9f0","Type":"ContainerStarted","Data":"7abc5837db8897ad8c012691f8c30edcf9d3a2802aa857f6939742f5388c8345"} Dec 01 14:53:27 crc kubenswrapper[4810]: I1201 14:53:27.001536 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lh6jm" Dec 01 14:53:27 crc kubenswrapper[4810]: I1201 14:53:27.019690 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lh6jm" podStartSLOduration=5.4555039260000004 podStartE2EDuration="53.019676112s" podCreationTimestamp="2025-12-01 14:52:34 +0000 UTC" firstStartedPulling="2025-12-01 14:52:38.625146841 +0000 UTC m=+1124.388656444" lastFinishedPulling="2025-12-01 14:53:26.189319027 +0000 UTC m=+1171.952828630" observedRunningTime="2025-12-01 14:53:27.013741801 +0000 UTC m=+1172.777251404" watchObservedRunningTime="2025-12-01 14:53:27.019676112 +0000 UTC m=+1172.783185715" Dec 01 14:53:34 crc kubenswrapper[4810]: I1201 14:53:34.746056 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ccwrk" Dec 01 14:53:35 crc kubenswrapper[4810]: I1201 14:53:35.157769 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-lh6jm" Dec 01 14:53:35 crc kubenswrapper[4810]: I1201 14:53:35.165177 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-hpln8" Dec 01 14:53:36 crc kubenswrapper[4810]: I1201 14:53:36.132521 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-z85kg" Dec 01 14:53:48 crc kubenswrapper[4810]: I1201 14:53:48.942867 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-557f57d995-vm4pr"] Dec 01 14:53:48 crc kubenswrapper[4810]: I1201 14:53:48.944549 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-557f57d995-vm4pr" Dec 01 14:53:48 crc kubenswrapper[4810]: I1201 14:53:48.951983 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 01 14:53:48 crc kubenswrapper[4810]: I1201 14:53:48.952109 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 01 14:53:48 crc kubenswrapper[4810]: I1201 14:53:48.952209 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 01 14:53:48 crc kubenswrapper[4810]: I1201 14:53:48.952224 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-gwpmb" Dec 01 14:53:48 crc kubenswrapper[4810]: I1201 14:53:48.966074 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-557f57d995-vm4pr"] Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.035720 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwk7g\" (UniqueName: \"kubernetes.io/projected/714a979a-695b-4879-82b8-13b7068f124c-kube-api-access-jwk7g\") pod \"dnsmasq-dns-557f57d995-vm4pr\" (UID: \"714a979a-695b-4879-82b8-13b7068f124c\") " pod="openstack/dnsmasq-dns-557f57d995-vm4pr" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.035809 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/714a979a-695b-4879-82b8-13b7068f124c-config\") pod \"dnsmasq-dns-557f57d995-vm4pr\" (UID: \"714a979a-695b-4879-82b8-13b7068f124c\") " pod="openstack/dnsmasq-dns-557f57d995-vm4pr" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.042512 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-766fdc659c-z6tgt"] Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.043922 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766fdc659c-z6tgt" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.046455 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.050903 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-766fdc659c-z6tgt"] Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.137351 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwk7g\" (UniqueName: \"kubernetes.io/projected/714a979a-695b-4879-82b8-13b7068f124c-kube-api-access-jwk7g\") pod \"dnsmasq-dns-557f57d995-vm4pr\" (UID: \"714a979a-695b-4879-82b8-13b7068f124c\") " pod="openstack/dnsmasq-dns-557f57d995-vm4pr" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.137419 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58151c02-5c33-4721-b2ec-17c1921b8e2a-dns-svc\") pod \"dnsmasq-dns-766fdc659c-z6tgt\" (UID: \"58151c02-5c33-4721-b2ec-17c1921b8e2a\") " pod="openstack/dnsmasq-dns-766fdc659c-z6tgt" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.137446 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/714a979a-695b-4879-82b8-13b7068f124c-config\") pod \"dnsmasq-dns-557f57d995-vm4pr\" (UID: \"714a979a-695b-4879-82b8-13b7068f124c\") " pod="openstack/dnsmasq-dns-557f57d995-vm4pr" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.137513 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqq4w\" (UniqueName: \"kubernetes.io/projected/58151c02-5c33-4721-b2ec-17c1921b8e2a-kube-api-access-jqq4w\") pod \"dnsmasq-dns-766fdc659c-z6tgt\" (UID: \"58151c02-5c33-4721-b2ec-17c1921b8e2a\") " pod="openstack/dnsmasq-dns-766fdc659c-z6tgt" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.137558 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58151c02-5c33-4721-b2ec-17c1921b8e2a-config\") pod \"dnsmasq-dns-766fdc659c-z6tgt\" (UID: \"58151c02-5c33-4721-b2ec-17c1921b8e2a\") " pod="openstack/dnsmasq-dns-766fdc659c-z6tgt" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.138386 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/714a979a-695b-4879-82b8-13b7068f124c-config\") pod \"dnsmasq-dns-557f57d995-vm4pr\" (UID: \"714a979a-695b-4879-82b8-13b7068f124c\") " pod="openstack/dnsmasq-dns-557f57d995-vm4pr" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.165605 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwk7g\" (UniqueName: \"kubernetes.io/projected/714a979a-695b-4879-82b8-13b7068f124c-kube-api-access-jwk7g\") pod \"dnsmasq-dns-557f57d995-vm4pr\" (UID: \"714a979a-695b-4879-82b8-13b7068f124c\") " pod="openstack/dnsmasq-dns-557f57d995-vm4pr" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.238511 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqq4w\" (UniqueName: \"kubernetes.io/projected/58151c02-5c33-4721-b2ec-17c1921b8e2a-kube-api-access-jqq4w\") pod \"dnsmasq-dns-766fdc659c-z6tgt\" (UID: \"58151c02-5c33-4721-b2ec-17c1921b8e2a\") " pod="openstack/dnsmasq-dns-766fdc659c-z6tgt" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.238637 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58151c02-5c33-4721-b2ec-17c1921b8e2a-config\") pod \"dnsmasq-dns-766fdc659c-z6tgt\" (UID: \"58151c02-5c33-4721-b2ec-17c1921b8e2a\") " pod="openstack/dnsmasq-dns-766fdc659c-z6tgt" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.238754 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58151c02-5c33-4721-b2ec-17c1921b8e2a-dns-svc\") pod \"dnsmasq-dns-766fdc659c-z6tgt\" (UID: \"58151c02-5c33-4721-b2ec-17c1921b8e2a\") " pod="openstack/dnsmasq-dns-766fdc659c-z6tgt" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.239630 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58151c02-5c33-4721-b2ec-17c1921b8e2a-config\") pod \"dnsmasq-dns-766fdc659c-z6tgt\" (UID: \"58151c02-5c33-4721-b2ec-17c1921b8e2a\") " pod="openstack/dnsmasq-dns-766fdc659c-z6tgt" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.240214 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58151c02-5c33-4721-b2ec-17c1921b8e2a-dns-svc\") pod \"dnsmasq-dns-766fdc659c-z6tgt\" (UID: \"58151c02-5c33-4721-b2ec-17c1921b8e2a\") " pod="openstack/dnsmasq-dns-766fdc659c-z6tgt" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.255454 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqq4w\" (UniqueName: \"kubernetes.io/projected/58151c02-5c33-4721-b2ec-17c1921b8e2a-kube-api-access-jqq4w\") pod \"dnsmasq-dns-766fdc659c-z6tgt\" (UID: \"58151c02-5c33-4721-b2ec-17c1921b8e2a\") " pod="openstack/dnsmasq-dns-766fdc659c-z6tgt" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.261759 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-557f57d995-vm4pr" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.368856 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766fdc659c-z6tgt" Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.679700 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-557f57d995-vm4pr"] Dec 01 14:53:49 crc kubenswrapper[4810]: I1201 14:53:49.797739 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-766fdc659c-z6tgt"] Dec 01 14:53:49 crc kubenswrapper[4810]: W1201 14:53:49.802147 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58151c02_5c33_4721_b2ec_17c1921b8e2a.slice/crio-a4877255051a85acd3a9b9e804b1fbc6333ec2ce241bf7b0a1907a3b1450d117 WatchSource:0}: Error finding container a4877255051a85acd3a9b9e804b1fbc6333ec2ce241bf7b0a1907a3b1450d117: Status 404 returned error can't find the container with id a4877255051a85acd3a9b9e804b1fbc6333ec2ce241bf7b0a1907a3b1450d117 Dec 01 14:53:50 crc kubenswrapper[4810]: I1201 14:53:50.150087 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-557f57d995-vm4pr" event={"ID":"714a979a-695b-4879-82b8-13b7068f124c","Type":"ContainerStarted","Data":"c1e59ed03268bec1fee66f243cf43e18233f67980409d3eab1c8d011bfc61fae"} Dec 01 14:53:50 crc kubenswrapper[4810]: I1201 14:53:50.151605 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766fdc659c-z6tgt" event={"ID":"58151c02-5c33-4721-b2ec-17c1921b8e2a","Type":"ContainerStarted","Data":"a4877255051a85acd3a9b9e804b1fbc6333ec2ce241bf7b0a1907a3b1450d117"} Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.079431 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-557f57d995-vm4pr"] Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.096293 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57dc4c6697-vw5rh"] Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.097403 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57dc4c6697-vw5rh" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.153992 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57dc4c6697-vw5rh"] Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.179050 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3383b93f-a184-4b6e-ad59-313e19cf4371-config\") pod \"dnsmasq-dns-57dc4c6697-vw5rh\" (UID: \"3383b93f-a184-4b6e-ad59-313e19cf4371\") " pod="openstack/dnsmasq-dns-57dc4c6697-vw5rh" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.179148 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3383b93f-a184-4b6e-ad59-313e19cf4371-dns-svc\") pod \"dnsmasq-dns-57dc4c6697-vw5rh\" (UID: \"3383b93f-a184-4b6e-ad59-313e19cf4371\") " pod="openstack/dnsmasq-dns-57dc4c6697-vw5rh" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.179230 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94c5j\" (UniqueName: \"kubernetes.io/projected/3383b93f-a184-4b6e-ad59-313e19cf4371-kube-api-access-94c5j\") pod \"dnsmasq-dns-57dc4c6697-vw5rh\" (UID: \"3383b93f-a184-4b6e-ad59-313e19cf4371\") " pod="openstack/dnsmasq-dns-57dc4c6697-vw5rh" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.280646 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3383b93f-a184-4b6e-ad59-313e19cf4371-config\") pod \"dnsmasq-dns-57dc4c6697-vw5rh\" (UID: \"3383b93f-a184-4b6e-ad59-313e19cf4371\") " pod="openstack/dnsmasq-dns-57dc4c6697-vw5rh" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.280711 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3383b93f-a184-4b6e-ad59-313e19cf4371-dns-svc\") pod \"dnsmasq-dns-57dc4c6697-vw5rh\" (UID: \"3383b93f-a184-4b6e-ad59-313e19cf4371\") " pod="openstack/dnsmasq-dns-57dc4c6697-vw5rh" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.280839 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94c5j\" (UniqueName: \"kubernetes.io/projected/3383b93f-a184-4b6e-ad59-313e19cf4371-kube-api-access-94c5j\") pod \"dnsmasq-dns-57dc4c6697-vw5rh\" (UID: \"3383b93f-a184-4b6e-ad59-313e19cf4371\") " pod="openstack/dnsmasq-dns-57dc4c6697-vw5rh" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.284682 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3383b93f-a184-4b6e-ad59-313e19cf4371-config\") pod \"dnsmasq-dns-57dc4c6697-vw5rh\" (UID: \"3383b93f-a184-4b6e-ad59-313e19cf4371\") " pod="openstack/dnsmasq-dns-57dc4c6697-vw5rh" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.285382 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3383b93f-a184-4b6e-ad59-313e19cf4371-dns-svc\") pod \"dnsmasq-dns-57dc4c6697-vw5rh\" (UID: \"3383b93f-a184-4b6e-ad59-313e19cf4371\") " pod="openstack/dnsmasq-dns-57dc4c6697-vw5rh" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.320668 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94c5j\" (UniqueName: \"kubernetes.io/projected/3383b93f-a184-4b6e-ad59-313e19cf4371-kube-api-access-94c5j\") pod \"dnsmasq-dns-57dc4c6697-vw5rh\" (UID: \"3383b93f-a184-4b6e-ad59-313e19cf4371\") " pod="openstack/dnsmasq-dns-57dc4c6697-vw5rh" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.482970 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57dc4c6697-vw5rh" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.722202 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-766fdc659c-z6tgt"] Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.729248 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8446fd7c75-b8hvv"] Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.742258 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8446fd7c75-b8hvv" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.760813 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8446fd7c75-b8hvv"] Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.889694 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d19854e-9f51-4d60-a6c1-6c1487f31d92-config\") pod \"dnsmasq-dns-8446fd7c75-b8hvv\" (UID: \"2d19854e-9f51-4d60-a6c1-6c1487f31d92\") " pod="openstack/dnsmasq-dns-8446fd7c75-b8hvv" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.889744 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d19854e-9f51-4d60-a6c1-6c1487f31d92-dns-svc\") pod \"dnsmasq-dns-8446fd7c75-b8hvv\" (UID: \"2d19854e-9f51-4d60-a6c1-6c1487f31d92\") " pod="openstack/dnsmasq-dns-8446fd7c75-b8hvv" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.889809 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgpb9\" (UniqueName: \"kubernetes.io/projected/2d19854e-9f51-4d60-a6c1-6c1487f31d92-kube-api-access-vgpb9\") pod \"dnsmasq-dns-8446fd7c75-b8hvv\" (UID: \"2d19854e-9f51-4d60-a6c1-6c1487f31d92\") " pod="openstack/dnsmasq-dns-8446fd7c75-b8hvv" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.990962 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgpb9\" (UniqueName: \"kubernetes.io/projected/2d19854e-9f51-4d60-a6c1-6c1487f31d92-kube-api-access-vgpb9\") pod \"dnsmasq-dns-8446fd7c75-b8hvv\" (UID: \"2d19854e-9f51-4d60-a6c1-6c1487f31d92\") " pod="openstack/dnsmasq-dns-8446fd7c75-b8hvv" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.991034 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d19854e-9f51-4d60-a6c1-6c1487f31d92-config\") pod \"dnsmasq-dns-8446fd7c75-b8hvv\" (UID: \"2d19854e-9f51-4d60-a6c1-6c1487f31d92\") " pod="openstack/dnsmasq-dns-8446fd7c75-b8hvv" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.991057 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d19854e-9f51-4d60-a6c1-6c1487f31d92-dns-svc\") pod \"dnsmasq-dns-8446fd7c75-b8hvv\" (UID: \"2d19854e-9f51-4d60-a6c1-6c1487f31d92\") " pod="openstack/dnsmasq-dns-8446fd7c75-b8hvv" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.991877 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d19854e-9f51-4d60-a6c1-6c1487f31d92-dns-svc\") pod \"dnsmasq-dns-8446fd7c75-b8hvv\" (UID: \"2d19854e-9f51-4d60-a6c1-6c1487f31d92\") " pod="openstack/dnsmasq-dns-8446fd7c75-b8hvv" Dec 01 14:53:51 crc kubenswrapper[4810]: I1201 14:53:51.992673 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d19854e-9f51-4d60-a6c1-6c1487f31d92-config\") pod \"dnsmasq-dns-8446fd7c75-b8hvv\" (UID: \"2d19854e-9f51-4d60-a6c1-6c1487f31d92\") " pod="openstack/dnsmasq-dns-8446fd7c75-b8hvv" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.014153 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgpb9\" (UniqueName: \"kubernetes.io/projected/2d19854e-9f51-4d60-a6c1-6c1487f31d92-kube-api-access-vgpb9\") pod \"dnsmasq-dns-8446fd7c75-b8hvv\" (UID: \"2d19854e-9f51-4d60-a6c1-6c1487f31d92\") " pod="openstack/dnsmasq-dns-8446fd7c75-b8hvv" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.028345 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57dc4c6697-vw5rh"] Dec 01 14:53:52 crc kubenswrapper[4810]: W1201 14:53:52.040826 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3383b93f_a184_4b6e_ad59_313e19cf4371.slice/crio-5277866b40136ef3325b00ba3ad0cea015c3b81aa433de751da0201942b891d1 WatchSource:0}: Error finding container 5277866b40136ef3325b00ba3ad0cea015c3b81aa433de751da0201942b891d1: Status 404 returned error can't find the container with id 5277866b40136ef3325b00ba3ad0cea015c3b81aa433de751da0201942b891d1 Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.074290 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8446fd7c75-b8hvv" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.227704 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57dc4c6697-vw5rh" event={"ID":"3383b93f-a184-4b6e-ad59-313e19cf4371","Type":"ContainerStarted","Data":"5277866b40136ef3325b00ba3ad0cea015c3b81aa433de751da0201942b891d1"} Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.261364 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.263962 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.266567 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.266828 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.266945 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.267060 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.271175 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-fh65m" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.278138 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.278345 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.283534 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.397190 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.397233 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.397263 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.397284 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwvm6\" (UniqueName: \"kubernetes.io/projected/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-kube-api-access-zwvm6\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.397391 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.397501 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-config-data\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.397547 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.397574 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.397602 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.397616 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.398171 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.498903 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.498938 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.498965 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.499368 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwvm6\" (UniqueName: \"kubernetes.io/projected/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-kube-api-access-zwvm6\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.499429 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.499481 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-config-data\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.499507 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.499527 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.499544 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.499559 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.499605 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.499954 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.501074 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.502652 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.505166 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-config-data\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.505178 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.505286 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.506209 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.507744 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.507947 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.515452 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.522219 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8446fd7c75-b8hvv"] Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.522504 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwvm6\" (UniqueName: \"kubernetes.io/projected/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-kube-api-access-zwvm6\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.530505 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: W1201 14:53:52.560541 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d19854e_9f51_4d60_a6c1_6c1487f31d92.slice/crio-930fb6d111e16e19fbec709b72b94b609a9457f939308683b336f28471f7d648 WatchSource:0}: Error finding container 930fb6d111e16e19fbec709b72b94b609a9457f939308683b336f28471f7d648: Status 404 returned error can't find the container with id 930fb6d111e16e19fbec709b72b94b609a9457f939308683b336f28471f7d648 Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.598598 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.883695 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.885169 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.887824 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.887841 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.890850 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.891107 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.891235 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-mcxlp" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.891600 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.891891 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 01 14:53:52 crc kubenswrapper[4810]: I1201 14:53:52.898051 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.005280 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.005364 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.005396 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.005457 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.005554 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.005579 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.005594 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.005618 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9hct\" (UniqueName: \"kubernetes.io/projected/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-kube-api-access-d9hct\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.006205 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.006289 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.006319 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.017614 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 14:53:53 crc kubenswrapper[4810]: W1201 14:53:53.038081 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b81b6a9_0be6_4ce1_81b4_043edb1c221f.slice/crio-e48c8a12d183e258596aa4c10eacf8ff2cbe94f89874e3e923e2c7d69077f67d WatchSource:0}: Error finding container e48c8a12d183e258596aa4c10eacf8ff2cbe94f89874e3e923e2c7d69077f67d: Status 404 returned error can't find the container with id e48c8a12d183e258596aa4c10eacf8ff2cbe94f89874e3e923e2c7d69077f67d Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.108078 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.109054 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.109147 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.109181 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.109216 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.109237 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.109252 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.109271 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9hct\" (UniqueName: \"kubernetes.io/projected/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-kube-api-access-d9hct\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.109289 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.109570 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.109589 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.109777 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.110729 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.110867 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.112881 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.113723 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.115591 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.127434 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.127613 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.127846 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.129871 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.130073 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9hct\" (UniqueName: \"kubernetes.io/projected/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-kube-api-access-d9hct\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.146410 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.207813 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.242603 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3b81b6a9-0be6-4ce1-81b4-043edb1c221f","Type":"ContainerStarted","Data":"e48c8a12d183e258596aa4c10eacf8ff2cbe94f89874e3e923e2c7d69077f67d"} Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.246974 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8446fd7c75-b8hvv" event={"ID":"2d19854e-9f51-4d60-a6c1-6c1487f31d92","Type":"ContainerStarted","Data":"930fb6d111e16e19fbec709b72b94b609a9457f939308683b336f28471f7d648"} Dec 01 14:53:53 crc kubenswrapper[4810]: I1201 14:53:53.922840 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.257877 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"56bbf32f-35b5-4f92-bfc0-9266b0ecccac","Type":"ContainerStarted","Data":"93aeae8258afd7bcc2b35dc20a4f79760421b80ce5ada14ec46ca0fa7970ab90"} Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.378176 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.379706 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.386061 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.388509 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.388610 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.388750 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-5jwrv" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.390035 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.394701 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.532912 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c309abf-824f-4264-859b-ded1f0a9acaf-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.533617 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28jjc\" (UniqueName: \"kubernetes.io/projected/4c309abf-824f-4264-859b-ded1f0a9acaf-kube-api-access-28jjc\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.533685 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c309abf-824f-4264-859b-ded1f0a9acaf-operator-scripts\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.533738 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4c309abf-824f-4264-859b-ded1f0a9acaf-config-data-generated\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.533764 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c309abf-824f-4264-859b-ded1f0a9acaf-kolla-config\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.533796 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c309abf-824f-4264-859b-ded1f0a9acaf-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.533899 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.534004 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4c309abf-824f-4264-859b-ded1f0a9acaf-config-data-default\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.635122 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4c309abf-824f-4264-859b-ded1f0a9acaf-config-data-default\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.635206 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c309abf-824f-4264-859b-ded1f0a9acaf-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.635236 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28jjc\" (UniqueName: \"kubernetes.io/projected/4c309abf-824f-4264-859b-ded1f0a9acaf-kube-api-access-28jjc\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.635283 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c309abf-824f-4264-859b-ded1f0a9acaf-operator-scripts\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.635317 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4c309abf-824f-4264-859b-ded1f0a9acaf-config-data-generated\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.635343 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c309abf-824f-4264-859b-ded1f0a9acaf-kolla-config\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.635376 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c309abf-824f-4264-859b-ded1f0a9acaf-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.635552 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.636115 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.637007 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4c309abf-824f-4264-859b-ded1f0a9acaf-config-data-generated\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.637954 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.638421 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.638596 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.642525 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.647327 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c309abf-824f-4264-859b-ded1f0a9acaf-operator-scripts\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.648407 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c309abf-824f-4264-859b-ded1f0a9acaf-kolla-config\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.648645 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4c309abf-824f-4264-859b-ded1f0a9acaf-config-data-default\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.653480 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28jjc\" (UniqueName: \"kubernetes.io/projected/4c309abf-824f-4264-859b-ded1f0a9acaf-kube-api-access-28jjc\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.654033 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c309abf-824f-4264-859b-ded1f0a9acaf-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.654884 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c309abf-824f-4264-859b-ded1f0a9acaf-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.664689 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " pod="openstack/openstack-galera-0" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.708774 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-5jwrv" Dec 01 14:53:54 crc kubenswrapper[4810]: I1201 14:53:54.717904 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 01 14:53:55 crc kubenswrapper[4810]: I1201 14:53:55.304205 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 14:53:55 crc kubenswrapper[4810]: I1201 14:53:55.797992 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 14:53:55 crc kubenswrapper[4810]: I1201 14:53:55.800501 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:55 crc kubenswrapper[4810]: I1201 14:53:55.805451 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 14:53:55 crc kubenswrapper[4810]: I1201 14:53:55.808548 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 01 14:53:55 crc kubenswrapper[4810]: I1201 14:53:55.808748 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 01 14:53:55 crc kubenswrapper[4810]: I1201 14:53:55.809757 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 01 14:53:55 crc kubenswrapper[4810]: I1201 14:53:55.814724 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-vd46r" Dec 01 14:53:55 crc kubenswrapper[4810]: I1201 14:53:55.960690 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d202704a-7a2e-46a6-9e20-2eeabdc11814-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:55 crc kubenswrapper[4810]: I1201 14:53:55.960755 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d202704a-7a2e-46a6-9e20-2eeabdc11814-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:55 crc kubenswrapper[4810]: I1201 14:53:55.960799 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d202704a-7a2e-46a6-9e20-2eeabdc11814-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:55 crc kubenswrapper[4810]: I1201 14:53:55.960825 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:55 crc kubenswrapper[4810]: I1201 14:53:55.960869 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d202704a-7a2e-46a6-9e20-2eeabdc11814-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:55 crc kubenswrapper[4810]: I1201 14:53:55.960892 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d202704a-7a2e-46a6-9e20-2eeabdc11814-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:55 crc kubenswrapper[4810]: I1201 14:53:55.960918 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d202704a-7a2e-46a6-9e20-2eeabdc11814-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:55 crc kubenswrapper[4810]: I1201 14:53:55.960947 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8gbp\" (UniqueName: \"kubernetes.io/projected/d202704a-7a2e-46a6-9e20-2eeabdc11814-kube-api-access-d8gbp\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.067853 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d202704a-7a2e-46a6-9e20-2eeabdc11814-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.067919 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d202704a-7a2e-46a6-9e20-2eeabdc11814-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.067954 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.068005 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d202704a-7a2e-46a6-9e20-2eeabdc11814-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.068038 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d202704a-7a2e-46a6-9e20-2eeabdc11814-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.068072 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d202704a-7a2e-46a6-9e20-2eeabdc11814-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.068112 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8gbp\" (UniqueName: \"kubernetes.io/projected/d202704a-7a2e-46a6-9e20-2eeabdc11814-kube-api-access-d8gbp\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.068148 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d202704a-7a2e-46a6-9e20-2eeabdc11814-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.068339 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.068641 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d202704a-7a2e-46a6-9e20-2eeabdc11814-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.068827 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d202704a-7a2e-46a6-9e20-2eeabdc11814-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.068890 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d202704a-7a2e-46a6-9e20-2eeabdc11814-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.078867 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d202704a-7a2e-46a6-9e20-2eeabdc11814-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.082560 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d202704a-7a2e-46a6-9e20-2eeabdc11814-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.093410 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d202704a-7a2e-46a6-9e20-2eeabdc11814-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.094523 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8gbp\" (UniqueName: \"kubernetes.io/projected/d202704a-7a2e-46a6-9e20-2eeabdc11814-kube-api-access-d8gbp\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.134313 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.240236 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.241533 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.244893 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.246447 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-fhcqk" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.246595 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.258602 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.373073 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtzq4\" (UniqueName: \"kubernetes.io/projected/e16305af-6573-4cfd-ae76-f596d410920b-kube-api-access-wtzq4\") pod \"memcached-0\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " pod="openstack/memcached-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.373163 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e16305af-6573-4cfd-ae76-f596d410920b-config-data\") pod \"memcached-0\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " pod="openstack/memcached-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.373205 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e16305af-6573-4cfd-ae76-f596d410920b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " pod="openstack/memcached-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.373252 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e16305af-6573-4cfd-ae76-f596d410920b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " pod="openstack/memcached-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.373270 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e16305af-6573-4cfd-ae76-f596d410920b-kolla-config\") pod \"memcached-0\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " pod="openstack/memcached-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.437498 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.474512 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e16305af-6573-4cfd-ae76-f596d410920b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " pod="openstack/memcached-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.474565 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e16305af-6573-4cfd-ae76-f596d410920b-kolla-config\") pod \"memcached-0\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " pod="openstack/memcached-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.474648 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtzq4\" (UniqueName: \"kubernetes.io/projected/e16305af-6573-4cfd-ae76-f596d410920b-kube-api-access-wtzq4\") pod \"memcached-0\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " pod="openstack/memcached-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.474705 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e16305af-6573-4cfd-ae76-f596d410920b-config-data\") pod \"memcached-0\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " pod="openstack/memcached-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.474741 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e16305af-6573-4cfd-ae76-f596d410920b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " pod="openstack/memcached-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.475843 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e16305af-6573-4cfd-ae76-f596d410920b-kolla-config\") pod \"memcached-0\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " pod="openstack/memcached-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.478541 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e16305af-6573-4cfd-ae76-f596d410920b-config-data\") pod \"memcached-0\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " pod="openstack/memcached-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.480783 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e16305af-6573-4cfd-ae76-f596d410920b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " pod="openstack/memcached-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.503806 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e16305af-6573-4cfd-ae76-f596d410920b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " pod="openstack/memcached-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.506208 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtzq4\" (UniqueName: \"kubernetes.io/projected/e16305af-6573-4cfd-ae76-f596d410920b-kube-api-access-wtzq4\") pod \"memcached-0\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " pod="openstack/memcached-0" Dec 01 14:53:56 crc kubenswrapper[4810]: I1201 14:53:56.581666 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 01 14:53:57 crc kubenswrapper[4810]: I1201 14:53:57.723648 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 14:53:57 crc kubenswrapper[4810]: I1201 14:53:57.727103 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 14:53:57 crc kubenswrapper[4810]: I1201 14:53:57.729558 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-5787b" Dec 01 14:53:57 crc kubenswrapper[4810]: I1201 14:53:57.733551 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 14:53:57 crc kubenswrapper[4810]: I1201 14:53:57.902850 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxgq5\" (UniqueName: \"kubernetes.io/projected/563bcc28-df47-4e24-a246-9b713ad5cbb3-kube-api-access-hxgq5\") pod \"kube-state-metrics-0\" (UID: \"563bcc28-df47-4e24-a246-9b713ad5cbb3\") " pod="openstack/kube-state-metrics-0" Dec 01 14:53:58 crc kubenswrapper[4810]: I1201 14:53:58.004392 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxgq5\" (UniqueName: \"kubernetes.io/projected/563bcc28-df47-4e24-a246-9b713ad5cbb3-kube-api-access-hxgq5\") pod \"kube-state-metrics-0\" (UID: \"563bcc28-df47-4e24-a246-9b713ad5cbb3\") " pod="openstack/kube-state-metrics-0" Dec 01 14:53:58 crc kubenswrapper[4810]: I1201 14:53:58.023075 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxgq5\" (UniqueName: \"kubernetes.io/projected/563bcc28-df47-4e24-a246-9b713ad5cbb3-kube-api-access-hxgq5\") pod \"kube-state-metrics-0\" (UID: \"563bcc28-df47-4e24-a246-9b713ad5cbb3\") " pod="openstack/kube-state-metrics-0" Dec 01 14:53:58 crc kubenswrapper[4810]: I1201 14:53:58.046219 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 14:54:00 crc kubenswrapper[4810]: I1201 14:54:00.317409 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4c309abf-824f-4264-859b-ded1f0a9acaf","Type":"ContainerStarted","Data":"37b9fc728b3901032eb1ebaa116b2a4617e7fa0ed61602d41ae1488dd53881a9"} Dec 01 14:54:01 crc kubenswrapper[4810]: I1201 14:54:01.978411 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-hfmpn"] Dec 01 14:54:01 crc kubenswrapper[4810]: I1201 14:54:01.979965 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:01 crc kubenswrapper[4810]: I1201 14:54:01.981933 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 01 14:54:01 crc kubenswrapper[4810]: I1201 14:54:01.982337 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 01 14:54:01 crc kubenswrapper[4810]: I1201 14:54:01.982343 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-8n7sp" Dec 01 14:54:01 crc kubenswrapper[4810]: I1201 14:54:01.986725 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-x82dv"] Dec 01 14:54:01 crc kubenswrapper[4810]: I1201 14:54:01.989054 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:01 crc kubenswrapper[4810]: I1201 14:54:01.995021 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hfmpn"] Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.005394 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-x82dv"] Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.058957 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88635a0d-cb9d-43c1-9086-5623c9a780f8-scripts\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.059005 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c699a0a-f268-419b-a829-a9ebaa52b322-scripts\") pod \"ovn-controller-ovs-x82dv\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.059026 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dm7m\" (UniqueName: \"kubernetes.io/projected/9c699a0a-f268-419b-a829-a9ebaa52b322-kube-api-access-7dm7m\") pod \"ovn-controller-ovs-x82dv\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.059074 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-var-run\") pod \"ovn-controller-ovs-x82dv\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.059123 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-etc-ovs\") pod \"ovn-controller-ovs-x82dv\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.059142 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/88635a0d-cb9d-43c1-9086-5623c9a780f8-var-log-ovn\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.059415 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/88635a0d-cb9d-43c1-9086-5623c9a780f8-var-run-ovn\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.059636 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/88635a0d-cb9d-43c1-9086-5623c9a780f8-var-run\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.059668 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/88635a0d-cb9d-43c1-9086-5623c9a780f8-ovn-controller-tls-certs\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.059734 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-var-lib\") pod \"ovn-controller-ovs-x82dv\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.059863 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6pz4\" (UniqueName: \"kubernetes.io/projected/88635a0d-cb9d-43c1-9086-5623c9a780f8-kube-api-access-b6pz4\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.060089 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-var-log\") pod \"ovn-controller-ovs-x82dv\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.060171 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88635a0d-cb9d-43c1-9086-5623c9a780f8-combined-ca-bundle\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162135 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-etc-ovs\") pod \"ovn-controller-ovs-x82dv\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162189 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/88635a0d-cb9d-43c1-9086-5623c9a780f8-var-log-ovn\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162241 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/88635a0d-cb9d-43c1-9086-5623c9a780f8-var-run-ovn\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162291 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/88635a0d-cb9d-43c1-9086-5623c9a780f8-var-run\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162325 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/88635a0d-cb9d-43c1-9086-5623c9a780f8-ovn-controller-tls-certs\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162358 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-var-lib\") pod \"ovn-controller-ovs-x82dv\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162385 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6pz4\" (UniqueName: \"kubernetes.io/projected/88635a0d-cb9d-43c1-9086-5623c9a780f8-kube-api-access-b6pz4\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162403 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-var-log\") pod \"ovn-controller-ovs-x82dv\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162428 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88635a0d-cb9d-43c1-9086-5623c9a780f8-combined-ca-bundle\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162458 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88635a0d-cb9d-43c1-9086-5623c9a780f8-scripts\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162498 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c699a0a-f268-419b-a829-a9ebaa52b322-scripts\") pod \"ovn-controller-ovs-x82dv\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162518 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dm7m\" (UniqueName: \"kubernetes.io/projected/9c699a0a-f268-419b-a829-a9ebaa52b322-kube-api-access-7dm7m\") pod \"ovn-controller-ovs-x82dv\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162541 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-var-run\") pod \"ovn-controller-ovs-x82dv\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162694 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-etc-ovs\") pod \"ovn-controller-ovs-x82dv\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162808 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/88635a0d-cb9d-43c1-9086-5623c9a780f8-var-log-ovn\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162853 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-var-log\") pod \"ovn-controller-ovs-x82dv\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162897 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/88635a0d-cb9d-43c1-9086-5623c9a780f8-var-run-ovn\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162936 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/88635a0d-cb9d-43c1-9086-5623c9a780f8-var-run\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.162948 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-var-run\") pod \"ovn-controller-ovs-x82dv\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.163401 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-var-lib\") pod \"ovn-controller-ovs-x82dv\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.165957 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88635a0d-cb9d-43c1-9086-5623c9a780f8-scripts\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.171046 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/88635a0d-cb9d-43c1-9086-5623c9a780f8-ovn-controller-tls-certs\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.176259 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c699a0a-f268-419b-a829-a9ebaa52b322-scripts\") pod \"ovn-controller-ovs-x82dv\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.178588 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88635a0d-cb9d-43c1-9086-5623c9a780f8-combined-ca-bundle\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.178609 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dm7m\" (UniqueName: \"kubernetes.io/projected/9c699a0a-f268-419b-a829-a9ebaa52b322-kube-api-access-7dm7m\") pod \"ovn-controller-ovs-x82dv\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.182986 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6pz4\" (UniqueName: \"kubernetes.io/projected/88635a0d-cb9d-43c1-9086-5623c9a780f8-kube-api-access-b6pz4\") pod \"ovn-controller-hfmpn\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.311036 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.318145 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.864897 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.866460 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.868634 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.868915 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-bkw2b" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.869046 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.869155 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.874711 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.877924 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.972402 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.972454 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.978862 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3798bfc4-fb37-48d1-8146-42d177cf7861-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.978917 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.978954 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3798bfc4-fb37-48d1-8146-42d177cf7861-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.978982 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3798bfc4-fb37-48d1-8146-42d177cf7861-config\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.979036 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3798bfc4-fb37-48d1-8146-42d177cf7861-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.979088 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3798bfc4-fb37-48d1-8146-42d177cf7861-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.979113 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gzl8\" (UniqueName: \"kubernetes.io/projected/3798bfc4-fb37-48d1-8146-42d177cf7861-kube-api-access-9gzl8\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:02 crc kubenswrapper[4810]: I1201 14:54:02.979168 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3798bfc4-fb37-48d1-8146-42d177cf7861-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.084325 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3798bfc4-fb37-48d1-8146-42d177cf7861-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.084372 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3798bfc4-fb37-48d1-8146-42d177cf7861-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.084398 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.084423 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3798bfc4-fb37-48d1-8146-42d177cf7861-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.084441 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3798bfc4-fb37-48d1-8146-42d177cf7861-config\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.084493 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3798bfc4-fb37-48d1-8146-42d177cf7861-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.084530 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3798bfc4-fb37-48d1-8146-42d177cf7861-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.084550 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gzl8\" (UniqueName: \"kubernetes.io/projected/3798bfc4-fb37-48d1-8146-42d177cf7861-kube-api-access-9gzl8\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.093336 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3798bfc4-fb37-48d1-8146-42d177cf7861-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.094227 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3798bfc4-fb37-48d1-8146-42d177cf7861-config\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.094537 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3798bfc4-fb37-48d1-8146-42d177cf7861-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.094814 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.118129 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3798bfc4-fb37-48d1-8146-42d177cf7861-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.118150 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3798bfc4-fb37-48d1-8146-42d177cf7861-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.120109 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3798bfc4-fb37-48d1-8146-42d177cf7861-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.151387 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gzl8\" (UniqueName: \"kubernetes.io/projected/3798bfc4-fb37-48d1-8146-42d177cf7861-kube-api-access-9gzl8\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.187334 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.427502 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 01 14:54:03 crc kubenswrapper[4810]: I1201 14:54:03.487506 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.232352 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.234059 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.236644 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.236774 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.236795 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-smplj" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.236844 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.244297 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.335869 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.336243 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.336362 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.336534 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.336690 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-config\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.336946 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.337263 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kczs5\" (UniqueName: \"kubernetes.io/projected/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-kube-api-access-kczs5\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.337304 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.438416 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-config\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.438556 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.438692 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kczs5\" (UniqueName: \"kubernetes.io/projected/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-kube-api-access-kczs5\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.438718 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.438755 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.438804 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.438822 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.438845 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.439217 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.439541 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.440139 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.440739 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-config\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.448233 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.448420 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.450338 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.455682 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kczs5\" (UniqueName: \"kubernetes.io/projected/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-kube-api-access-kczs5\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.460160 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:05 crc kubenswrapper[4810]: I1201 14:54:05.557071 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:08 crc kubenswrapper[4810]: W1201 14:54:08.034442 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode16305af_6573_4cfd_ae76_f596d410920b.slice/crio-6f18f95df324cde3ccd3262cfa697fd4db9f0bc111e923c5a0bca2ed88be7313 WatchSource:0}: Error finding container 6f18f95df324cde3ccd3262cfa697fd4db9f0bc111e923c5a0bca2ed88be7313: Status 404 returned error can't find the container with id 6f18f95df324cde3ccd3262cfa697fd4db9f0bc111e923c5a0bca2ed88be7313 Dec 01 14:54:08 crc kubenswrapper[4810]: I1201 14:54:08.378403 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e16305af-6573-4cfd-ae76-f596d410920b","Type":"ContainerStarted","Data":"6f18f95df324cde3ccd3262cfa697fd4db9f0bc111e923c5a0bca2ed88be7313"} Dec 01 14:54:08 crc kubenswrapper[4810]: I1201 14:54:08.434594 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 14:54:08 crc kubenswrapper[4810]: E1201 14:54:08.769811 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627" Dec 01 14:54:08 crc kubenswrapper[4810]: E1201 14:54:08.770008 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jqq4w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-766fdc659c-z6tgt_openstack(58151c02-5c33-4721-b2ec-17c1921b8e2a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:54:08 crc kubenswrapper[4810]: E1201 14:54:08.772238 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-766fdc659c-z6tgt" podUID="58151c02-5c33-4721-b2ec-17c1921b8e2a" Dec 01 14:54:15 crc kubenswrapper[4810]: E1201 14:54:15.526628 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627" Dec 01 14:54:15 crc kubenswrapper[4810]: E1201 14:54:15.527246 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-94c5j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57dc4c6697-vw5rh_openstack(3383b93f-a184-4b6e-ad59-313e19cf4371): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:54:15 crc kubenswrapper[4810]: E1201 14:54:15.528426 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57dc4c6697-vw5rh" podUID="3383b93f-a184-4b6e-ad59-313e19cf4371" Dec 01 14:54:15 crc kubenswrapper[4810]: E1201 14:54:15.545306 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627" Dec 01 14:54:15 crc kubenswrapper[4810]: E1201 14:54:15.546029 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jwk7g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-557f57d995-vm4pr_openstack(714a979a-695b-4879-82b8-13b7068f124c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:54:15 crc kubenswrapper[4810]: E1201 14:54:15.546900 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627" Dec 01 14:54:15 crc kubenswrapper[4810]: E1201 14:54:15.546979 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vgpb9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-8446fd7c75-b8hvv_openstack(2d19854e-9f51-4d60-a6c1-6c1487f31d92): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:54:15 crc kubenswrapper[4810]: E1201 14:54:15.547433 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-557f57d995-vm4pr" podUID="714a979a-695b-4879-82b8-13b7068f124c" Dec 01 14:54:15 crc kubenswrapper[4810]: E1201 14:54:15.551741 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-8446fd7c75-b8hvv" podUID="2d19854e-9f51-4d60-a6c1-6c1487f31d92" Dec 01 14:54:15 crc kubenswrapper[4810]: I1201 14:54:15.641743 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766fdc659c-z6tgt" Dec 01 14:54:15 crc kubenswrapper[4810]: I1201 14:54:15.816879 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqq4w\" (UniqueName: \"kubernetes.io/projected/58151c02-5c33-4721-b2ec-17c1921b8e2a-kube-api-access-jqq4w\") pod \"58151c02-5c33-4721-b2ec-17c1921b8e2a\" (UID: \"58151c02-5c33-4721-b2ec-17c1921b8e2a\") " Dec 01 14:54:15 crc kubenswrapper[4810]: I1201 14:54:15.817325 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58151c02-5c33-4721-b2ec-17c1921b8e2a-config\") pod \"58151c02-5c33-4721-b2ec-17c1921b8e2a\" (UID: \"58151c02-5c33-4721-b2ec-17c1921b8e2a\") " Dec 01 14:54:15 crc kubenswrapper[4810]: I1201 14:54:15.817387 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58151c02-5c33-4721-b2ec-17c1921b8e2a-dns-svc\") pod \"58151c02-5c33-4721-b2ec-17c1921b8e2a\" (UID: \"58151c02-5c33-4721-b2ec-17c1921b8e2a\") " Dec 01 14:54:15 crc kubenswrapper[4810]: I1201 14:54:15.818128 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58151c02-5c33-4721-b2ec-17c1921b8e2a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "58151c02-5c33-4721-b2ec-17c1921b8e2a" (UID: "58151c02-5c33-4721-b2ec-17c1921b8e2a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:15 crc kubenswrapper[4810]: I1201 14:54:15.818506 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58151c02-5c33-4721-b2ec-17c1921b8e2a-config" (OuterVolumeSpecName: "config") pod "58151c02-5c33-4721-b2ec-17c1921b8e2a" (UID: "58151c02-5c33-4721-b2ec-17c1921b8e2a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:15 crc kubenswrapper[4810]: I1201 14:54:15.823654 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58151c02-5c33-4721-b2ec-17c1921b8e2a-kube-api-access-jqq4w" (OuterVolumeSpecName: "kube-api-access-jqq4w") pod "58151c02-5c33-4721-b2ec-17c1921b8e2a" (UID: "58151c02-5c33-4721-b2ec-17c1921b8e2a"). InnerVolumeSpecName "kube-api-access-jqq4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:54:15 crc kubenswrapper[4810]: I1201 14:54:15.918626 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/58151c02-5c33-4721-b2ec-17c1921b8e2a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:15 crc kubenswrapper[4810]: I1201 14:54:15.918657 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqq4w\" (UniqueName: \"kubernetes.io/projected/58151c02-5c33-4721-b2ec-17c1921b8e2a-kube-api-access-jqq4w\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:15 crc kubenswrapper[4810]: I1201 14:54:15.918672 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58151c02-5c33-4721-b2ec-17c1921b8e2a-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:15 crc kubenswrapper[4810]: I1201 14:54:15.945545 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hfmpn"] Dec 01 14:54:15 crc kubenswrapper[4810]: W1201 14:54:15.951919 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88635a0d_cb9d_43c1_9086_5623c9a780f8.slice/crio-b83750ee8d6e70fe32615bc44082261ab5ed545cea6cd0bf0ff4b0eacb8f8dc5 WatchSource:0}: Error finding container b83750ee8d6e70fe32615bc44082261ab5ed545cea6cd0bf0ff4b0eacb8f8dc5: Status 404 returned error can't find the container with id b83750ee8d6e70fe32615bc44082261ab5ed545cea6cd0bf0ff4b0eacb8f8dc5 Dec 01 14:54:16 crc kubenswrapper[4810]: I1201 14:54:16.034543 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 14:54:16 crc kubenswrapper[4810]: W1201 14:54:16.038089 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd202704a_7a2e_46a6_9e20_2eeabdc11814.slice/crio-118df97aecfd418517e3007d6833b5d4e2b18bd1489f3992ba59f20ddf95abdf WatchSource:0}: Error finding container 118df97aecfd418517e3007d6833b5d4e2b18bd1489f3992ba59f20ddf95abdf: Status 404 returned error can't find the container with id 118df97aecfd418517e3007d6833b5d4e2b18bd1489f3992ba59f20ddf95abdf Dec 01 14:54:16 crc kubenswrapper[4810]: I1201 14:54:16.187398 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 14:54:16 crc kubenswrapper[4810]: W1201 14:54:16.188767 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3798bfc4_fb37_48d1_8146_42d177cf7861.slice/crio-a30d683bfc270e3e352e7e4bd39449d321a965c5bf2257fce181fbfe593aa51c WatchSource:0}: Error finding container a30d683bfc270e3e352e7e4bd39449d321a965c5bf2257fce181fbfe593aa51c: Status 404 returned error can't find the container with id a30d683bfc270e3e352e7e4bd39449d321a965c5bf2257fce181fbfe593aa51c Dec 01 14:54:16 crc kubenswrapper[4810]: I1201 14:54:16.435260 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hfmpn" event={"ID":"88635a0d-cb9d-43c1-9086-5623c9a780f8","Type":"ContainerStarted","Data":"b83750ee8d6e70fe32615bc44082261ab5ed545cea6cd0bf0ff4b0eacb8f8dc5"} Dec 01 14:54:16 crc kubenswrapper[4810]: I1201 14:54:16.449051 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4c309abf-824f-4264-859b-ded1f0a9acaf","Type":"ContainerStarted","Data":"b5e8843e09485339399f86646edf6056d7c953642f1b41d38018280215001cb5"} Dec 01 14:54:16 crc kubenswrapper[4810]: I1201 14:54:16.453000 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-766fdc659c-z6tgt" event={"ID":"58151c02-5c33-4721-b2ec-17c1921b8e2a","Type":"ContainerDied","Data":"a4877255051a85acd3a9b9e804b1fbc6333ec2ce241bf7b0a1907a3b1450d117"} Dec 01 14:54:16 crc kubenswrapper[4810]: I1201 14:54:16.453081 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-766fdc659c-z6tgt" Dec 01 14:54:16 crc kubenswrapper[4810]: I1201 14:54:16.457838 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3798bfc4-fb37-48d1-8146-42d177cf7861","Type":"ContainerStarted","Data":"a30d683bfc270e3e352e7e4bd39449d321a965c5bf2257fce181fbfe593aa51c"} Dec 01 14:54:16 crc kubenswrapper[4810]: I1201 14:54:16.459262 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"563bcc28-df47-4e24-a246-9b713ad5cbb3","Type":"ContainerStarted","Data":"f3c2704bbef3d9d23a3b4ae03adb993df9e67183586c306521f4db470f8720c9"} Dec 01 14:54:16 crc kubenswrapper[4810]: I1201 14:54:16.464837 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d202704a-7a2e-46a6-9e20-2eeabdc11814","Type":"ContainerStarted","Data":"dbf605ce59fdfc665d554ecf7dd8225a9a123f40c558fd3683c4706be9b2e0a8"} Dec 01 14:54:16 crc kubenswrapper[4810]: I1201 14:54:16.464884 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d202704a-7a2e-46a6-9e20-2eeabdc11814","Type":"ContainerStarted","Data":"118df97aecfd418517e3007d6833b5d4e2b18bd1489f3992ba59f20ddf95abdf"} Dec 01 14:54:16 crc kubenswrapper[4810]: E1201 14:54:16.467170 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627\\\"\"" pod="openstack/dnsmasq-dns-8446fd7c75-b8hvv" podUID="2d19854e-9f51-4d60-a6c1-6c1487f31d92" Dec 01 14:54:16 crc kubenswrapper[4810]: E1201 14:54:16.467633 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627\\\"\"" pod="openstack/dnsmasq-dns-57dc4c6697-vw5rh" podUID="3383b93f-a184-4b6e-ad59-313e19cf4371" Dec 01 14:54:16 crc kubenswrapper[4810]: I1201 14:54:16.740938 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-766fdc659c-z6tgt"] Dec 01 14:54:16 crc kubenswrapper[4810]: I1201 14:54:16.748148 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-766fdc659c-z6tgt"] Dec 01 14:54:16 crc kubenswrapper[4810]: I1201 14:54:16.879335 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 14:54:17 crc kubenswrapper[4810]: W1201 14:54:17.235712 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod081ac19d_af0e_48f7_9c2e_0d282a9bb7ba.slice/crio-31554da4b756f2b0099d3a4a3d6c074cfcc4bca38c34e243462873eed47e271b WatchSource:0}: Error finding container 31554da4b756f2b0099d3a4a3d6c074cfcc4bca38c34e243462873eed47e271b: Status 404 returned error can't find the container with id 31554da4b756f2b0099d3a4a3d6c074cfcc4bca38c34e243462873eed47e271b Dec 01 14:54:17 crc kubenswrapper[4810]: I1201 14:54:17.247542 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-x82dv"] Dec 01 14:54:17 crc kubenswrapper[4810]: I1201 14:54:17.360007 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-557f57d995-vm4pr" Dec 01 14:54:17 crc kubenswrapper[4810]: I1201 14:54:17.447946 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwk7g\" (UniqueName: \"kubernetes.io/projected/714a979a-695b-4879-82b8-13b7068f124c-kube-api-access-jwk7g\") pod \"714a979a-695b-4879-82b8-13b7068f124c\" (UID: \"714a979a-695b-4879-82b8-13b7068f124c\") " Dec 01 14:54:17 crc kubenswrapper[4810]: I1201 14:54:17.448020 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/714a979a-695b-4879-82b8-13b7068f124c-config\") pod \"714a979a-695b-4879-82b8-13b7068f124c\" (UID: \"714a979a-695b-4879-82b8-13b7068f124c\") " Dec 01 14:54:17 crc kubenswrapper[4810]: I1201 14:54:17.448740 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/714a979a-695b-4879-82b8-13b7068f124c-config" (OuterVolumeSpecName: "config") pod "714a979a-695b-4879-82b8-13b7068f124c" (UID: "714a979a-695b-4879-82b8-13b7068f124c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:17 crc kubenswrapper[4810]: I1201 14:54:17.451078 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/714a979a-695b-4879-82b8-13b7068f124c-kube-api-access-jwk7g" (OuterVolumeSpecName: "kube-api-access-jwk7g") pod "714a979a-695b-4879-82b8-13b7068f124c" (UID: "714a979a-695b-4879-82b8-13b7068f124c"). InnerVolumeSpecName "kube-api-access-jwk7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:54:17 crc kubenswrapper[4810]: I1201 14:54:17.476228 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"56bbf32f-35b5-4f92-bfc0-9266b0ecccac","Type":"ContainerStarted","Data":"e23662080306dd3c3cc28ceb3e4c2195c124ff9bccb4a3034f0a68f6534ad3d3"} Dec 01 14:54:17 crc kubenswrapper[4810]: I1201 14:54:17.481408 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-557f57d995-vm4pr" Dec 01 14:54:17 crc kubenswrapper[4810]: I1201 14:54:17.481596 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-557f57d995-vm4pr" event={"ID":"714a979a-695b-4879-82b8-13b7068f124c","Type":"ContainerDied","Data":"c1e59ed03268bec1fee66f243cf43e18233f67980409d3eab1c8d011bfc61fae"} Dec 01 14:54:17 crc kubenswrapper[4810]: I1201 14:54:17.483499 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-x82dv" event={"ID":"9c699a0a-f268-419b-a829-a9ebaa52b322","Type":"ContainerStarted","Data":"ee51a4871d387f6ab294b0535c53206dcd1bf44b9ed26b71d9760d06883f5b7f"} Dec 01 14:54:17 crc kubenswrapper[4810]: I1201 14:54:17.486262 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3b81b6a9-0be6-4ce1-81b4-043edb1c221f","Type":"ContainerStarted","Data":"823b399b01a75b82b70c3653a5f597fb68c1ca3af1fbf06c04a2d840c74494cb"} Dec 01 14:54:17 crc kubenswrapper[4810]: I1201 14:54:17.489246 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba","Type":"ContainerStarted","Data":"31554da4b756f2b0099d3a4a3d6c074cfcc4bca38c34e243462873eed47e271b"} Dec 01 14:54:17 crc kubenswrapper[4810]: I1201 14:54:17.552954 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwk7g\" (UniqueName: \"kubernetes.io/projected/714a979a-695b-4879-82b8-13b7068f124c-kube-api-access-jwk7g\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:17 crc kubenswrapper[4810]: I1201 14:54:17.553249 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/714a979a-695b-4879-82b8-13b7068f124c-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:17 crc kubenswrapper[4810]: I1201 14:54:17.579622 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-557f57d995-vm4pr"] Dec 01 14:54:17 crc kubenswrapper[4810]: I1201 14:54:17.587117 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-557f57d995-vm4pr"] Dec 01 14:54:18 crc kubenswrapper[4810]: I1201 14:54:18.501594 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58151c02-5c33-4721-b2ec-17c1921b8e2a" path="/var/lib/kubelet/pods/58151c02-5c33-4721-b2ec-17c1921b8e2a/volumes" Dec 01 14:54:18 crc kubenswrapper[4810]: I1201 14:54:18.502026 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="714a979a-695b-4879-82b8-13b7068f124c" path="/var/lib/kubelet/pods/714a979a-695b-4879-82b8-13b7068f124c/volumes" Dec 01 14:54:18 crc kubenswrapper[4810]: I1201 14:54:18.502336 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e16305af-6573-4cfd-ae76-f596d410920b","Type":"ContainerStarted","Data":"6a66d0aa3094f13ae73e9e79b4f1f5db114907ff336ae2e84fa9c09a936675a3"} Dec 01 14:54:18 crc kubenswrapper[4810]: I1201 14:54:18.502359 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 01 14:54:18 crc kubenswrapper[4810]: I1201 14:54:18.517363 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=13.264325674 podStartE2EDuration="22.517347825s" podCreationTimestamp="2025-12-01 14:53:56 +0000 UTC" firstStartedPulling="2025-12-01 14:54:08.039397456 +0000 UTC m=+1213.802907059" lastFinishedPulling="2025-12-01 14:54:17.292419617 +0000 UTC m=+1223.055929210" observedRunningTime="2025-12-01 14:54:18.516910092 +0000 UTC m=+1224.280419695" watchObservedRunningTime="2025-12-01 14:54:18.517347825 +0000 UTC m=+1224.280857428" Dec 01 14:54:19 crc kubenswrapper[4810]: I1201 14:54:19.507201 4810 generic.go:334] "Generic (PLEG): container finished" podID="4c309abf-824f-4264-859b-ded1f0a9acaf" containerID="b5e8843e09485339399f86646edf6056d7c953642f1b41d38018280215001cb5" exitCode=0 Dec 01 14:54:19 crc kubenswrapper[4810]: I1201 14:54:19.507281 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4c309abf-824f-4264-859b-ded1f0a9acaf","Type":"ContainerDied","Data":"b5e8843e09485339399f86646edf6056d7c953642f1b41d38018280215001cb5"} Dec 01 14:54:20 crc kubenswrapper[4810]: I1201 14:54:20.524458 4810 generic.go:334] "Generic (PLEG): container finished" podID="d202704a-7a2e-46a6-9e20-2eeabdc11814" containerID="dbf605ce59fdfc665d554ecf7dd8225a9a123f40c558fd3683c4706be9b2e0a8" exitCode=0 Dec 01 14:54:20 crc kubenswrapper[4810]: I1201 14:54:20.524655 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d202704a-7a2e-46a6-9e20-2eeabdc11814","Type":"ContainerDied","Data":"dbf605ce59fdfc665d554ecf7dd8225a9a123f40c558fd3683c4706be9b2e0a8"} Dec 01 14:54:21 crc kubenswrapper[4810]: I1201 14:54:21.535052 4810 generic.go:334] "Generic (PLEG): container finished" podID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerID="0dd2c680c8d22899f94d6de624293194c03d0629c22fb4e022cc58fae7b6f0aa" exitCode=0 Dec 01 14:54:21 crc kubenswrapper[4810]: I1201 14:54:21.535102 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-x82dv" event={"ID":"9c699a0a-f268-419b-a829-a9ebaa52b322","Type":"ContainerDied","Data":"0dd2c680c8d22899f94d6de624293194c03d0629c22fb4e022cc58fae7b6f0aa"} Dec 01 14:54:21 crc kubenswrapper[4810]: I1201 14:54:21.539970 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba","Type":"ContainerStarted","Data":"2bb8d93447e17f91799220c99cae83b3109ca92b16e6bbce7a1bf9da7f081d60"} Dec 01 14:54:21 crc kubenswrapper[4810]: I1201 14:54:21.543253 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3798bfc4-fb37-48d1-8146-42d177cf7861","Type":"ContainerStarted","Data":"afd5408dd106d9875d318c4f58d0916d20a0514bcf3f71e7b57c27a91050bc80"} Dec 01 14:54:21 crc kubenswrapper[4810]: I1201 14:54:21.545721 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"563bcc28-df47-4e24-a246-9b713ad5cbb3","Type":"ContainerStarted","Data":"c8a73c8466b9b60faf6e449bb4dd1db5af7054b925a232c56b49850420483b2a"} Dec 01 14:54:21 crc kubenswrapper[4810]: I1201 14:54:21.545850 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 01 14:54:21 crc kubenswrapper[4810]: I1201 14:54:21.550862 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d202704a-7a2e-46a6-9e20-2eeabdc11814","Type":"ContainerStarted","Data":"f21f812c44b7972a5f6ecdd8dbe2ffd46fb63a8fa481dd14afa8b978ba21edfb"} Dec 01 14:54:21 crc kubenswrapper[4810]: I1201 14:54:21.556561 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hfmpn" event={"ID":"88635a0d-cb9d-43c1-9086-5623c9a780f8","Type":"ContainerStarted","Data":"1c3b60c932bcc8ce61ca6d0f28d631b382aa2b237627432ebf90c92ecb1485d0"} Dec 01 14:54:21 crc kubenswrapper[4810]: I1201 14:54:21.557554 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-hfmpn" Dec 01 14:54:21 crc kubenswrapper[4810]: I1201 14:54:21.561390 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4c309abf-824f-4264-859b-ded1f0a9acaf","Type":"ContainerStarted","Data":"1b9b162f9bfb794c639b6d75ebcad8b08b8c31a6a252c25ffdb8e930eaa77acf"} Dec 01 14:54:21 crc kubenswrapper[4810]: I1201 14:54:21.585019 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=19.616736025 podStartE2EDuration="24.585000349s" podCreationTimestamp="2025-12-01 14:53:57 +0000 UTC" firstStartedPulling="2025-12-01 14:54:15.463030753 +0000 UTC m=+1221.226540346" lastFinishedPulling="2025-12-01 14:54:20.431295067 +0000 UTC m=+1226.194804670" observedRunningTime="2025-12-01 14:54:21.576451647 +0000 UTC m=+1227.339961270" watchObservedRunningTime="2025-12-01 14:54:21.585000349 +0000 UTC m=+1227.348509962" Dec 01 14:54:21 crc kubenswrapper[4810]: I1201 14:54:21.637992 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-hfmpn" podStartSLOduration=16.122258221 podStartE2EDuration="20.637973647s" podCreationTimestamp="2025-12-01 14:54:01 +0000 UTC" firstStartedPulling="2025-12-01 14:54:15.954782826 +0000 UTC m=+1221.718292429" lastFinishedPulling="2025-12-01 14:54:20.470498242 +0000 UTC m=+1226.234007855" observedRunningTime="2025-12-01 14:54:21.629910699 +0000 UTC m=+1227.393420302" watchObservedRunningTime="2025-12-01 14:54:21.637973647 +0000 UTC m=+1227.401483250" Dec 01 14:54:21 crc kubenswrapper[4810]: I1201 14:54:21.638907 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=27.638901893 podStartE2EDuration="27.638901893s" podCreationTimestamp="2025-12-01 14:53:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:54:21.611417057 +0000 UTC m=+1227.374926700" watchObservedRunningTime="2025-12-01 14:54:21.638901893 +0000 UTC m=+1227.402411496" Dec 01 14:54:21 crc kubenswrapper[4810]: I1201 14:54:21.655245 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=12.466390285 podStartE2EDuration="28.655225627s" podCreationTimestamp="2025-12-01 14:53:53 +0000 UTC" firstStartedPulling="2025-12-01 14:53:59.457175764 +0000 UTC m=+1205.220685357" lastFinishedPulling="2025-12-01 14:54:15.646011096 +0000 UTC m=+1221.409520699" observedRunningTime="2025-12-01 14:54:21.653223812 +0000 UTC m=+1227.416733415" watchObservedRunningTime="2025-12-01 14:54:21.655225627 +0000 UTC m=+1227.418735230" Dec 01 14:54:22 crc kubenswrapper[4810]: I1201 14:54:22.573299 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-x82dv" event={"ID":"9c699a0a-f268-419b-a829-a9ebaa52b322","Type":"ContainerStarted","Data":"74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6"} Dec 01 14:54:22 crc kubenswrapper[4810]: I1201 14:54:22.573687 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:22 crc kubenswrapper[4810]: I1201 14:54:22.573735 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-x82dv" event={"ID":"9c699a0a-f268-419b-a829-a9ebaa52b322","Type":"ContainerStarted","Data":"0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22"} Dec 01 14:54:22 crc kubenswrapper[4810]: I1201 14:54:22.596436 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-x82dv" podStartSLOduration=18.41764059 podStartE2EDuration="21.596412104s" podCreationTimestamp="2025-12-01 14:54:01 +0000 UTC" firstStartedPulling="2025-12-01 14:54:17.255207636 +0000 UTC m=+1223.018717239" lastFinishedPulling="2025-12-01 14:54:20.43397915 +0000 UTC m=+1226.197488753" observedRunningTime="2025-12-01 14:54:22.592212859 +0000 UTC m=+1228.355722492" watchObservedRunningTime="2025-12-01 14:54:22.596412104 +0000 UTC m=+1228.359921707" Dec 01 14:54:23 crc kubenswrapper[4810]: I1201 14:54:23.582091 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:23 crc kubenswrapper[4810]: E1201 14:54:23.808418 4810 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.182:45900->38.102.83.182:41089: read tcp 38.102.83.182:45900->38.102.83.182:41089: read: connection reset by peer Dec 01 14:54:24 crc kubenswrapper[4810]: I1201 14:54:24.591797 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba","Type":"ContainerStarted","Data":"6fcdb0e74e2281103d173be66fd77a226460e47c85adf65b5481538a371e0a73"} Dec 01 14:54:24 crc kubenswrapper[4810]: I1201 14:54:24.593515 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3798bfc4-fb37-48d1-8146-42d177cf7861","Type":"ContainerStarted","Data":"d74f9aed9f56dc44f786f3d28286286934f841d05512a1633b5fe3b4cc217204"} Dec 01 14:54:24 crc kubenswrapper[4810]: I1201 14:54:24.623631 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=14.016845193 podStartE2EDuration="20.623610653s" podCreationTimestamp="2025-12-01 14:54:04 +0000 UTC" firstStartedPulling="2025-12-01 14:54:17.24138859 +0000 UTC m=+1223.004898183" lastFinishedPulling="2025-12-01 14:54:23.84815404 +0000 UTC m=+1229.611663643" observedRunningTime="2025-12-01 14:54:24.612295296 +0000 UTC m=+1230.375804899" watchObservedRunningTime="2025-12-01 14:54:24.623610653 +0000 UTC m=+1230.387120266" Dec 01 14:54:24 crc kubenswrapper[4810]: I1201 14:54:24.637949 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=15.99962148 podStartE2EDuration="23.637929492s" podCreationTimestamp="2025-12-01 14:54:01 +0000 UTC" firstStartedPulling="2025-12-01 14:54:16.19114082 +0000 UTC m=+1221.954650423" lastFinishedPulling="2025-12-01 14:54:23.829448832 +0000 UTC m=+1229.592958435" observedRunningTime="2025-12-01 14:54:24.628975819 +0000 UTC m=+1230.392485452" watchObservedRunningTime="2025-12-01 14:54:24.637929492 +0000 UTC m=+1230.401439105" Dec 01 14:54:24 crc kubenswrapper[4810]: I1201 14:54:24.719031 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 01 14:54:24 crc kubenswrapper[4810]: I1201 14:54:24.719093 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.066515 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-6lqnm"] Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.067590 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.072426 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.078716 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-6lqnm"] Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.178199 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bp2d\" (UniqueName: \"kubernetes.io/projected/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-kube-api-access-7bp2d\") pod \"ovn-controller-metrics-6lqnm\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.178268 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-ovn-rundir\") pod \"ovn-controller-metrics-6lqnm\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.178320 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-config\") pod \"ovn-controller-metrics-6lqnm\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.178393 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-ovs-rundir\") pod \"ovn-controller-metrics-6lqnm\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.178431 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-6lqnm\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.178450 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-combined-ca-bundle\") pod \"ovn-controller-metrics-6lqnm\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.230379 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57dc4c6697-vw5rh"] Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.270576 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b4845d8d5-hfkzl"] Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.271761 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.273483 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.279734 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-ovs-rundir\") pod \"ovn-controller-metrics-6lqnm\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.279785 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-6lqnm\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.279812 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-combined-ca-bundle\") pod \"ovn-controller-metrics-6lqnm\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.279844 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bp2d\" (UniqueName: \"kubernetes.io/projected/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-kube-api-access-7bp2d\") pod \"ovn-controller-metrics-6lqnm\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.279885 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-ovn-rundir\") pod \"ovn-controller-metrics-6lqnm\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.279906 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-config\") pod \"ovn-controller-metrics-6lqnm\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.280822 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-config\") pod \"ovn-controller-metrics-6lqnm\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.281102 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-ovn-rundir\") pod \"ovn-controller-metrics-6lqnm\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.284248 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-ovs-rundir\") pod \"ovn-controller-metrics-6lqnm\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.295519 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-combined-ca-bundle\") pod \"ovn-controller-metrics-6lqnm\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.296672 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-6lqnm\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.298626 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b4845d8d5-hfkzl"] Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.307638 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bp2d\" (UniqueName: \"kubernetes.io/projected/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-kube-api-access-7bp2d\") pod \"ovn-controller-metrics-6lqnm\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.385170 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-ovsdbserver-nb\") pod \"dnsmasq-dns-b4845d8d5-hfkzl\" (UID: \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\") " pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.385490 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6475\" (UniqueName: \"kubernetes.io/projected/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-kube-api-access-d6475\") pod \"dnsmasq-dns-b4845d8d5-hfkzl\" (UID: \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\") " pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.385672 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-config\") pod \"dnsmasq-dns-b4845d8d5-hfkzl\" (UID: \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\") " pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.385763 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-dns-svc\") pod \"dnsmasq-dns-b4845d8d5-hfkzl\" (UID: \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\") " pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.398123 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.427949 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8446fd7c75-b8hvv"] Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.479132 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bd7c66845-thth9"] Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.480631 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.482958 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.493196 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-dns-svc\") pod \"dnsmasq-dns-b4845d8d5-hfkzl\" (UID: \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\") " pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.493408 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-ovsdbserver-nb\") pod \"dnsmasq-dns-b4845d8d5-hfkzl\" (UID: \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\") " pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.493543 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6475\" (UniqueName: \"kubernetes.io/projected/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-kube-api-access-d6475\") pod \"dnsmasq-dns-b4845d8d5-hfkzl\" (UID: \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\") " pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.493735 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-config\") pod \"dnsmasq-dns-b4845d8d5-hfkzl\" (UID: \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\") " pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.496977 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-ovsdbserver-nb\") pod \"dnsmasq-dns-b4845d8d5-hfkzl\" (UID: \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\") " pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.497101 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-config\") pod \"dnsmasq-dns-b4845d8d5-hfkzl\" (UID: \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\") " pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.504946 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-dns-svc\") pod \"dnsmasq-dns-b4845d8d5-hfkzl\" (UID: \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\") " pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.514204 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6475\" (UniqueName: \"kubernetes.io/projected/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-kube-api-access-d6475\") pod \"dnsmasq-dns-b4845d8d5-hfkzl\" (UID: \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\") " pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.527340 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bd7c66845-thth9"] Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.559136 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.600013 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-ovsdbserver-nb\") pod \"dnsmasq-dns-5bd7c66845-thth9\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.600312 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-config\") pod \"dnsmasq-dns-5bd7c66845-thth9\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.600340 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5vkf\" (UniqueName: \"kubernetes.io/projected/dbc8bb26-f165-4960-becf-661fc9b67eed-kube-api-access-g5vkf\") pod \"dnsmasq-dns-5bd7c66845-thth9\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.600368 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-ovsdbserver-sb\") pod \"dnsmasq-dns-5bd7c66845-thth9\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.600405 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-dns-svc\") pod \"dnsmasq-dns-5bd7c66845-thth9\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.645433 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.695935 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57dc4c6697-vw5rh" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.701669 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-ovsdbserver-nb\") pod \"dnsmasq-dns-5bd7c66845-thth9\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.701727 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-config\") pod \"dnsmasq-dns-5bd7c66845-thth9\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.701756 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5vkf\" (UniqueName: \"kubernetes.io/projected/dbc8bb26-f165-4960-becf-661fc9b67eed-kube-api-access-g5vkf\") pod \"dnsmasq-dns-5bd7c66845-thth9\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.701804 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-ovsdbserver-sb\") pod \"dnsmasq-dns-5bd7c66845-thth9\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.701830 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-dns-svc\") pod \"dnsmasq-dns-5bd7c66845-thth9\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.703827 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-config\") pod \"dnsmasq-dns-5bd7c66845-thth9\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.704023 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-ovsdbserver-sb\") pod \"dnsmasq-dns-5bd7c66845-thth9\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.704046 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-dns-svc\") pod \"dnsmasq-dns-5bd7c66845-thth9\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.704361 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-ovsdbserver-nb\") pod \"dnsmasq-dns-5bd7c66845-thth9\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.720414 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5vkf\" (UniqueName: \"kubernetes.io/projected/dbc8bb26-f165-4960-becf-661fc9b67eed-kube-api-access-g5vkf\") pod \"dnsmasq-dns-5bd7c66845-thth9\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.803644 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3383b93f-a184-4b6e-ad59-313e19cf4371-config\") pod \"3383b93f-a184-4b6e-ad59-313e19cf4371\" (UID: \"3383b93f-a184-4b6e-ad59-313e19cf4371\") " Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.804099 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3383b93f-a184-4b6e-ad59-313e19cf4371-config" (OuterVolumeSpecName: "config") pod "3383b93f-a184-4b6e-ad59-313e19cf4371" (UID: "3383b93f-a184-4b6e-ad59-313e19cf4371"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.804375 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3383b93f-a184-4b6e-ad59-313e19cf4371-dns-svc\") pod \"3383b93f-a184-4b6e-ad59-313e19cf4371\" (UID: \"3383b93f-a184-4b6e-ad59-313e19cf4371\") " Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.804831 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3383b93f-a184-4b6e-ad59-313e19cf4371-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3383b93f-a184-4b6e-ad59-313e19cf4371" (UID: "3383b93f-a184-4b6e-ad59-313e19cf4371"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.804984 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94c5j\" (UniqueName: \"kubernetes.io/projected/3383b93f-a184-4b6e-ad59-313e19cf4371-kube-api-access-94c5j\") pod \"3383b93f-a184-4b6e-ad59-313e19cf4371\" (UID: \"3383b93f-a184-4b6e-ad59-313e19cf4371\") " Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.805932 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3383b93f-a184-4b6e-ad59-313e19cf4371-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.805953 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3383b93f-a184-4b6e-ad59-313e19cf4371-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.812447 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3383b93f-a184-4b6e-ad59-313e19cf4371-kube-api-access-94c5j" (OuterVolumeSpecName: "kube-api-access-94c5j") pod "3383b93f-a184-4b6e-ad59-313e19cf4371" (UID: "3383b93f-a184-4b6e-ad59-313e19cf4371"). InnerVolumeSpecName "kube-api-access-94c5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.856196 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8446fd7c75-b8hvv" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.870087 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.908625 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgpb9\" (UniqueName: \"kubernetes.io/projected/2d19854e-9f51-4d60-a6c1-6c1487f31d92-kube-api-access-vgpb9\") pod \"2d19854e-9f51-4d60-a6c1-6c1487f31d92\" (UID: \"2d19854e-9f51-4d60-a6c1-6c1487f31d92\") " Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.908698 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d19854e-9f51-4d60-a6c1-6c1487f31d92-config\") pod \"2d19854e-9f51-4d60-a6c1-6c1487f31d92\" (UID: \"2d19854e-9f51-4d60-a6c1-6c1487f31d92\") " Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.908731 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d19854e-9f51-4d60-a6c1-6c1487f31d92-dns-svc\") pod \"2d19854e-9f51-4d60-a6c1-6c1487f31d92\" (UID: \"2d19854e-9f51-4d60-a6c1-6c1487f31d92\") " Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.909161 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94c5j\" (UniqueName: \"kubernetes.io/projected/3383b93f-a184-4b6e-ad59-313e19cf4371-kube-api-access-94c5j\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.909627 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d19854e-9f51-4d60-a6c1-6c1487f31d92-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2d19854e-9f51-4d60-a6c1-6c1487f31d92" (UID: "2d19854e-9f51-4d60-a6c1-6c1487f31d92"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.912571 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d19854e-9f51-4d60-a6c1-6c1487f31d92-config" (OuterVolumeSpecName: "config") pod "2d19854e-9f51-4d60-a6c1-6c1487f31d92" (UID: "2d19854e-9f51-4d60-a6c1-6c1487f31d92"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:25 crc kubenswrapper[4810]: I1201 14:54:25.913635 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d19854e-9f51-4d60-a6c1-6c1487f31d92-kube-api-access-vgpb9" (OuterVolumeSpecName: "kube-api-access-vgpb9") pod "2d19854e-9f51-4d60-a6c1-6c1487f31d92" (UID: "2d19854e-9f51-4d60-a6c1-6c1487f31d92"). InnerVolumeSpecName "kube-api-access-vgpb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.012776 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgpb9\" (UniqueName: \"kubernetes.io/projected/2d19854e-9f51-4d60-a6c1-6c1487f31d92-kube-api-access-vgpb9\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.012815 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d19854e-9f51-4d60-a6c1-6c1487f31d92-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.012827 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d19854e-9f51-4d60-a6c1-6c1487f31d92-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.046756 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-6lqnm"] Dec 01 14:54:26 crc kubenswrapper[4810]: W1201 14:54:26.056870 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd2094a7_9a7b_417b_b4ea_b1da8b69cac6.slice/crio-d63d727c2af3793d4c63ed53b1bc16ab310573f3fe15ec6f8f985db2fab60875 WatchSource:0}: Error finding container d63d727c2af3793d4c63ed53b1bc16ab310573f3fe15ec6f8f985db2fab60875: Status 404 returned error can't find the container with id d63d727c2af3793d4c63ed53b1bc16ab310573f3fe15ec6f8f985db2fab60875 Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.223285 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b4845d8d5-hfkzl"] Dec 01 14:54:26 crc kubenswrapper[4810]: W1201 14:54:26.233194 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fb61aa6_dd8f_4fb5_9697_76dc914ef2a9.slice/crio-a3372657fecd36a885e3d0b8b019c28a58ca7686ff6984e079f04bb72ac1c0d2 WatchSource:0}: Error finding container a3372657fecd36a885e3d0b8b019c28a58ca7686ff6984e079f04bb72ac1c0d2: Status 404 returned error can't find the container with id a3372657fecd36a885e3d0b8b019c28a58ca7686ff6984e079f04bb72ac1c0d2 Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.387065 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bd7c66845-thth9"] Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.438637 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.438680 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.504412 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.557186 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.583420 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.597509 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.630100 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" event={"ID":"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9","Type":"ContainerStarted","Data":"a3372657fecd36a885e3d0b8b019c28a58ca7686ff6984e079f04bb72ac1c0d2"} Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.631965 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57dc4c6697-vw5rh" Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.631936 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57dc4c6697-vw5rh" event={"ID":"3383b93f-a184-4b6e-ad59-313e19cf4371","Type":"ContainerDied","Data":"5277866b40136ef3325b00ba3ad0cea015c3b81aa433de751da0201942b891d1"} Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.633403 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-6lqnm" event={"ID":"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6","Type":"ContainerStarted","Data":"d78defca92531f2b0db0c17c5224178008057f1cf21db48f31f088bda1fc039b"} Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.633444 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-6lqnm" event={"ID":"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6","Type":"ContainerStarted","Data":"d63d727c2af3793d4c63ed53b1bc16ab310573f3fe15ec6f8f985db2fab60875"} Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.634660 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd7c66845-thth9" event={"ID":"dbc8bb26-f165-4960-becf-661fc9b67eed","Type":"ContainerStarted","Data":"c8ce3082ea49e1b755c58768d704ccd938686b5077e56bc186480851cff88c95"} Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.643386 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8446fd7c75-b8hvv" event={"ID":"2d19854e-9f51-4d60-a6c1-6c1487f31d92","Type":"ContainerDied","Data":"930fb6d111e16e19fbec709b72b94b609a9457f939308683b336f28471f7d648"} Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.643511 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8446fd7c75-b8hvv" Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.678610 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-6lqnm" podStartSLOduration=1.678586417 podStartE2EDuration="1.678586417s" podCreationTimestamp="2025-12-01 14:54:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:54:26.673145029 +0000 UTC m=+1232.436654632" watchObservedRunningTime="2025-12-01 14:54:26.678586417 +0000 UTC m=+1232.442096020" Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.714260 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8446fd7c75-b8hvv"] Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.717027 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.719330 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8446fd7c75-b8hvv"] Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.760061 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57dc4c6697-vw5rh"] Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.776267 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57dc4c6697-vw5rh"] Dec 01 14:54:26 crc kubenswrapper[4810]: I1201 14:54:26.835322 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 01 14:54:27 crc kubenswrapper[4810]: I1201 14:54:27.487805 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:27 crc kubenswrapper[4810]: I1201 14:54:27.538980 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:27 crc kubenswrapper[4810]: I1201 14:54:27.657999 4810 generic.go:334] "Generic (PLEG): container finished" podID="2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9" containerID="f153efb28861380862b3940a8f955f575482e7b5a15dab4443b07455b5d8c644" exitCode=0 Dec 01 14:54:27 crc kubenswrapper[4810]: I1201 14:54:27.658106 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" event={"ID":"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9","Type":"ContainerDied","Data":"f153efb28861380862b3940a8f955f575482e7b5a15dab4443b07455b5d8c644"} Dec 01 14:54:27 crc kubenswrapper[4810]: I1201 14:54:27.659931 4810 generic.go:334] "Generic (PLEG): container finished" podID="dbc8bb26-f165-4960-becf-661fc9b67eed" containerID="b417098a11c1d128d79133f905054b971c187e966f0cc4420591b5298add921d" exitCode=0 Dec 01 14:54:27 crc kubenswrapper[4810]: I1201 14:54:27.660774 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd7c66845-thth9" event={"ID":"dbc8bb26-f165-4960-becf-661fc9b67eed","Type":"ContainerDied","Data":"b417098a11c1d128d79133f905054b971c187e966f0cc4420591b5298add921d"} Dec 01 14:54:27 crc kubenswrapper[4810]: I1201 14:54:27.661521 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:27 crc kubenswrapper[4810]: I1201 14:54:27.702889 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 01 14:54:27 crc kubenswrapper[4810]: I1201 14:54:27.932594 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 01 14:54:27 crc kubenswrapper[4810]: I1201 14:54:27.934365 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 01 14:54:27 crc kubenswrapper[4810]: I1201 14:54:27.937421 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-8qmwp" Dec 01 14:54:27 crc kubenswrapper[4810]: I1201 14:54:27.944360 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 01 14:54:27 crc kubenswrapper[4810]: I1201 14:54:27.944604 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 01 14:54:27 crc kubenswrapper[4810]: I1201 14:54:27.944737 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 01 14:54:27 crc kubenswrapper[4810]: I1201 14:54:27.959228 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.029498 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bd7c66845-thth9"] Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.047039 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.047120 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.047192 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.047248 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.047348 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-scripts\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.047488 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8cqd\" (UniqueName: \"kubernetes.io/projected/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-kube-api-access-f8cqd\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.047520 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-config\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.052096 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.087300 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f6d79597f-wvf92"] Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.088582 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.111766 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f6d79597f-wvf92"] Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.150361 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8cqd\" (UniqueName: \"kubernetes.io/projected/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-kube-api-access-f8cqd\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.150408 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-config\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.150485 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.150523 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.150561 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.150607 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.150628 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-scripts\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.151488 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-scripts\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.151738 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-config\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.152615 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.159556 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.159757 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.162205 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.172855 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8cqd\" (UniqueName: \"kubernetes.io/projected/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-kube-api-access-f8cqd\") pod \"ovn-northd-0\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.252188 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-config\") pod \"dnsmasq-dns-5f6d79597f-wvf92\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.252252 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42ztn\" (UniqueName: \"kubernetes.io/projected/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-kube-api-access-42ztn\") pod \"dnsmasq-dns-5f6d79597f-wvf92\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.252364 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-ovsdbserver-sb\") pod \"dnsmasq-dns-5f6d79597f-wvf92\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.252411 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-ovsdbserver-nb\") pod \"dnsmasq-dns-5f6d79597f-wvf92\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.252429 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-dns-svc\") pod \"dnsmasq-dns-5f6d79597f-wvf92\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.267032 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.353436 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-ovsdbserver-sb\") pod \"dnsmasq-dns-5f6d79597f-wvf92\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.353665 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-ovsdbserver-nb\") pod \"dnsmasq-dns-5f6d79597f-wvf92\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.353688 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-dns-svc\") pod \"dnsmasq-dns-5f6d79597f-wvf92\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.353731 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-config\") pod \"dnsmasq-dns-5f6d79597f-wvf92\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.353766 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42ztn\" (UniqueName: \"kubernetes.io/projected/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-kube-api-access-42ztn\") pod \"dnsmasq-dns-5f6d79597f-wvf92\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.354870 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-dns-svc\") pod \"dnsmasq-dns-5f6d79597f-wvf92\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.355025 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-ovsdbserver-nb\") pod \"dnsmasq-dns-5f6d79597f-wvf92\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.355932 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-config\") pod \"dnsmasq-dns-5f6d79597f-wvf92\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.359272 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-ovsdbserver-sb\") pod \"dnsmasq-dns-5f6d79597f-wvf92\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.376506 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42ztn\" (UniqueName: \"kubernetes.io/projected/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-kube-api-access-42ztn\") pod \"dnsmasq-dns-5f6d79597f-wvf92\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.411029 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.532434 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d19854e-9f51-4d60-a6c1-6c1487f31d92" path="/var/lib/kubelet/pods/2d19854e-9f51-4d60-a6c1-6c1487f31d92/volumes" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.532903 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3383b93f-a184-4b6e-ad59-313e19cf4371" path="/var/lib/kubelet/pods/3383b93f-a184-4b6e-ad59-313e19cf4371/volumes" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.672162 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd7c66845-thth9" event={"ID":"dbc8bb26-f165-4960-becf-661fc9b67eed","Type":"ContainerStarted","Data":"f6b900fc5fdfeafa793d43e0f7dcb02177c3b8a55e7a6d80a6ab7d620fb80c0c"} Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.672599 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.677391 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" event={"ID":"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9","Type":"ContainerStarted","Data":"913cd11463e9569b76011529c0c9a2c497c0ebef75e361ac42746df7eb821abb"} Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.710420 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bd7c66845-thth9" podStartSLOduration=2.848014526 podStartE2EDuration="3.710402242s" podCreationTimestamp="2025-12-01 14:54:25 +0000 UTC" firstStartedPulling="2025-12-01 14:54:26.39278924 +0000 UTC m=+1232.156298843" lastFinishedPulling="2025-12-01 14:54:27.255176956 +0000 UTC m=+1233.018686559" observedRunningTime="2025-12-01 14:54:28.69781071 +0000 UTC m=+1234.461320313" watchObservedRunningTime="2025-12-01 14:54:28.710402242 +0000 UTC m=+1234.473911845" Dec 01 14:54:28 crc kubenswrapper[4810]: W1201 14:54:28.722299 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod604a8b79_fcf3_40cf_82e6_8af2cf667b0f.slice/crio-1610573c62e58cd66296d7120033026e8ac8365ced8ff8d8bc935cffdb387816 WatchSource:0}: Error finding container 1610573c62e58cd66296d7120033026e8ac8365ced8ff8d8bc935cffdb387816: Status 404 returned error can't find the container with id 1610573c62e58cd66296d7120033026e8ac8365ced8ff8d8bc935cffdb387816 Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.726357 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.730274 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" podStartSLOduration=3.235001172 podStartE2EDuration="3.730259351s" podCreationTimestamp="2025-12-01 14:54:25 +0000 UTC" firstStartedPulling="2025-12-01 14:54:26.235109436 +0000 UTC m=+1231.998619039" lastFinishedPulling="2025-12-01 14:54:26.730367615 +0000 UTC m=+1232.493877218" observedRunningTime="2025-12-01 14:54:28.717598627 +0000 UTC m=+1234.481108250" watchObservedRunningTime="2025-12-01 14:54:28.730259351 +0000 UTC m=+1234.493768954" Dec 01 14:54:28 crc kubenswrapper[4810]: I1201 14:54:28.912848 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f6d79597f-wvf92"] Dec 01 14:54:28 crc kubenswrapper[4810]: W1201 14:54:28.914244 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1aec0ffb_75a0_43e6_a721_4010c6ee0f81.slice/crio-1689fa3a24b341f5c3eca87491a11d142e9373c69b475f908affd61a79a767b6 WatchSource:0}: Error finding container 1689fa3a24b341f5c3eca87491a11d142e9373c69b475f908affd61a79a767b6: Status 404 returned error can't find the container with id 1689fa3a24b341f5c3eca87491a11d142e9373c69b475f908affd61a79a767b6 Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.071101 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.145502 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.153216 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.159133 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.160861 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.161252 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.161706 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-s9lmr" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.161760 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.178238 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.271154 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mq9v\" (UniqueName: \"kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-kube-api-access-5mq9v\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.271205 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.271239 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.271262 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1f455986-23c9-4088-9829-51c74e5b7770-cache\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.271317 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1f455986-23c9-4088-9829-51c74e5b7770-lock\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.373276 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mq9v\" (UniqueName: \"kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-kube-api-access-5mq9v\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.373340 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.373385 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.373416 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1f455986-23c9-4088-9829-51c74e5b7770-cache\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.373498 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1f455986-23c9-4088-9829-51c74e5b7770-lock\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.373847 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/swift-storage-0" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.374190 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1f455986-23c9-4088-9829-51c74e5b7770-cache\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:29 crc kubenswrapper[4810]: E1201 14:54:29.374298 4810 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 14:54:29 crc kubenswrapper[4810]: E1201 14:54:29.374315 4810 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 14:54:29 crc kubenswrapper[4810]: E1201 14:54:29.374369 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift podName:1f455986-23c9-4088-9829-51c74e5b7770 nodeName:}" failed. No retries permitted until 2025-12-01 14:54:29.874353785 +0000 UTC m=+1235.637863388 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift") pod "swift-storage-0" (UID: "1f455986-23c9-4088-9829-51c74e5b7770") : configmap "swift-ring-files" not found Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.374493 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1f455986-23c9-4088-9829-51c74e5b7770-lock\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.391170 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mq9v\" (UniqueName: \"kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-kube-api-access-5mq9v\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.394177 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.684196 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"604a8b79-fcf3-40cf-82e6-8af2cf667b0f","Type":"ContainerStarted","Data":"1610573c62e58cd66296d7120033026e8ac8365ced8ff8d8bc935cffdb387816"} Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.685693 4810 generic.go:334] "Generic (PLEG): container finished" podID="1aec0ffb-75a0-43e6-a721-4010c6ee0f81" containerID="47157a1db99702e1e8ffb9402d3bdb1c9952d455b59c92707ba64c98952bb76c" exitCode=0 Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.685833 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" event={"ID":"1aec0ffb-75a0-43e6-a721-4010c6ee0f81","Type":"ContainerDied","Data":"47157a1db99702e1e8ffb9402d3bdb1c9952d455b59c92707ba64c98952bb76c"} Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.685881 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" event={"ID":"1aec0ffb-75a0-43e6-a721-4010c6ee0f81","Type":"ContainerStarted","Data":"1689fa3a24b341f5c3eca87491a11d142e9373c69b475f908affd61a79a767b6"} Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.688596 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.688631 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bd7c66845-thth9" podUID="dbc8bb26-f165-4960-becf-661fc9b67eed" containerName="dnsmasq-dns" containerID="cri-o://f6b900fc5fdfeafa793d43e0f7dcb02177c3b8a55e7a6d80a6ab7d620fb80c0c" gracePeriod=10 Dec 01 14:54:29 crc kubenswrapper[4810]: I1201 14:54:29.880746 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:29 crc kubenswrapper[4810]: E1201 14:54:29.880945 4810 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 14:54:29 crc kubenswrapper[4810]: E1201 14:54:29.880975 4810 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 14:54:29 crc kubenswrapper[4810]: E1201 14:54:29.881032 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift podName:1f455986-23c9-4088-9829-51c74e5b7770 nodeName:}" failed. No retries permitted until 2025-12-01 14:54:30.881014013 +0000 UTC m=+1236.644523616 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift") pod "swift-storage-0" (UID: "1f455986-23c9-4088-9829-51c74e5b7770") : configmap "swift-ring-files" not found Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.206092 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.288451 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-ovsdbserver-nb\") pod \"dbc8bb26-f165-4960-becf-661fc9b67eed\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.288520 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5vkf\" (UniqueName: \"kubernetes.io/projected/dbc8bb26-f165-4960-becf-661fc9b67eed-kube-api-access-g5vkf\") pod \"dbc8bb26-f165-4960-becf-661fc9b67eed\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.288703 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-config\") pod \"dbc8bb26-f165-4960-becf-661fc9b67eed\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.288733 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-dns-svc\") pod \"dbc8bb26-f165-4960-becf-661fc9b67eed\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.288771 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-ovsdbserver-sb\") pod \"dbc8bb26-f165-4960-becf-661fc9b67eed\" (UID: \"dbc8bb26-f165-4960-becf-661fc9b67eed\") " Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.296073 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbc8bb26-f165-4960-becf-661fc9b67eed-kube-api-access-g5vkf" (OuterVolumeSpecName: "kube-api-access-g5vkf") pod "dbc8bb26-f165-4960-becf-661fc9b67eed" (UID: "dbc8bb26-f165-4960-becf-661fc9b67eed"). InnerVolumeSpecName "kube-api-access-g5vkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.337044 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-config" (OuterVolumeSpecName: "config") pod "dbc8bb26-f165-4960-becf-661fc9b67eed" (UID: "dbc8bb26-f165-4960-becf-661fc9b67eed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.337938 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dbc8bb26-f165-4960-becf-661fc9b67eed" (UID: "dbc8bb26-f165-4960-becf-661fc9b67eed"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.339133 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dbc8bb26-f165-4960-becf-661fc9b67eed" (UID: "dbc8bb26-f165-4960-becf-661fc9b67eed"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.356688 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dbc8bb26-f165-4960-becf-661fc9b67eed" (UID: "dbc8bb26-f165-4960-becf-661fc9b67eed"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.391215 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.391286 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.391296 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.391309 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dbc8bb26-f165-4960-becf-661fc9b67eed-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.391318 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5vkf\" (UniqueName: \"kubernetes.io/projected/dbc8bb26-f165-4960-becf-661fc9b67eed-kube-api-access-g5vkf\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.693636 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"604a8b79-fcf3-40cf-82e6-8af2cf667b0f","Type":"ContainerStarted","Data":"87439695570ef9b3e9da98681ae2d1dec7fb73d967ccb8912a2e40ca13529cf7"} Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.697894 4810 generic.go:334] "Generic (PLEG): container finished" podID="dbc8bb26-f165-4960-becf-661fc9b67eed" containerID="f6b900fc5fdfeafa793d43e0f7dcb02177c3b8a55e7a6d80a6ab7d620fb80c0c" exitCode=0 Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.697972 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd7c66845-thth9" event={"ID":"dbc8bb26-f165-4960-becf-661fc9b67eed","Type":"ContainerDied","Data":"f6b900fc5fdfeafa793d43e0f7dcb02177c3b8a55e7a6d80a6ab7d620fb80c0c"} Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.698005 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd7c66845-thth9" event={"ID":"dbc8bb26-f165-4960-becf-661fc9b67eed","Type":"ContainerDied","Data":"c8ce3082ea49e1b755c58768d704ccd938686b5077e56bc186480851cff88c95"} Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.698028 4810 scope.go:117] "RemoveContainer" containerID="f6b900fc5fdfeafa793d43e0f7dcb02177c3b8a55e7a6d80a6ab7d620fb80c0c" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.698188 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd7c66845-thth9" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.712335 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" event={"ID":"1aec0ffb-75a0-43e6-a721-4010c6ee0f81","Type":"ContainerStarted","Data":"d0da55680ea404f4e37e5d9ef3fc938bbec4754bf125f24758a733873b008d4e"} Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.712790 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.731557 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bd7c66845-thth9"] Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.734008 4810 scope.go:117] "RemoveContainer" containerID="b417098a11c1d128d79133f905054b971c187e966f0cc4420591b5298add921d" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.739419 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bd7c66845-thth9"] Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.745748 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" podStartSLOduration=2.745729352 podStartE2EDuration="2.745729352s" podCreationTimestamp="2025-12-01 14:54:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:54:30.742794822 +0000 UTC m=+1236.506304435" watchObservedRunningTime="2025-12-01 14:54:30.745729352 +0000 UTC m=+1236.509238945" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.759050 4810 scope.go:117] "RemoveContainer" containerID="f6b900fc5fdfeafa793d43e0f7dcb02177c3b8a55e7a6d80a6ab7d620fb80c0c" Dec 01 14:54:30 crc kubenswrapper[4810]: E1201 14:54:30.759529 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6b900fc5fdfeafa793d43e0f7dcb02177c3b8a55e7a6d80a6ab7d620fb80c0c\": container with ID starting with f6b900fc5fdfeafa793d43e0f7dcb02177c3b8a55e7a6d80a6ab7d620fb80c0c not found: ID does not exist" containerID="f6b900fc5fdfeafa793d43e0f7dcb02177c3b8a55e7a6d80a6ab7d620fb80c0c" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.759579 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6b900fc5fdfeafa793d43e0f7dcb02177c3b8a55e7a6d80a6ab7d620fb80c0c"} err="failed to get container status \"f6b900fc5fdfeafa793d43e0f7dcb02177c3b8a55e7a6d80a6ab7d620fb80c0c\": rpc error: code = NotFound desc = could not find container \"f6b900fc5fdfeafa793d43e0f7dcb02177c3b8a55e7a6d80a6ab7d620fb80c0c\": container with ID starting with f6b900fc5fdfeafa793d43e0f7dcb02177c3b8a55e7a6d80a6ab7d620fb80c0c not found: ID does not exist" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.759610 4810 scope.go:117] "RemoveContainer" containerID="b417098a11c1d128d79133f905054b971c187e966f0cc4420591b5298add921d" Dec 01 14:54:30 crc kubenswrapper[4810]: E1201 14:54:30.760412 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b417098a11c1d128d79133f905054b971c187e966f0cc4420591b5298add921d\": container with ID starting with b417098a11c1d128d79133f905054b971c187e966f0cc4420591b5298add921d not found: ID does not exist" containerID="b417098a11c1d128d79133f905054b971c187e966f0cc4420591b5298add921d" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.760451 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b417098a11c1d128d79133f905054b971c187e966f0cc4420591b5298add921d"} err="failed to get container status \"b417098a11c1d128d79133f905054b971c187e966f0cc4420591b5298add921d\": rpc error: code = NotFound desc = could not find container \"b417098a11c1d128d79133f905054b971c187e966f0cc4420591b5298add921d\": container with ID starting with b417098a11c1d128d79133f905054b971c187e966f0cc4420591b5298add921d not found: ID does not exist" Dec 01 14:54:30 crc kubenswrapper[4810]: I1201 14:54:30.898276 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:30 crc kubenswrapper[4810]: E1201 14:54:30.898512 4810 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 14:54:30 crc kubenswrapper[4810]: E1201 14:54:30.898741 4810 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 14:54:30 crc kubenswrapper[4810]: E1201 14:54:30.898790 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift podName:1f455986-23c9-4088-9829-51c74e5b7770 nodeName:}" failed. No retries permitted until 2025-12-01 14:54:32.898775181 +0000 UTC m=+1238.662284784 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift") pod "swift-storage-0" (UID: "1f455986-23c9-4088-9829-51c74e5b7770") : configmap "swift-ring-files" not found Dec 01 14:54:31 crc kubenswrapper[4810]: I1201 14:54:31.723084 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"604a8b79-fcf3-40cf-82e6-8af2cf667b0f","Type":"ContainerStarted","Data":"4403ed72fa83e2083ad29edab75aeefc505950c9b561718246753461398f9c33"} Dec 01 14:54:32 crc kubenswrapper[4810]: I1201 14:54:32.503748 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbc8bb26-f165-4960-becf-661fc9b67eed" path="/var/lib/kubelet/pods/dbc8bb26-f165-4960-becf-661fc9b67eed/volumes" Dec 01 14:54:32 crc kubenswrapper[4810]: I1201 14:54:32.732233 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 01 14:54:32 crc kubenswrapper[4810]: I1201 14:54:32.929877 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:32 crc kubenswrapper[4810]: E1201 14:54:32.930171 4810 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 14:54:32 crc kubenswrapper[4810]: E1201 14:54:32.930435 4810 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 14:54:32 crc kubenswrapper[4810]: E1201 14:54:32.930510 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift podName:1f455986-23c9-4088-9829-51c74e5b7770 nodeName:}" failed. No retries permitted until 2025-12-01 14:54:36.930491563 +0000 UTC m=+1242.694001166 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift") pod "swift-storage-0" (UID: "1f455986-23c9-4088-9829-51c74e5b7770") : configmap "swift-ring-files" not found Dec 01 14:54:32 crc kubenswrapper[4810]: I1201 14:54:32.972373 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:54:32 crc kubenswrapper[4810]: I1201 14:54:32.972448 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.123876 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=4.6046425840000005 podStartE2EDuration="6.123854888s" podCreationTimestamp="2025-12-01 14:54:27 +0000 UTC" firstStartedPulling="2025-12-01 14:54:28.725316677 +0000 UTC m=+1234.488826270" lastFinishedPulling="2025-12-01 14:54:30.244528961 +0000 UTC m=+1236.008038574" observedRunningTime="2025-12-01 14:54:31.756022117 +0000 UTC m=+1237.519531720" watchObservedRunningTime="2025-12-01 14:54:33.123854888 +0000 UTC m=+1238.887364491" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.130705 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-mslgc"] Dec 01 14:54:33 crc kubenswrapper[4810]: E1201 14:54:33.131081 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbc8bb26-f165-4960-becf-661fc9b67eed" containerName="dnsmasq-dns" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.131104 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbc8bb26-f165-4960-becf-661fc9b67eed" containerName="dnsmasq-dns" Dec 01 14:54:33 crc kubenswrapper[4810]: E1201 14:54:33.131141 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbc8bb26-f165-4960-becf-661fc9b67eed" containerName="init" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.131149 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbc8bb26-f165-4960-becf-661fc9b67eed" containerName="init" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.131319 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbc8bb26-f165-4960-becf-661fc9b67eed" containerName="dnsmasq-dns" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.131984 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.133945 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.134447 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.135144 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.152591 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-mslgc"] Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.235771 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw6hp\" (UniqueName: \"kubernetes.io/projected/beec7c3d-1e8b-42c7-9e07-978388973b80-kube-api-access-dw6hp\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.235837 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/beec7c3d-1e8b-42c7-9e07-978388973b80-etc-swift\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.235867 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/beec7c3d-1e8b-42c7-9e07-978388973b80-scripts\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.235927 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/beec7c3d-1e8b-42c7-9e07-978388973b80-combined-ca-bundle\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.235962 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/beec7c3d-1e8b-42c7-9e07-978388973b80-ring-data-devices\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.236024 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/beec7c3d-1e8b-42c7-9e07-978388973b80-dispersionconf\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.236046 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/beec7c3d-1e8b-42c7-9e07-978388973b80-swiftconf\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.337278 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/beec7c3d-1e8b-42c7-9e07-978388973b80-dispersionconf\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.337592 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/beec7c3d-1e8b-42c7-9e07-978388973b80-swiftconf\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.338391 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw6hp\" (UniqueName: \"kubernetes.io/projected/beec7c3d-1e8b-42c7-9e07-978388973b80-kube-api-access-dw6hp\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.338465 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/beec7c3d-1e8b-42c7-9e07-978388973b80-etc-swift\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.338520 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/beec7c3d-1e8b-42c7-9e07-978388973b80-scripts\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.338673 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/beec7c3d-1e8b-42c7-9e07-978388973b80-combined-ca-bundle\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.338737 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/beec7c3d-1e8b-42c7-9e07-978388973b80-ring-data-devices\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.339143 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/beec7c3d-1e8b-42c7-9e07-978388973b80-etc-swift\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.339525 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/beec7c3d-1e8b-42c7-9e07-978388973b80-ring-data-devices\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.339688 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/beec7c3d-1e8b-42c7-9e07-978388973b80-scripts\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.342520 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/beec7c3d-1e8b-42c7-9e07-978388973b80-swiftconf\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.346111 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/beec7c3d-1e8b-42c7-9e07-978388973b80-combined-ca-bundle\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.346379 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/beec7c3d-1e8b-42c7-9e07-978388973b80-dispersionconf\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.354309 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw6hp\" (UniqueName: \"kubernetes.io/projected/beec7c3d-1e8b-42c7-9e07-978388973b80-kube-api-access-dw6hp\") pod \"swift-ring-rebalance-mslgc\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.453595 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:33 crc kubenswrapper[4810]: W1201 14:54:33.924872 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbeec7c3d_1e8b_42c7_9e07_978388973b80.slice/crio-3bce6aa8345ea924a8105d9facbd40eb8d6b05c3de6b2094a38304f8822fbf05 WatchSource:0}: Error finding container 3bce6aa8345ea924a8105d9facbd40eb8d6b05c3de6b2094a38304f8822fbf05: Status 404 returned error can't find the container with id 3bce6aa8345ea924a8105d9facbd40eb8d6b05c3de6b2094a38304f8822fbf05 Dec 01 14:54:33 crc kubenswrapper[4810]: I1201 14:54:33.927939 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-mslgc"] Dec 01 14:54:34 crc kubenswrapper[4810]: I1201 14:54:34.749660 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mslgc" event={"ID":"beec7c3d-1e8b-42c7-9e07-978388973b80","Type":"ContainerStarted","Data":"3bce6aa8345ea924a8105d9facbd40eb8d6b05c3de6b2094a38304f8822fbf05"} Dec 01 14:54:35 crc kubenswrapper[4810]: I1201 14:54:35.647833 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.225383 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8cd1-account-create-update-xj8tj"] Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.228956 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8cd1-account-create-update-xj8tj" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.232252 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.237420 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8cd1-account-create-update-xj8tj"] Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.278961 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-gcsr9"] Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.279978 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gcsr9" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.288707 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gcsr9"] Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.318070 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzqdm\" (UniqueName: \"kubernetes.io/projected/d3b0f7e7-9035-4c33-940b-8bdd905d97b5-kube-api-access-lzqdm\") pod \"keystone-8cd1-account-create-update-xj8tj\" (UID: \"d3b0f7e7-9035-4c33-940b-8bdd905d97b5\") " pod="openstack/keystone-8cd1-account-create-update-xj8tj" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.318649 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3b0f7e7-9035-4c33-940b-8bdd905d97b5-operator-scripts\") pod \"keystone-8cd1-account-create-update-xj8tj\" (UID: \"d3b0f7e7-9035-4c33-940b-8bdd905d97b5\") " pod="openstack/keystone-8cd1-account-create-update-xj8tj" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.420010 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3b0f7e7-9035-4c33-940b-8bdd905d97b5-operator-scripts\") pod \"keystone-8cd1-account-create-update-xj8tj\" (UID: \"d3b0f7e7-9035-4c33-940b-8bdd905d97b5\") " pod="openstack/keystone-8cd1-account-create-update-xj8tj" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.420106 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55732ab9-9610-4b4f-bae3-516b623e0608-operator-scripts\") pod \"keystone-db-create-gcsr9\" (UID: \"55732ab9-9610-4b4f-bae3-516b623e0608\") " pod="openstack/keystone-db-create-gcsr9" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.420131 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzqdm\" (UniqueName: \"kubernetes.io/projected/d3b0f7e7-9035-4c33-940b-8bdd905d97b5-kube-api-access-lzqdm\") pod \"keystone-8cd1-account-create-update-xj8tj\" (UID: \"d3b0f7e7-9035-4c33-940b-8bdd905d97b5\") " pod="openstack/keystone-8cd1-account-create-update-xj8tj" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.420173 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kcqs\" (UniqueName: \"kubernetes.io/projected/55732ab9-9610-4b4f-bae3-516b623e0608-kube-api-access-8kcqs\") pod \"keystone-db-create-gcsr9\" (UID: \"55732ab9-9610-4b4f-bae3-516b623e0608\") " pod="openstack/keystone-db-create-gcsr9" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.421836 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3b0f7e7-9035-4c33-940b-8bdd905d97b5-operator-scripts\") pod \"keystone-8cd1-account-create-update-xj8tj\" (UID: \"d3b0f7e7-9035-4c33-940b-8bdd905d97b5\") " pod="openstack/keystone-8cd1-account-create-update-xj8tj" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.456800 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-prlzg"] Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.458510 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-prlzg" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.465278 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-prlzg"] Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.466093 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzqdm\" (UniqueName: \"kubernetes.io/projected/d3b0f7e7-9035-4c33-940b-8bdd905d97b5-kube-api-access-lzqdm\") pod \"keystone-8cd1-account-create-update-xj8tj\" (UID: \"d3b0f7e7-9035-4c33-940b-8bdd905d97b5\") " pod="openstack/keystone-8cd1-account-create-update-xj8tj" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.521118 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30a1b114-0b4e-43ff-815b-0983ca7a86e6-operator-scripts\") pod \"placement-db-create-prlzg\" (UID: \"30a1b114-0b4e-43ff-815b-0983ca7a86e6\") " pod="openstack/placement-db-create-prlzg" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.521421 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55732ab9-9610-4b4f-bae3-516b623e0608-operator-scripts\") pod \"keystone-db-create-gcsr9\" (UID: \"55732ab9-9610-4b4f-bae3-516b623e0608\") " pod="openstack/keystone-db-create-gcsr9" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.521521 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z96p6\" (UniqueName: \"kubernetes.io/projected/30a1b114-0b4e-43ff-815b-0983ca7a86e6-kube-api-access-z96p6\") pod \"placement-db-create-prlzg\" (UID: \"30a1b114-0b4e-43ff-815b-0983ca7a86e6\") " pod="openstack/placement-db-create-prlzg" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.521597 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kcqs\" (UniqueName: \"kubernetes.io/projected/55732ab9-9610-4b4f-bae3-516b623e0608-kube-api-access-8kcqs\") pod \"keystone-db-create-gcsr9\" (UID: \"55732ab9-9610-4b4f-bae3-516b623e0608\") " pod="openstack/keystone-db-create-gcsr9" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.522115 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55732ab9-9610-4b4f-bae3-516b623e0608-operator-scripts\") pod \"keystone-db-create-gcsr9\" (UID: \"55732ab9-9610-4b4f-bae3-516b623e0608\") " pod="openstack/keystone-db-create-gcsr9" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.545082 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kcqs\" (UniqueName: \"kubernetes.io/projected/55732ab9-9610-4b4f-bae3-516b623e0608-kube-api-access-8kcqs\") pod \"keystone-db-create-gcsr9\" (UID: \"55732ab9-9610-4b4f-bae3-516b623e0608\") " pod="openstack/keystone-db-create-gcsr9" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.574006 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8cd1-account-create-update-xj8tj" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.582945 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-8657-account-create-update-jsmmz"] Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.584002 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8657-account-create-update-jsmmz" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.586116 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.591856 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8657-account-create-update-jsmmz"] Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.599175 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gcsr9" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.623519 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z96p6\" (UniqueName: \"kubernetes.io/projected/30a1b114-0b4e-43ff-815b-0983ca7a86e6-kube-api-access-z96p6\") pod \"placement-db-create-prlzg\" (UID: \"30a1b114-0b4e-43ff-815b-0983ca7a86e6\") " pod="openstack/placement-db-create-prlzg" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.623907 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30a1b114-0b4e-43ff-815b-0983ca7a86e6-operator-scripts\") pod \"placement-db-create-prlzg\" (UID: \"30a1b114-0b4e-43ff-815b-0983ca7a86e6\") " pod="openstack/placement-db-create-prlzg" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.624815 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30a1b114-0b4e-43ff-815b-0983ca7a86e6-operator-scripts\") pod \"placement-db-create-prlzg\" (UID: \"30a1b114-0b4e-43ff-815b-0983ca7a86e6\") " pod="openstack/placement-db-create-prlzg" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.640099 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z96p6\" (UniqueName: \"kubernetes.io/projected/30a1b114-0b4e-43ff-815b-0983ca7a86e6-kube-api-access-z96p6\") pod \"placement-db-create-prlzg\" (UID: \"30a1b114-0b4e-43ff-815b-0983ca7a86e6\") " pod="openstack/placement-db-create-prlzg" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.725669 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e29ef9a3-9060-4301-ad2c-3d62fe66c6b9-operator-scripts\") pod \"placement-8657-account-create-update-jsmmz\" (UID: \"e29ef9a3-9060-4301-ad2c-3d62fe66c6b9\") " pod="openstack/placement-8657-account-create-update-jsmmz" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.725756 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mkjn\" (UniqueName: \"kubernetes.io/projected/e29ef9a3-9060-4301-ad2c-3d62fe66c6b9-kube-api-access-9mkjn\") pod \"placement-8657-account-create-update-jsmmz\" (UID: \"e29ef9a3-9060-4301-ad2c-3d62fe66c6b9\") " pod="openstack/placement-8657-account-create-update-jsmmz" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.741060 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-n85zz"] Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.742315 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-n85zz" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.754558 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-n85zz"] Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.811680 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-prlzg" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.830188 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6a42d09-fb12-4466-a792-49998f29ccc4-operator-scripts\") pod \"glance-db-create-n85zz\" (UID: \"d6a42d09-fb12-4466-a792-49998f29ccc4\") " pod="openstack/glance-db-create-n85zz" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.830380 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e29ef9a3-9060-4301-ad2c-3d62fe66c6b9-operator-scripts\") pod \"placement-8657-account-create-update-jsmmz\" (UID: \"e29ef9a3-9060-4301-ad2c-3d62fe66c6b9\") " pod="openstack/placement-8657-account-create-update-jsmmz" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.830593 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mkjn\" (UniqueName: \"kubernetes.io/projected/e29ef9a3-9060-4301-ad2c-3d62fe66c6b9-kube-api-access-9mkjn\") pod \"placement-8657-account-create-update-jsmmz\" (UID: \"e29ef9a3-9060-4301-ad2c-3d62fe66c6b9\") " pod="openstack/placement-8657-account-create-update-jsmmz" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.830710 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6mzs\" (UniqueName: \"kubernetes.io/projected/d6a42d09-fb12-4466-a792-49998f29ccc4-kube-api-access-l6mzs\") pod \"glance-db-create-n85zz\" (UID: \"d6a42d09-fb12-4466-a792-49998f29ccc4\") " pod="openstack/glance-db-create-n85zz" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.831661 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e29ef9a3-9060-4301-ad2c-3d62fe66c6b9-operator-scripts\") pod \"placement-8657-account-create-update-jsmmz\" (UID: \"e29ef9a3-9060-4301-ad2c-3d62fe66c6b9\") " pod="openstack/placement-8657-account-create-update-jsmmz" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.853125 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mkjn\" (UniqueName: \"kubernetes.io/projected/e29ef9a3-9060-4301-ad2c-3d62fe66c6b9-kube-api-access-9mkjn\") pod \"placement-8657-account-create-update-jsmmz\" (UID: \"e29ef9a3-9060-4301-ad2c-3d62fe66c6b9\") " pod="openstack/placement-8657-account-create-update-jsmmz" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.884686 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-324e-account-create-update-rdb5j"] Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.885792 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-324e-account-create-update-rdb5j" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.888181 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.903265 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-324e-account-create-update-rdb5j"] Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.906946 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8657-account-create-update-jsmmz" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.932782 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6mzs\" (UniqueName: \"kubernetes.io/projected/d6a42d09-fb12-4466-a792-49998f29ccc4-kube-api-access-l6mzs\") pod \"glance-db-create-n85zz\" (UID: \"d6a42d09-fb12-4466-a792-49998f29ccc4\") " pod="openstack/glance-db-create-n85zz" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.932831 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.932919 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6a42d09-fb12-4466-a792-49998f29ccc4-operator-scripts\") pod \"glance-db-create-n85zz\" (UID: \"d6a42d09-fb12-4466-a792-49998f29ccc4\") " pod="openstack/glance-db-create-n85zz" Dec 01 14:54:36 crc kubenswrapper[4810]: E1201 14:54:36.933221 4810 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 14:54:36 crc kubenswrapper[4810]: E1201 14:54:36.933301 4810 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 14:54:36 crc kubenswrapper[4810]: E1201 14:54:36.933396 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift podName:1f455986-23c9-4088-9829-51c74e5b7770 nodeName:}" failed. No retries permitted until 2025-12-01 14:54:44.933364101 +0000 UTC m=+1250.696873694 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift") pod "swift-storage-0" (UID: "1f455986-23c9-4088-9829-51c74e5b7770") : configmap "swift-ring-files" not found Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.933572 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6a42d09-fb12-4466-a792-49998f29ccc4-operator-scripts\") pod \"glance-db-create-n85zz\" (UID: \"d6a42d09-fb12-4466-a792-49998f29ccc4\") " pod="openstack/glance-db-create-n85zz" Dec 01 14:54:36 crc kubenswrapper[4810]: I1201 14:54:36.949267 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6mzs\" (UniqueName: \"kubernetes.io/projected/d6a42d09-fb12-4466-a792-49998f29ccc4-kube-api-access-l6mzs\") pod \"glance-db-create-n85zz\" (UID: \"d6a42d09-fb12-4466-a792-49998f29ccc4\") " pod="openstack/glance-db-create-n85zz" Dec 01 14:54:37 crc kubenswrapper[4810]: I1201 14:54:37.034880 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fb6a553-c5fd-427f-96e6-b5bd6d3a4971-operator-scripts\") pod \"glance-324e-account-create-update-rdb5j\" (UID: \"8fb6a553-c5fd-427f-96e6-b5bd6d3a4971\") " pod="openstack/glance-324e-account-create-update-rdb5j" Dec 01 14:54:37 crc kubenswrapper[4810]: I1201 14:54:37.034944 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl9wl\" (UniqueName: \"kubernetes.io/projected/8fb6a553-c5fd-427f-96e6-b5bd6d3a4971-kube-api-access-rl9wl\") pod \"glance-324e-account-create-update-rdb5j\" (UID: \"8fb6a553-c5fd-427f-96e6-b5bd6d3a4971\") " pod="openstack/glance-324e-account-create-update-rdb5j" Dec 01 14:54:37 crc kubenswrapper[4810]: I1201 14:54:37.064958 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-n85zz" Dec 01 14:54:37 crc kubenswrapper[4810]: I1201 14:54:37.136658 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fb6a553-c5fd-427f-96e6-b5bd6d3a4971-operator-scripts\") pod \"glance-324e-account-create-update-rdb5j\" (UID: \"8fb6a553-c5fd-427f-96e6-b5bd6d3a4971\") " pod="openstack/glance-324e-account-create-update-rdb5j" Dec 01 14:54:37 crc kubenswrapper[4810]: I1201 14:54:37.136719 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl9wl\" (UniqueName: \"kubernetes.io/projected/8fb6a553-c5fd-427f-96e6-b5bd6d3a4971-kube-api-access-rl9wl\") pod \"glance-324e-account-create-update-rdb5j\" (UID: \"8fb6a553-c5fd-427f-96e6-b5bd6d3a4971\") " pod="openstack/glance-324e-account-create-update-rdb5j" Dec 01 14:54:37 crc kubenswrapper[4810]: I1201 14:54:37.137503 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fb6a553-c5fd-427f-96e6-b5bd6d3a4971-operator-scripts\") pod \"glance-324e-account-create-update-rdb5j\" (UID: \"8fb6a553-c5fd-427f-96e6-b5bd6d3a4971\") " pod="openstack/glance-324e-account-create-update-rdb5j" Dec 01 14:54:37 crc kubenswrapper[4810]: I1201 14:54:37.155234 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl9wl\" (UniqueName: \"kubernetes.io/projected/8fb6a553-c5fd-427f-96e6-b5bd6d3a4971-kube-api-access-rl9wl\") pod \"glance-324e-account-create-update-rdb5j\" (UID: \"8fb6a553-c5fd-427f-96e6-b5bd6d3a4971\") " pod="openstack/glance-324e-account-create-update-rdb5j" Dec 01 14:54:37 crc kubenswrapper[4810]: I1201 14:54:37.205227 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-324e-account-create-update-rdb5j" Dec 01 14:54:38 crc kubenswrapper[4810]: I1201 14:54:38.413017 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:54:38 crc kubenswrapper[4810]: I1201 14:54:38.502946 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b4845d8d5-hfkzl"] Dec 01 14:54:38 crc kubenswrapper[4810]: I1201 14:54:38.503168 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" podUID="2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9" containerName="dnsmasq-dns" containerID="cri-o://913cd11463e9569b76011529c0c9a2c497c0ebef75e361ac42746df7eb821abb" gracePeriod=10 Dec 01 14:54:38 crc kubenswrapper[4810]: I1201 14:54:38.794498 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mslgc" event={"ID":"beec7c3d-1e8b-42c7-9e07-978388973b80","Type":"ContainerStarted","Data":"6f00749a221c1f7e6ab22d685c419d81ad12d39e73efcda6cf4bc60436e158e9"} Dec 01 14:54:38 crc kubenswrapper[4810]: I1201 14:54:38.805588 4810 generic.go:334] "Generic (PLEG): container finished" podID="2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9" containerID="913cd11463e9569b76011529c0c9a2c497c0ebef75e361ac42746df7eb821abb" exitCode=0 Dec 01 14:54:38 crc kubenswrapper[4810]: I1201 14:54:38.805710 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" event={"ID":"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9","Type":"ContainerDied","Data":"913cd11463e9569b76011529c0c9a2c497c0ebef75e361ac42746df7eb821abb"} Dec 01 14:54:38 crc kubenswrapper[4810]: I1201 14:54:38.830350 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-mslgc" podStartSLOduration=1.6232542639999998 podStartE2EDuration="5.830326761s" podCreationTimestamp="2025-12-01 14:54:33 +0000 UTC" firstStartedPulling="2025-12-01 14:54:33.927885457 +0000 UTC m=+1239.691395060" lastFinishedPulling="2025-12-01 14:54:38.134957954 +0000 UTC m=+1243.898467557" observedRunningTime="2025-12-01 14:54:38.815670803 +0000 UTC m=+1244.579180416" watchObservedRunningTime="2025-12-01 14:54:38.830326761 +0000 UTC m=+1244.593836364" Dec 01 14:54:38 crc kubenswrapper[4810]: I1201 14:54:38.862857 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gcsr9"] Dec 01 14:54:38 crc kubenswrapper[4810]: I1201 14:54:38.870127 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-324e-account-create-update-rdb5j"] Dec 01 14:54:38 crc kubenswrapper[4810]: I1201 14:54:38.877413 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8cd1-account-create-update-xj8tj"] Dec 01 14:54:38 crc kubenswrapper[4810]: I1201 14:54:38.997142 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8657-account-create-update-jsmmz"] Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.005355 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-n85zz"] Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.011946 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-prlzg"] Dec 01 14:54:39 crc kubenswrapper[4810]: W1201 14:54:39.035521 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30a1b114_0b4e_43ff_815b_0983ca7a86e6.slice/crio-3a3228122918ddc595705de79d5698e91b10af2c1731458e5afc4391fa25a498 WatchSource:0}: Error finding container 3a3228122918ddc595705de79d5698e91b10af2c1731458e5afc4391fa25a498: Status 404 returned error can't find the container with id 3a3228122918ddc595705de79d5698e91b10af2c1731458e5afc4391fa25a498 Dec 01 14:54:39 crc kubenswrapper[4810]: W1201 14:54:39.036082 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode29ef9a3_9060_4301_ad2c_3d62fe66c6b9.slice/crio-579174cd63e33fe2cfc74401d12dba42d7aacbab154a62e4a73338a9c3402d24 WatchSource:0}: Error finding container 579174cd63e33fe2cfc74401d12dba42d7aacbab154a62e4a73338a9c3402d24: Status 404 returned error can't find the container with id 579174cd63e33fe2cfc74401d12dba42d7aacbab154a62e4a73338a9c3402d24 Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.037073 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.192431 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6475\" (UniqueName: \"kubernetes.io/projected/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-kube-api-access-d6475\") pod \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\" (UID: \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\") " Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.192799 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-ovsdbserver-nb\") pod \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\" (UID: \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\") " Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.192925 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-dns-svc\") pod \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\" (UID: \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\") " Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.193027 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-config\") pod \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\" (UID: \"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9\") " Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.202673 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-kube-api-access-d6475" (OuterVolumeSpecName: "kube-api-access-d6475") pod "2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9" (UID: "2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9"). InnerVolumeSpecName "kube-api-access-d6475". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.296029 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6475\" (UniqueName: \"kubernetes.io/projected/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-kube-api-access-d6475\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.315910 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9" (UID: "2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.320160 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-config" (OuterVolumeSpecName: "config") pod "2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9" (UID: "2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.324349 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9" (UID: "2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.397414 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.397756 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.397848 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.813745 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" event={"ID":"2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9","Type":"ContainerDied","Data":"a3372657fecd36a885e3d0b8b019c28a58ca7686ff6984e079f04bb72ac1c0d2"} Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.813808 4810 scope.go:117] "RemoveContainer" containerID="913cd11463e9569b76011529c0c9a2c497c0ebef75e361ac42746df7eb821abb" Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.813754 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b4845d8d5-hfkzl" Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.816580 4810 generic.go:334] "Generic (PLEG): container finished" podID="8fb6a553-c5fd-427f-96e6-b5bd6d3a4971" containerID="9870c1b8a5276fda6d94cdc32f4bc57664dfbe1d025f8da26d8eb70045b38620" exitCode=0 Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.816644 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-324e-account-create-update-rdb5j" event={"ID":"8fb6a553-c5fd-427f-96e6-b5bd6d3a4971","Type":"ContainerDied","Data":"9870c1b8a5276fda6d94cdc32f4bc57664dfbe1d025f8da26d8eb70045b38620"} Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.816670 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-324e-account-create-update-rdb5j" event={"ID":"8fb6a553-c5fd-427f-96e6-b5bd6d3a4971","Type":"ContainerStarted","Data":"6453e00b8333d08833df4f469cd5005df0d4b9894a281253ffa62a8ea3909e6a"} Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.818944 4810 generic.go:334] "Generic (PLEG): container finished" podID="e29ef9a3-9060-4301-ad2c-3d62fe66c6b9" containerID="4b9246edd94b5a8cf6a041a9b8421b6220f01030cf4806b8da8546eafef406a4" exitCode=0 Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.819058 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8657-account-create-update-jsmmz" event={"ID":"e29ef9a3-9060-4301-ad2c-3d62fe66c6b9","Type":"ContainerDied","Data":"4b9246edd94b5a8cf6a041a9b8421b6220f01030cf4806b8da8546eafef406a4"} Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.819074 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8657-account-create-update-jsmmz" event={"ID":"e29ef9a3-9060-4301-ad2c-3d62fe66c6b9","Type":"ContainerStarted","Data":"579174cd63e33fe2cfc74401d12dba42d7aacbab154a62e4a73338a9c3402d24"} Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.821731 4810 generic.go:334] "Generic (PLEG): container finished" podID="d6a42d09-fb12-4466-a792-49998f29ccc4" containerID="8fbfb2930b964fb785ed1a3925eda27b84fe588e38fc85afb7e23742343427af" exitCode=0 Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.821788 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-n85zz" event={"ID":"d6a42d09-fb12-4466-a792-49998f29ccc4","Type":"ContainerDied","Data":"8fbfb2930b964fb785ed1a3925eda27b84fe588e38fc85afb7e23742343427af"} Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.821810 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-n85zz" event={"ID":"d6a42d09-fb12-4466-a792-49998f29ccc4","Type":"ContainerStarted","Data":"0d27fd2f6257b6ef69603863befc7f8eeb5546cc23d612d9aab8abd0b62c9ffa"} Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.823036 4810 generic.go:334] "Generic (PLEG): container finished" podID="d3b0f7e7-9035-4c33-940b-8bdd905d97b5" containerID="1d0d48e84f1be3cb4796f921b940c2af50fbf54d7bc7bc24d59d735872cb1f8c" exitCode=0 Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.823072 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8cd1-account-create-update-xj8tj" event={"ID":"d3b0f7e7-9035-4c33-940b-8bdd905d97b5","Type":"ContainerDied","Data":"1d0d48e84f1be3cb4796f921b940c2af50fbf54d7bc7bc24d59d735872cb1f8c"} Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.823085 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8cd1-account-create-update-xj8tj" event={"ID":"d3b0f7e7-9035-4c33-940b-8bdd905d97b5","Type":"ContainerStarted","Data":"18201aca890626fc74281693ad8dcda7a9e82c0ec305201d3f8ed6a5028b21c3"} Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.824966 4810 generic.go:334] "Generic (PLEG): container finished" podID="55732ab9-9610-4b4f-bae3-516b623e0608" containerID="7995ec5ad10e6bed11e113536d305833e1e514a42699e70c76cc0ce45be3bbc2" exitCode=0 Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.825067 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gcsr9" event={"ID":"55732ab9-9610-4b4f-bae3-516b623e0608","Type":"ContainerDied","Data":"7995ec5ad10e6bed11e113536d305833e1e514a42699e70c76cc0ce45be3bbc2"} Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.825087 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gcsr9" event={"ID":"55732ab9-9610-4b4f-bae3-516b623e0608","Type":"ContainerStarted","Data":"930c7e0a2cd5986f7b5a71f6263ad9e6ab47464bc738b97dd96428bab2d3eae3"} Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.828674 4810 generic.go:334] "Generic (PLEG): container finished" podID="30a1b114-0b4e-43ff-815b-0983ca7a86e6" containerID="20908c4e323538658a97e4515a4c9cfc48a3249e1229339681a663eb84fee0df" exitCode=0 Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.828749 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-prlzg" event={"ID":"30a1b114-0b4e-43ff-815b-0983ca7a86e6","Type":"ContainerDied","Data":"20908c4e323538658a97e4515a4c9cfc48a3249e1229339681a663eb84fee0df"} Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.828807 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-prlzg" event={"ID":"30a1b114-0b4e-43ff-815b-0983ca7a86e6","Type":"ContainerStarted","Data":"3a3228122918ddc595705de79d5698e91b10af2c1731458e5afc4391fa25a498"} Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.832094 4810 scope.go:117] "RemoveContainer" containerID="f153efb28861380862b3940a8f955f575482e7b5a15dab4443b07455b5d8c644" Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.884632 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b4845d8d5-hfkzl"] Dec 01 14:54:39 crc kubenswrapper[4810]: I1201 14:54:39.894156 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b4845d8d5-hfkzl"] Dec 01 14:54:40 crc kubenswrapper[4810]: I1201 14:54:40.500313 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9" path="/var/lib/kubelet/pods/2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9/volumes" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.281603 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-324e-account-create-update-rdb5j" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.358848 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rl9wl\" (UniqueName: \"kubernetes.io/projected/8fb6a553-c5fd-427f-96e6-b5bd6d3a4971-kube-api-access-rl9wl\") pod \"8fb6a553-c5fd-427f-96e6-b5bd6d3a4971\" (UID: \"8fb6a553-c5fd-427f-96e6-b5bd6d3a4971\") " Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.358901 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fb6a553-c5fd-427f-96e6-b5bd6d3a4971-operator-scripts\") pod \"8fb6a553-c5fd-427f-96e6-b5bd6d3a4971\" (UID: \"8fb6a553-c5fd-427f-96e6-b5bd6d3a4971\") " Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.362120 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fb6a553-c5fd-427f-96e6-b5bd6d3a4971-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8fb6a553-c5fd-427f-96e6-b5bd6d3a4971" (UID: "8fb6a553-c5fd-427f-96e6-b5bd6d3a4971"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.366276 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fb6a553-c5fd-427f-96e6-b5bd6d3a4971-kube-api-access-rl9wl" (OuterVolumeSpecName: "kube-api-access-rl9wl") pod "8fb6a553-c5fd-427f-96e6-b5bd6d3a4971" (UID: "8fb6a553-c5fd-427f-96e6-b5bd6d3a4971"). InnerVolumeSpecName "kube-api-access-rl9wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.475021 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rl9wl\" (UniqueName: \"kubernetes.io/projected/8fb6a553-c5fd-427f-96e6-b5bd6d3a4971-kube-api-access-rl9wl\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.475067 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fb6a553-c5fd-427f-96e6-b5bd6d3a4971-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.489053 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gcsr9" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.498321 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8657-account-create-update-jsmmz" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.517826 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-prlzg" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.524880 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8cd1-account-create-update-xj8tj" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.526274 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-n85zz" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.576368 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3b0f7e7-9035-4c33-940b-8bdd905d97b5-operator-scripts\") pod \"d3b0f7e7-9035-4c33-940b-8bdd905d97b5\" (UID: \"d3b0f7e7-9035-4c33-940b-8bdd905d97b5\") " Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.576447 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kcqs\" (UniqueName: \"kubernetes.io/projected/55732ab9-9610-4b4f-bae3-516b623e0608-kube-api-access-8kcqs\") pod \"55732ab9-9610-4b4f-bae3-516b623e0608\" (UID: \"55732ab9-9610-4b4f-bae3-516b623e0608\") " Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.576884 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3b0f7e7-9035-4c33-940b-8bdd905d97b5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d3b0f7e7-9035-4c33-940b-8bdd905d97b5" (UID: "d3b0f7e7-9035-4c33-940b-8bdd905d97b5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.577573 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3b0f7e7-9035-4c33-940b-8bdd905d97b5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.579651 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55732ab9-9610-4b4f-bae3-516b623e0608-kube-api-access-8kcqs" (OuterVolumeSpecName: "kube-api-access-8kcqs") pod "55732ab9-9610-4b4f-bae3-516b623e0608" (UID: "55732ab9-9610-4b4f-bae3-516b623e0608"). InnerVolumeSpecName "kube-api-access-8kcqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.679193 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6a42d09-fb12-4466-a792-49998f29ccc4-operator-scripts\") pod \"d6a42d09-fb12-4466-a792-49998f29ccc4\" (UID: \"d6a42d09-fb12-4466-a792-49998f29ccc4\") " Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.679263 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55732ab9-9610-4b4f-bae3-516b623e0608-operator-scripts\") pod \"55732ab9-9610-4b4f-bae3-516b623e0608\" (UID: \"55732ab9-9610-4b4f-bae3-516b623e0608\") " Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.679289 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6mzs\" (UniqueName: \"kubernetes.io/projected/d6a42d09-fb12-4466-a792-49998f29ccc4-kube-api-access-l6mzs\") pod \"d6a42d09-fb12-4466-a792-49998f29ccc4\" (UID: \"d6a42d09-fb12-4466-a792-49998f29ccc4\") " Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.679313 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e29ef9a3-9060-4301-ad2c-3d62fe66c6b9-operator-scripts\") pod \"e29ef9a3-9060-4301-ad2c-3d62fe66c6b9\" (UID: \"e29ef9a3-9060-4301-ad2c-3d62fe66c6b9\") " Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.679333 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzqdm\" (UniqueName: \"kubernetes.io/projected/d3b0f7e7-9035-4c33-940b-8bdd905d97b5-kube-api-access-lzqdm\") pod \"d3b0f7e7-9035-4c33-940b-8bdd905d97b5\" (UID: \"d3b0f7e7-9035-4c33-940b-8bdd905d97b5\") " Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.679409 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30a1b114-0b4e-43ff-815b-0983ca7a86e6-operator-scripts\") pod \"30a1b114-0b4e-43ff-815b-0983ca7a86e6\" (UID: \"30a1b114-0b4e-43ff-815b-0983ca7a86e6\") " Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.679438 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z96p6\" (UniqueName: \"kubernetes.io/projected/30a1b114-0b4e-43ff-815b-0983ca7a86e6-kube-api-access-z96p6\") pod \"30a1b114-0b4e-43ff-815b-0983ca7a86e6\" (UID: \"30a1b114-0b4e-43ff-815b-0983ca7a86e6\") " Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.679513 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mkjn\" (UniqueName: \"kubernetes.io/projected/e29ef9a3-9060-4301-ad2c-3d62fe66c6b9-kube-api-access-9mkjn\") pod \"e29ef9a3-9060-4301-ad2c-3d62fe66c6b9\" (UID: \"e29ef9a3-9060-4301-ad2c-3d62fe66c6b9\") " Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.679827 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kcqs\" (UniqueName: \"kubernetes.io/projected/55732ab9-9610-4b4f-bae3-516b623e0608-kube-api-access-8kcqs\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.680368 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e29ef9a3-9060-4301-ad2c-3d62fe66c6b9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e29ef9a3-9060-4301-ad2c-3d62fe66c6b9" (UID: "e29ef9a3-9060-4301-ad2c-3d62fe66c6b9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.680675 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55732ab9-9610-4b4f-bae3-516b623e0608-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "55732ab9-9610-4b4f-bae3-516b623e0608" (UID: "55732ab9-9610-4b4f-bae3-516b623e0608"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.680796 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6a42d09-fb12-4466-a792-49998f29ccc4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d6a42d09-fb12-4466-a792-49998f29ccc4" (UID: "d6a42d09-fb12-4466-a792-49998f29ccc4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.680985 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30a1b114-0b4e-43ff-815b-0983ca7a86e6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "30a1b114-0b4e-43ff-815b-0983ca7a86e6" (UID: "30a1b114-0b4e-43ff-815b-0983ca7a86e6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.683617 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3b0f7e7-9035-4c33-940b-8bdd905d97b5-kube-api-access-lzqdm" (OuterVolumeSpecName: "kube-api-access-lzqdm") pod "d3b0f7e7-9035-4c33-940b-8bdd905d97b5" (UID: "d3b0f7e7-9035-4c33-940b-8bdd905d97b5"). InnerVolumeSpecName "kube-api-access-lzqdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.683721 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6a42d09-fb12-4466-a792-49998f29ccc4-kube-api-access-l6mzs" (OuterVolumeSpecName: "kube-api-access-l6mzs") pod "d6a42d09-fb12-4466-a792-49998f29ccc4" (UID: "d6a42d09-fb12-4466-a792-49998f29ccc4"). InnerVolumeSpecName "kube-api-access-l6mzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.683761 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30a1b114-0b4e-43ff-815b-0983ca7a86e6-kube-api-access-z96p6" (OuterVolumeSpecName: "kube-api-access-z96p6") pod "30a1b114-0b4e-43ff-815b-0983ca7a86e6" (UID: "30a1b114-0b4e-43ff-815b-0983ca7a86e6"). InnerVolumeSpecName "kube-api-access-z96p6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.685092 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e29ef9a3-9060-4301-ad2c-3d62fe66c6b9-kube-api-access-9mkjn" (OuterVolumeSpecName: "kube-api-access-9mkjn") pod "e29ef9a3-9060-4301-ad2c-3d62fe66c6b9" (UID: "e29ef9a3-9060-4301-ad2c-3d62fe66c6b9"). InnerVolumeSpecName "kube-api-access-9mkjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.780938 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30a1b114-0b4e-43ff-815b-0983ca7a86e6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.781203 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z96p6\" (UniqueName: \"kubernetes.io/projected/30a1b114-0b4e-43ff-815b-0983ca7a86e6-kube-api-access-z96p6\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.781301 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mkjn\" (UniqueName: \"kubernetes.io/projected/e29ef9a3-9060-4301-ad2c-3d62fe66c6b9-kube-api-access-9mkjn\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.781359 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6a42d09-fb12-4466-a792-49998f29ccc4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.781423 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55732ab9-9610-4b4f-bae3-516b623e0608-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.781507 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6mzs\" (UniqueName: \"kubernetes.io/projected/d6a42d09-fb12-4466-a792-49998f29ccc4-kube-api-access-l6mzs\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.781584 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e29ef9a3-9060-4301-ad2c-3d62fe66c6b9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.781792 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzqdm\" (UniqueName: \"kubernetes.io/projected/d3b0f7e7-9035-4c33-940b-8bdd905d97b5-kube-api-access-lzqdm\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.848937 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-324e-account-create-update-rdb5j" event={"ID":"8fb6a553-c5fd-427f-96e6-b5bd6d3a4971","Type":"ContainerDied","Data":"6453e00b8333d08833df4f469cd5005df0d4b9894a281253ffa62a8ea3909e6a"} Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.848987 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6453e00b8333d08833df4f469cd5005df0d4b9894a281253ffa62a8ea3909e6a" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.849052 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-324e-account-create-update-rdb5j" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.857375 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8657-account-create-update-jsmmz" event={"ID":"e29ef9a3-9060-4301-ad2c-3d62fe66c6b9","Type":"ContainerDied","Data":"579174cd63e33fe2cfc74401d12dba42d7aacbab154a62e4a73338a9c3402d24"} Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.857420 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="579174cd63e33fe2cfc74401d12dba42d7aacbab154a62e4a73338a9c3402d24" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.857713 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8657-account-create-update-jsmmz" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.858448 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8cd1-account-create-update-xj8tj" event={"ID":"d3b0f7e7-9035-4c33-940b-8bdd905d97b5","Type":"ContainerDied","Data":"18201aca890626fc74281693ad8dcda7a9e82c0ec305201d3f8ed6a5028b21c3"} Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.858456 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8cd1-account-create-update-xj8tj" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.858499 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18201aca890626fc74281693ad8dcda7a9e82c0ec305201d3f8ed6a5028b21c3" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.866604 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-n85zz" event={"ID":"d6a42d09-fb12-4466-a792-49998f29ccc4","Type":"ContainerDied","Data":"0d27fd2f6257b6ef69603863befc7f8eeb5546cc23d612d9aab8abd0b62c9ffa"} Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.866651 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d27fd2f6257b6ef69603863befc7f8eeb5546cc23d612d9aab8abd0b62c9ffa" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.866803 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-n85zz" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.868197 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gcsr9" event={"ID":"55732ab9-9610-4b4f-bae3-516b623e0608","Type":"ContainerDied","Data":"930c7e0a2cd5986f7b5a71f6263ad9e6ab47464bc738b97dd96428bab2d3eae3"} Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.868222 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="930c7e0a2cd5986f7b5a71f6263ad9e6ab47464bc738b97dd96428bab2d3eae3" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.868240 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gcsr9" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.869529 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-prlzg" event={"ID":"30a1b114-0b4e-43ff-815b-0983ca7a86e6","Type":"ContainerDied","Data":"3a3228122918ddc595705de79d5698e91b10af2c1731458e5afc4391fa25a498"} Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.869552 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-prlzg" Dec 01 14:54:41 crc kubenswrapper[4810]: I1201 14:54:41.869558 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a3228122918ddc595705de79d5698e91b10af2c1731458e5afc4391fa25a498" Dec 01 14:54:43 crc kubenswrapper[4810]: I1201 14:54:43.359972 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 01 14:54:44 crc kubenswrapper[4810]: I1201 14:54:44.946320 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:54:44 crc kubenswrapper[4810]: E1201 14:54:44.946536 4810 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 14:54:44 crc kubenswrapper[4810]: E1201 14:54:44.946709 4810 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 14:54:44 crc kubenswrapper[4810]: E1201 14:54:44.946766 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift podName:1f455986-23c9-4088-9829-51c74e5b7770 nodeName:}" failed. No retries permitted until 2025-12-01 14:55:00.946750075 +0000 UTC m=+1266.710259678 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift") pod "swift-storage-0" (UID: "1f455986-23c9-4088-9829-51c74e5b7770") : configmap "swift-ring-files" not found Dec 01 14:54:46 crc kubenswrapper[4810]: I1201 14:54:46.904671 4810 generic.go:334] "Generic (PLEG): container finished" podID="beec7c3d-1e8b-42c7-9e07-978388973b80" containerID="6f00749a221c1f7e6ab22d685c419d81ad12d39e73efcda6cf4bc60436e158e9" exitCode=0 Dec 01 14:54:46 crc kubenswrapper[4810]: I1201 14:54:46.904718 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mslgc" event={"ID":"beec7c3d-1e8b-42c7-9e07-978388973b80","Type":"ContainerDied","Data":"6f00749a221c1f7e6ab22d685c419d81ad12d39e73efcda6cf4bc60436e158e9"} Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.047784 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-h4brs"] Dec 01 14:54:47 crc kubenswrapper[4810]: E1201 14:54:47.048162 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55732ab9-9610-4b4f-bae3-516b623e0608" containerName="mariadb-database-create" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.048186 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="55732ab9-9610-4b4f-bae3-516b623e0608" containerName="mariadb-database-create" Dec 01 14:54:47 crc kubenswrapper[4810]: E1201 14:54:47.048199 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e29ef9a3-9060-4301-ad2c-3d62fe66c6b9" containerName="mariadb-account-create-update" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.048207 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e29ef9a3-9060-4301-ad2c-3d62fe66c6b9" containerName="mariadb-account-create-update" Dec 01 14:54:47 crc kubenswrapper[4810]: E1201 14:54:47.048224 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a1b114-0b4e-43ff-815b-0983ca7a86e6" containerName="mariadb-database-create" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.048232 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a1b114-0b4e-43ff-815b-0983ca7a86e6" containerName="mariadb-database-create" Dec 01 14:54:47 crc kubenswrapper[4810]: E1201 14:54:47.048247 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9" containerName="init" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.048254 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9" containerName="init" Dec 01 14:54:47 crc kubenswrapper[4810]: E1201 14:54:47.048275 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9" containerName="dnsmasq-dns" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.048282 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9" containerName="dnsmasq-dns" Dec 01 14:54:47 crc kubenswrapper[4810]: E1201 14:54:47.048290 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6a42d09-fb12-4466-a792-49998f29ccc4" containerName="mariadb-database-create" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.048298 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6a42d09-fb12-4466-a792-49998f29ccc4" containerName="mariadb-database-create" Dec 01 14:54:47 crc kubenswrapper[4810]: E1201 14:54:47.048313 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3b0f7e7-9035-4c33-940b-8bdd905d97b5" containerName="mariadb-account-create-update" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.048323 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3b0f7e7-9035-4c33-940b-8bdd905d97b5" containerName="mariadb-account-create-update" Dec 01 14:54:47 crc kubenswrapper[4810]: E1201 14:54:47.048338 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fb6a553-c5fd-427f-96e6-b5bd6d3a4971" containerName="mariadb-account-create-update" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.048346 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fb6a553-c5fd-427f-96e6-b5bd6d3a4971" containerName="mariadb-account-create-update" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.048561 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fb6a553-c5fd-427f-96e6-b5bd6d3a4971" containerName="mariadb-account-create-update" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.048587 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fb61aa6-dd8f-4fb5-9697-76dc914ef2a9" containerName="dnsmasq-dns" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.048598 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="30a1b114-0b4e-43ff-815b-0983ca7a86e6" containerName="mariadb-database-create" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.048611 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6a42d09-fb12-4466-a792-49998f29ccc4" containerName="mariadb-database-create" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.048628 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="55732ab9-9610-4b4f-bae3-516b623e0608" containerName="mariadb-database-create" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.048639 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e29ef9a3-9060-4301-ad2c-3d62fe66c6b9" containerName="mariadb-account-create-update" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.048649 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3b0f7e7-9035-4c33-940b-8bdd905d97b5" containerName="mariadb-account-create-update" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.049321 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-h4brs" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.051007 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.051254 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-6kdsr" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.060806 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-h4brs"] Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.183685 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/986d954e-89dd-4878-bbdd-b93e6a0ca155-db-sync-config-data\") pod \"glance-db-sync-h4brs\" (UID: \"986d954e-89dd-4878-bbdd-b93e6a0ca155\") " pod="openstack/glance-db-sync-h4brs" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.183792 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/986d954e-89dd-4878-bbdd-b93e6a0ca155-combined-ca-bundle\") pod \"glance-db-sync-h4brs\" (UID: \"986d954e-89dd-4878-bbdd-b93e6a0ca155\") " pod="openstack/glance-db-sync-h4brs" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.183822 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whcww\" (UniqueName: \"kubernetes.io/projected/986d954e-89dd-4878-bbdd-b93e6a0ca155-kube-api-access-whcww\") pod \"glance-db-sync-h4brs\" (UID: \"986d954e-89dd-4878-bbdd-b93e6a0ca155\") " pod="openstack/glance-db-sync-h4brs" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.183879 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/986d954e-89dd-4878-bbdd-b93e6a0ca155-config-data\") pod \"glance-db-sync-h4brs\" (UID: \"986d954e-89dd-4878-bbdd-b93e6a0ca155\") " pod="openstack/glance-db-sync-h4brs" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.285326 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/986d954e-89dd-4878-bbdd-b93e6a0ca155-combined-ca-bundle\") pod \"glance-db-sync-h4brs\" (UID: \"986d954e-89dd-4878-bbdd-b93e6a0ca155\") " pod="openstack/glance-db-sync-h4brs" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.285736 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whcww\" (UniqueName: \"kubernetes.io/projected/986d954e-89dd-4878-bbdd-b93e6a0ca155-kube-api-access-whcww\") pod \"glance-db-sync-h4brs\" (UID: \"986d954e-89dd-4878-bbdd-b93e6a0ca155\") " pod="openstack/glance-db-sync-h4brs" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.285805 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/986d954e-89dd-4878-bbdd-b93e6a0ca155-config-data\") pod \"glance-db-sync-h4brs\" (UID: \"986d954e-89dd-4878-bbdd-b93e6a0ca155\") " pod="openstack/glance-db-sync-h4brs" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.285891 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/986d954e-89dd-4878-bbdd-b93e6a0ca155-db-sync-config-data\") pod \"glance-db-sync-h4brs\" (UID: \"986d954e-89dd-4878-bbdd-b93e6a0ca155\") " pod="openstack/glance-db-sync-h4brs" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.291689 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/986d954e-89dd-4878-bbdd-b93e6a0ca155-combined-ca-bundle\") pod \"glance-db-sync-h4brs\" (UID: \"986d954e-89dd-4878-bbdd-b93e6a0ca155\") " pod="openstack/glance-db-sync-h4brs" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.292635 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/986d954e-89dd-4878-bbdd-b93e6a0ca155-db-sync-config-data\") pod \"glance-db-sync-h4brs\" (UID: \"986d954e-89dd-4878-bbdd-b93e6a0ca155\") " pod="openstack/glance-db-sync-h4brs" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.294085 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/986d954e-89dd-4878-bbdd-b93e6a0ca155-config-data\") pod \"glance-db-sync-h4brs\" (UID: \"986d954e-89dd-4878-bbdd-b93e6a0ca155\") " pod="openstack/glance-db-sync-h4brs" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.304271 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whcww\" (UniqueName: \"kubernetes.io/projected/986d954e-89dd-4878-bbdd-b93e6a0ca155-kube-api-access-whcww\") pod \"glance-db-sync-h4brs\" (UID: \"986d954e-89dd-4878-bbdd-b93e6a0ca155\") " pod="openstack/glance-db-sync-h4brs" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.364794 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-h4brs" Dec 01 14:54:47 crc kubenswrapper[4810]: I1201 14:54:47.972529 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-h4brs"] Dec 01 14:54:47 crc kubenswrapper[4810]: W1201 14:54:47.977909 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod986d954e_89dd_4878_bbdd_b93e6a0ca155.slice/crio-4d22615abe344366b599db7410cafca71f369ce6428155ee9a3711714c028493 WatchSource:0}: Error finding container 4d22615abe344366b599db7410cafca71f369ce6428155ee9a3711714c028493: Status 404 returned error can't find the container with id 4d22615abe344366b599db7410cafca71f369ce6428155ee9a3711714c028493 Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.168059 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.303919 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/beec7c3d-1e8b-42c7-9e07-978388973b80-swiftconf\") pod \"beec7c3d-1e8b-42c7-9e07-978388973b80\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.304029 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/beec7c3d-1e8b-42c7-9e07-978388973b80-dispersionconf\") pod \"beec7c3d-1e8b-42c7-9e07-978388973b80\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.304082 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/beec7c3d-1e8b-42c7-9e07-978388973b80-combined-ca-bundle\") pod \"beec7c3d-1e8b-42c7-9e07-978388973b80\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.304115 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/beec7c3d-1e8b-42c7-9e07-978388973b80-ring-data-devices\") pod \"beec7c3d-1e8b-42c7-9e07-978388973b80\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.304161 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw6hp\" (UniqueName: \"kubernetes.io/projected/beec7c3d-1e8b-42c7-9e07-978388973b80-kube-api-access-dw6hp\") pod \"beec7c3d-1e8b-42c7-9e07-978388973b80\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.304199 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/beec7c3d-1e8b-42c7-9e07-978388973b80-scripts\") pod \"beec7c3d-1e8b-42c7-9e07-978388973b80\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.304252 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/beec7c3d-1e8b-42c7-9e07-978388973b80-etc-swift\") pod \"beec7c3d-1e8b-42c7-9e07-978388973b80\" (UID: \"beec7c3d-1e8b-42c7-9e07-978388973b80\") " Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.306366 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/beec7c3d-1e8b-42c7-9e07-978388973b80-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "beec7c3d-1e8b-42c7-9e07-978388973b80" (UID: "beec7c3d-1e8b-42c7-9e07-978388973b80"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.308721 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/beec7c3d-1e8b-42c7-9e07-978388973b80-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "beec7c3d-1e8b-42c7-9e07-978388973b80" (UID: "beec7c3d-1e8b-42c7-9e07-978388973b80"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.312786 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beec7c3d-1e8b-42c7-9e07-978388973b80-kube-api-access-dw6hp" (OuterVolumeSpecName: "kube-api-access-dw6hp") pod "beec7c3d-1e8b-42c7-9e07-978388973b80" (UID: "beec7c3d-1e8b-42c7-9e07-978388973b80"). InnerVolumeSpecName "kube-api-access-dw6hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.319801 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/beec7c3d-1e8b-42c7-9e07-978388973b80-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "beec7c3d-1e8b-42c7-9e07-978388973b80" (UID: "beec7c3d-1e8b-42c7-9e07-978388973b80"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.333638 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/beec7c3d-1e8b-42c7-9e07-978388973b80-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "beec7c3d-1e8b-42c7-9e07-978388973b80" (UID: "beec7c3d-1e8b-42c7-9e07-978388973b80"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.334404 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/beec7c3d-1e8b-42c7-9e07-978388973b80-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "beec7c3d-1e8b-42c7-9e07-978388973b80" (UID: "beec7c3d-1e8b-42c7-9e07-978388973b80"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.339362 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/beec7c3d-1e8b-42c7-9e07-978388973b80-scripts" (OuterVolumeSpecName: "scripts") pod "beec7c3d-1e8b-42c7-9e07-978388973b80" (UID: "beec7c3d-1e8b-42c7-9e07-978388973b80"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.405640 4810 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/beec7c3d-1e8b-42c7-9e07-978388973b80-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.405682 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/beec7c3d-1e8b-42c7-9e07-978388973b80-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.405695 4810 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/beec7c3d-1e8b-42c7-9e07-978388973b80-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.405705 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw6hp\" (UniqueName: \"kubernetes.io/projected/beec7c3d-1e8b-42c7-9e07-978388973b80-kube-api-access-dw6hp\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.405720 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/beec7c3d-1e8b-42c7-9e07-978388973b80-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.405730 4810 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/beec7c3d-1e8b-42c7-9e07-978388973b80-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.405741 4810 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/beec7c3d-1e8b-42c7-9e07-978388973b80-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.919509 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mslgc" event={"ID":"beec7c3d-1e8b-42c7-9e07-978388973b80","Type":"ContainerDied","Data":"3bce6aa8345ea924a8105d9facbd40eb8d6b05c3de6b2094a38304f8822fbf05"} Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.919545 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bce6aa8345ea924a8105d9facbd40eb8d6b05c3de6b2094a38304f8822fbf05" Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.919530 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mslgc" Dec 01 14:54:48 crc kubenswrapper[4810]: I1201 14:54:48.920708 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-h4brs" event={"ID":"986d954e-89dd-4878-bbdd-b93e6a0ca155","Type":"ContainerStarted","Data":"4d22615abe344366b599db7410cafca71f369ce6428155ee9a3711714c028493"} Dec 01 14:54:49 crc kubenswrapper[4810]: I1201 14:54:49.932453 4810 generic.go:334] "Generic (PLEG): container finished" podID="56bbf32f-35b5-4f92-bfc0-9266b0ecccac" containerID="e23662080306dd3c3cc28ceb3e4c2195c124ff9bccb4a3034f0a68f6534ad3d3" exitCode=0 Dec 01 14:54:49 crc kubenswrapper[4810]: I1201 14:54:49.932586 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"56bbf32f-35b5-4f92-bfc0-9266b0ecccac","Type":"ContainerDied","Data":"e23662080306dd3c3cc28ceb3e4c2195c124ff9bccb4a3034f0a68f6534ad3d3"} Dec 01 14:54:49 crc kubenswrapper[4810]: I1201 14:54:49.939871 4810 generic.go:334] "Generic (PLEG): container finished" podID="3b81b6a9-0be6-4ce1-81b4-043edb1c221f" containerID="823b399b01a75b82b70c3653a5f597fb68c1ca3af1fbf06c04a2d840c74494cb" exitCode=0 Dec 01 14:54:49 crc kubenswrapper[4810]: I1201 14:54:49.940019 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3b81b6a9-0be6-4ce1-81b4-043edb1c221f","Type":"ContainerDied","Data":"823b399b01a75b82b70c3653a5f597fb68c1ca3af1fbf06c04a2d840c74494cb"} Dec 01 14:54:50 crc kubenswrapper[4810]: I1201 14:54:50.952242 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3b81b6a9-0be6-4ce1-81b4-043edb1c221f","Type":"ContainerStarted","Data":"74e6d90ac1ab1306dd26c5e8c6cd4cd8b45f47eb0ab68865c9c1575f423fce76"} Dec 01 14:54:50 crc kubenswrapper[4810]: I1201 14:54:50.952846 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 01 14:54:50 crc kubenswrapper[4810]: I1201 14:54:50.954931 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"56bbf32f-35b5-4f92-bfc0-9266b0ecccac","Type":"ContainerStarted","Data":"82a26c84ad224e60a4fa117cf5e1d03643cfd46868d1eeb656fac71a268ad1e5"} Dec 01 14:54:50 crc kubenswrapper[4810]: I1201 14:54:50.955372 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:54:50 crc kubenswrapper[4810]: I1201 14:54:50.981441 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.559851051 podStartE2EDuration="59.981426398s" podCreationTimestamp="2025-12-01 14:53:51 +0000 UTC" firstStartedPulling="2025-12-01 14:53:53.043026709 +0000 UTC m=+1198.806536302" lastFinishedPulling="2025-12-01 14:54:15.464602046 +0000 UTC m=+1221.228111649" observedRunningTime="2025-12-01 14:54:50.975075925 +0000 UTC m=+1256.738585538" watchObservedRunningTime="2025-12-01 14:54:50.981426398 +0000 UTC m=+1256.744936001" Dec 01 14:54:51 crc kubenswrapper[4810]: I1201 14:54:51.001750 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.295530442 podStartE2EDuration="1m0.001723509s" podCreationTimestamp="2025-12-01 14:53:51 +0000 UTC" firstStartedPulling="2025-12-01 14:53:53.947976151 +0000 UTC m=+1199.711485754" lastFinishedPulling="2025-12-01 14:54:15.654169218 +0000 UTC m=+1221.417678821" observedRunningTime="2025-12-01 14:54:50.996045145 +0000 UTC m=+1256.759554778" watchObservedRunningTime="2025-12-01 14:54:51.001723509 +0000 UTC m=+1256.765233112" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.357547 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.360439 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-hfmpn" podUID="88635a0d-cb9d-43c1-9086-5623c9a780f8" containerName="ovn-controller" probeResult="failure" output=< Dec 01 14:54:52 crc kubenswrapper[4810]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 01 14:54:52 crc kubenswrapper[4810]: > Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.381133 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-x82dv" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.592923 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-hfmpn-config-gxx8k"] Dec 01 14:54:52 crc kubenswrapper[4810]: E1201 14:54:52.593293 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beec7c3d-1e8b-42c7-9e07-978388973b80" containerName="swift-ring-rebalance" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.593316 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="beec7c3d-1e8b-42c7-9e07-978388973b80" containerName="swift-ring-rebalance" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.593539 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="beec7c3d-1e8b-42c7-9e07-978388973b80" containerName="swift-ring-rebalance" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.594173 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.597058 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.612291 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hfmpn-config-gxx8k"] Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.744188 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d0e2997d-4287-4301-8796-426e07a744f6-var-log-ovn\") pod \"ovn-controller-hfmpn-config-gxx8k\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.744391 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d0e2997d-4287-4301-8796-426e07a744f6-additional-scripts\") pod \"ovn-controller-hfmpn-config-gxx8k\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.744409 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0e2997d-4287-4301-8796-426e07a744f6-scripts\") pod \"ovn-controller-hfmpn-config-gxx8k\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.744452 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d0e2997d-4287-4301-8796-426e07a744f6-var-run-ovn\") pod \"ovn-controller-hfmpn-config-gxx8k\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.744506 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d0e2997d-4287-4301-8796-426e07a744f6-var-run\") pod \"ovn-controller-hfmpn-config-gxx8k\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.744535 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fll6f\" (UniqueName: \"kubernetes.io/projected/d0e2997d-4287-4301-8796-426e07a744f6-kube-api-access-fll6f\") pod \"ovn-controller-hfmpn-config-gxx8k\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.846265 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d0e2997d-4287-4301-8796-426e07a744f6-additional-scripts\") pod \"ovn-controller-hfmpn-config-gxx8k\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.846401 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0e2997d-4287-4301-8796-426e07a744f6-scripts\") pod \"ovn-controller-hfmpn-config-gxx8k\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.847185 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d0e2997d-4287-4301-8796-426e07a744f6-additional-scripts\") pod \"ovn-controller-hfmpn-config-gxx8k\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.849339 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0e2997d-4287-4301-8796-426e07a744f6-scripts\") pod \"ovn-controller-hfmpn-config-gxx8k\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.849594 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d0e2997d-4287-4301-8796-426e07a744f6-var-run-ovn\") pod \"ovn-controller-hfmpn-config-gxx8k\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.849741 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d0e2997d-4287-4301-8796-426e07a744f6-var-run\") pod \"ovn-controller-hfmpn-config-gxx8k\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.849796 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fll6f\" (UniqueName: \"kubernetes.io/projected/d0e2997d-4287-4301-8796-426e07a744f6-kube-api-access-fll6f\") pod \"ovn-controller-hfmpn-config-gxx8k\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.849850 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d0e2997d-4287-4301-8796-426e07a744f6-var-log-ovn\") pod \"ovn-controller-hfmpn-config-gxx8k\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.850453 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d0e2997d-4287-4301-8796-426e07a744f6-var-log-ovn\") pod \"ovn-controller-hfmpn-config-gxx8k\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.850569 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d0e2997d-4287-4301-8796-426e07a744f6-var-run-ovn\") pod \"ovn-controller-hfmpn-config-gxx8k\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.850629 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d0e2997d-4287-4301-8796-426e07a744f6-var-run\") pod \"ovn-controller-hfmpn-config-gxx8k\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.870497 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fll6f\" (UniqueName: \"kubernetes.io/projected/d0e2997d-4287-4301-8796-426e07a744f6-kube-api-access-fll6f\") pod \"ovn-controller-hfmpn-config-gxx8k\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:52 crc kubenswrapper[4810]: I1201 14:54:52.912368 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:54:53 crc kubenswrapper[4810]: I1201 14:54:53.432800 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-hfmpn-config-gxx8k"] Dec 01 14:54:53 crc kubenswrapper[4810]: I1201 14:54:53.980879 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hfmpn-config-gxx8k" event={"ID":"d0e2997d-4287-4301-8796-426e07a744f6","Type":"ContainerStarted","Data":"b1d04ab295a44cc19fce17280e189d3a9861c9b22579a19c562d1fae072f4910"} Dec 01 14:54:54 crc kubenswrapper[4810]: I1201 14:54:54.993013 4810 generic.go:334] "Generic (PLEG): container finished" podID="d0e2997d-4287-4301-8796-426e07a744f6" containerID="87b827b5dea17db7f9051c490e22381eb54e6dc308ffffd4d8f479968b5b8aef" exitCode=0 Dec 01 14:54:54 crc kubenswrapper[4810]: I1201 14:54:54.993102 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hfmpn-config-gxx8k" event={"ID":"d0e2997d-4287-4301-8796-426e07a744f6","Type":"ContainerDied","Data":"87b827b5dea17db7f9051c490e22381eb54e6dc308ffffd4d8f479968b5b8aef"} Dec 01 14:54:57 crc kubenswrapper[4810]: I1201 14:54:57.359614 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-hfmpn" Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.551104 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.711571 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d0e2997d-4287-4301-8796-426e07a744f6-additional-scripts\") pod \"d0e2997d-4287-4301-8796-426e07a744f6\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.711681 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0e2997d-4287-4301-8796-426e07a744f6-scripts\") pod \"d0e2997d-4287-4301-8796-426e07a744f6\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.711718 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fll6f\" (UniqueName: \"kubernetes.io/projected/d0e2997d-4287-4301-8796-426e07a744f6-kube-api-access-fll6f\") pod \"d0e2997d-4287-4301-8796-426e07a744f6\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.711848 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d0e2997d-4287-4301-8796-426e07a744f6-var-run-ovn\") pod \"d0e2997d-4287-4301-8796-426e07a744f6\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.711877 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d0e2997d-4287-4301-8796-426e07a744f6-var-run\") pod \"d0e2997d-4287-4301-8796-426e07a744f6\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.711915 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d0e2997d-4287-4301-8796-426e07a744f6-var-log-ovn\") pod \"d0e2997d-4287-4301-8796-426e07a744f6\" (UID: \"d0e2997d-4287-4301-8796-426e07a744f6\") " Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.712317 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0e2997d-4287-4301-8796-426e07a744f6-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "d0e2997d-4287-4301-8796-426e07a744f6" (UID: "d0e2997d-4287-4301-8796-426e07a744f6"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.712329 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0e2997d-4287-4301-8796-426e07a744f6-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "d0e2997d-4287-4301-8796-426e07a744f6" (UID: "d0e2997d-4287-4301-8796-426e07a744f6"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.712370 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0e2997d-4287-4301-8796-426e07a744f6-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "d0e2997d-4287-4301-8796-426e07a744f6" (UID: "d0e2997d-4287-4301-8796-426e07a744f6"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.712743 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0e2997d-4287-4301-8796-426e07a744f6-var-run" (OuterVolumeSpecName: "var-run") pod "d0e2997d-4287-4301-8796-426e07a744f6" (UID: "d0e2997d-4287-4301-8796-426e07a744f6"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.713074 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0e2997d-4287-4301-8796-426e07a744f6-scripts" (OuterVolumeSpecName: "scripts") pod "d0e2997d-4287-4301-8796-426e07a744f6" (UID: "d0e2997d-4287-4301-8796-426e07a744f6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.715466 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0e2997d-4287-4301-8796-426e07a744f6-kube-api-access-fll6f" (OuterVolumeSpecName: "kube-api-access-fll6f") pod "d0e2997d-4287-4301-8796-426e07a744f6" (UID: "d0e2997d-4287-4301-8796-426e07a744f6"). InnerVolumeSpecName "kube-api-access-fll6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.814014 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fll6f\" (UniqueName: \"kubernetes.io/projected/d0e2997d-4287-4301-8796-426e07a744f6-kube-api-access-fll6f\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.814301 4810 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d0e2997d-4287-4301-8796-426e07a744f6-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.814385 4810 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d0e2997d-4287-4301-8796-426e07a744f6-var-run\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.814462 4810 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d0e2997d-4287-4301-8796-426e07a744f6-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.814670 4810 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d0e2997d-4287-4301-8796-426e07a744f6-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:00 crc kubenswrapper[4810]: I1201 14:55:00.814763 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0e2997d-4287-4301-8796-426e07a744f6-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:01 crc kubenswrapper[4810]: I1201 14:55:01.017870 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:55:01 crc kubenswrapper[4810]: I1201 14:55:01.024344 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift\") pod \"swift-storage-0\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " pod="openstack/swift-storage-0" Dec 01 14:55:01 crc kubenswrapper[4810]: I1201 14:55:01.035786 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hfmpn-config-gxx8k" event={"ID":"d0e2997d-4287-4301-8796-426e07a744f6","Type":"ContainerDied","Data":"b1d04ab295a44cc19fce17280e189d3a9861c9b22579a19c562d1fae072f4910"} Dec 01 14:55:01 crc kubenswrapper[4810]: I1201 14:55:01.035818 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1d04ab295a44cc19fce17280e189d3a9861c9b22579a19c562d1fae072f4910" Dec 01 14:55:01 crc kubenswrapper[4810]: I1201 14:55:01.035839 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hfmpn-config-gxx8k" Dec 01 14:55:01 crc kubenswrapper[4810]: I1201 14:55:01.313144 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 01 14:55:01 crc kubenswrapper[4810]: I1201 14:55:01.646883 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-hfmpn-config-gxx8k"] Dec 01 14:55:01 crc kubenswrapper[4810]: I1201 14:55:01.657981 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-hfmpn-config-gxx8k"] Dec 01 14:55:01 crc kubenswrapper[4810]: I1201 14:55:01.813957 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 01 14:55:02 crc kubenswrapper[4810]: I1201 14:55:02.042695 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerStarted","Data":"2f3e61aff3858aad8d5c49a32975a02fe7fc56a522b919189eb31626e1b14ee9"} Dec 01 14:55:02 crc kubenswrapper[4810]: I1201 14:55:02.044054 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-h4brs" event={"ID":"986d954e-89dd-4878-bbdd-b93e6a0ca155","Type":"ContainerStarted","Data":"8ebd6744e6dfcf18f3444cf3cf726b084641649414a0ce03fbfd64313fce06e2"} Dec 01 14:55:02 crc kubenswrapper[4810]: I1201 14:55:02.065147 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-h4brs" podStartSLOduration=2.452770166 podStartE2EDuration="15.065128988s" podCreationTimestamp="2025-12-01 14:54:47 +0000 UTC" firstStartedPulling="2025-12-01 14:54:47.980271511 +0000 UTC m=+1253.743781114" lastFinishedPulling="2025-12-01 14:55:00.592630333 +0000 UTC m=+1266.356139936" observedRunningTime="2025-12-01 14:55:02.056547765 +0000 UTC m=+1267.820057368" watchObservedRunningTime="2025-12-01 14:55:02.065128988 +0000 UTC m=+1267.828638591" Dec 01 14:55:02 crc kubenswrapper[4810]: I1201 14:55:02.501805 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0e2997d-4287-4301-8796-426e07a744f6" path="/var/lib/kubelet/pods/d0e2997d-4287-4301-8796-426e07a744f6/volumes" Dec 01 14:55:02 crc kubenswrapper[4810]: I1201 14:55:02.603682 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 01 14:55:02 crc kubenswrapper[4810]: I1201 14:55:02.878209 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-wt46c"] Dec 01 14:55:02 crc kubenswrapper[4810]: E1201 14:55:02.878844 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0e2997d-4287-4301-8796-426e07a744f6" containerName="ovn-config" Dec 01 14:55:02 crc kubenswrapper[4810]: I1201 14:55:02.878859 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0e2997d-4287-4301-8796-426e07a744f6" containerName="ovn-config" Dec 01 14:55:02 crc kubenswrapper[4810]: I1201 14:55:02.879037 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0e2997d-4287-4301-8796-426e07a744f6" containerName="ovn-config" Dec 01 14:55:02 crc kubenswrapper[4810]: I1201 14:55:02.879510 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-wt46c" Dec 01 14:55:02 crc kubenswrapper[4810]: I1201 14:55:02.892364 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-wt46c"] Dec 01 14:55:02 crc kubenswrapper[4810]: I1201 14:55:02.972108 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:55:02 crc kubenswrapper[4810]: I1201 14:55:02.972164 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:55:02 crc kubenswrapper[4810]: I1201 14:55:02.972205 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:55:02 crc kubenswrapper[4810]: I1201 14:55:02.972763 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8e1ade99d7746e14166be3755e17c0d1329bcaa748d059179439c1c0ba6fc843"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 14:55:02 crc kubenswrapper[4810]: I1201 14:55:02.972822 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://8e1ade99d7746e14166be3755e17c0d1329bcaa748d059179439c1c0ba6fc843" gracePeriod=600 Dec 01 14:55:02 crc kubenswrapper[4810]: I1201 14:55:02.991898 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-ff71-account-create-update-c62lr"] Dec 01 14:55:02 crc kubenswrapper[4810]: I1201 14:55:02.994135 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ff71-account-create-update-c62lr" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.000905 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.043157 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-ff71-account-create-update-c62lr"] Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.048451 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8nr7\" (UniqueName: \"kubernetes.io/projected/b96848ac-9038-4172-a36b-fe0889e3f6b6-kube-api-access-q8nr7\") pod \"barbican-db-create-wt46c\" (UID: \"b96848ac-9038-4172-a36b-fe0889e3f6b6\") " pod="openstack/barbican-db-create-wt46c" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.048709 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b96848ac-9038-4172-a36b-fe0889e3f6b6-operator-scripts\") pod \"barbican-db-create-wt46c\" (UID: \"b96848ac-9038-4172-a36b-fe0889e3f6b6\") " pod="openstack/barbican-db-create-wt46c" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.076948 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-jvlml"] Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.077945 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jvlml" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.089659 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-jvlml"] Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.150718 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0221e2e-817a-4fa9-9717-2bb622c966f0-operator-scripts\") pod \"barbican-ff71-account-create-update-c62lr\" (UID: \"f0221e2e-817a-4fa9-9717-2bb622c966f0\") " pod="openstack/barbican-ff71-account-create-update-c62lr" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.150958 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b96848ac-9038-4172-a36b-fe0889e3f6b6-operator-scripts\") pod \"barbican-db-create-wt46c\" (UID: \"b96848ac-9038-4172-a36b-fe0889e3f6b6\") " pod="openstack/barbican-db-create-wt46c" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.151039 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm475\" (UniqueName: \"kubernetes.io/projected/f0221e2e-817a-4fa9-9717-2bb622c966f0-kube-api-access-wm475\") pod \"barbican-ff71-account-create-update-c62lr\" (UID: \"f0221e2e-817a-4fa9-9717-2bb622c966f0\") " pod="openstack/barbican-ff71-account-create-update-c62lr" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.151072 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8nr7\" (UniqueName: \"kubernetes.io/projected/b96848ac-9038-4172-a36b-fe0889e3f6b6-kube-api-access-q8nr7\") pod \"barbican-db-create-wt46c\" (UID: \"b96848ac-9038-4172-a36b-fe0889e3f6b6\") " pod="openstack/barbican-db-create-wt46c" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.152377 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b96848ac-9038-4172-a36b-fe0889e3f6b6-operator-scripts\") pod \"barbican-db-create-wt46c\" (UID: \"b96848ac-9038-4172-a36b-fe0889e3f6b6\") " pod="openstack/barbican-db-create-wt46c" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.178773 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8nr7\" (UniqueName: \"kubernetes.io/projected/b96848ac-9038-4172-a36b-fe0889e3f6b6-kube-api-access-q8nr7\") pod \"barbican-db-create-wt46c\" (UID: \"b96848ac-9038-4172-a36b-fe0889e3f6b6\") " pod="openstack/barbican-db-create-wt46c" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.202097 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-wt46c" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.219704 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.226194 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-17c6-account-create-update-kt4nm"] Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.229146 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-17c6-account-create-update-kt4nm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.240855 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.247692 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-17c6-account-create-update-kt4nm"] Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.252216 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm475\" (UniqueName: \"kubernetes.io/projected/f0221e2e-817a-4fa9-9717-2bb622c966f0-kube-api-access-wm475\") pod \"barbican-ff71-account-create-update-c62lr\" (UID: \"f0221e2e-817a-4fa9-9717-2bb622c966f0\") " pod="openstack/barbican-ff71-account-create-update-c62lr" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.252278 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0221e2e-817a-4fa9-9717-2bb622c966f0-operator-scripts\") pod \"barbican-ff71-account-create-update-c62lr\" (UID: \"f0221e2e-817a-4fa9-9717-2bb622c966f0\") " pod="openstack/barbican-ff71-account-create-update-c62lr" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.252347 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2dmx\" (UniqueName: \"kubernetes.io/projected/01e51b04-eb2c-4102-9a86-d19ca8c566d6-kube-api-access-l2dmx\") pod \"cinder-db-create-jvlml\" (UID: \"01e51b04-eb2c-4102-9a86-d19ca8c566d6\") " pod="openstack/cinder-db-create-jvlml" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.252372 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01e51b04-eb2c-4102-9a86-d19ca8c566d6-operator-scripts\") pod \"cinder-db-create-jvlml\" (UID: \"01e51b04-eb2c-4102-9a86-d19ca8c566d6\") " pod="openstack/cinder-db-create-jvlml" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.253270 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0221e2e-817a-4fa9-9717-2bb622c966f0-operator-scripts\") pod \"barbican-ff71-account-create-update-c62lr\" (UID: \"f0221e2e-817a-4fa9-9717-2bb622c966f0\") " pod="openstack/barbican-ff71-account-create-update-c62lr" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.314116 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm475\" (UniqueName: \"kubernetes.io/projected/f0221e2e-817a-4fa9-9717-2bb622c966f0-kube-api-access-wm475\") pod \"barbican-ff71-account-create-update-c62lr\" (UID: \"f0221e2e-817a-4fa9-9717-2bb622c966f0\") " pod="openstack/barbican-ff71-account-create-update-c62lr" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.328404 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ff71-account-create-update-c62lr" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.341635 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-tmqcq"] Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.342830 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tmqcq" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.352951 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-tmqcq"] Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.354046 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2dmx\" (UniqueName: \"kubernetes.io/projected/01e51b04-eb2c-4102-9a86-d19ca8c566d6-kube-api-access-l2dmx\") pod \"cinder-db-create-jvlml\" (UID: \"01e51b04-eb2c-4102-9a86-d19ca8c566d6\") " pod="openstack/cinder-db-create-jvlml" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.354117 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrmz5\" (UniqueName: \"kubernetes.io/projected/21d83a41-8ec9-424e-be10-fac93f693310-kube-api-access-qrmz5\") pod \"cinder-17c6-account-create-update-kt4nm\" (UID: \"21d83a41-8ec9-424e-be10-fac93f693310\") " pod="openstack/cinder-17c6-account-create-update-kt4nm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.354153 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01e51b04-eb2c-4102-9a86-d19ca8c566d6-operator-scripts\") pod \"cinder-db-create-jvlml\" (UID: \"01e51b04-eb2c-4102-9a86-d19ca8c566d6\") " pod="openstack/cinder-db-create-jvlml" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.354178 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21d83a41-8ec9-424e-be10-fac93f693310-operator-scripts\") pod \"cinder-17c6-account-create-update-kt4nm\" (UID: \"21d83a41-8ec9-424e-be10-fac93f693310\") " pod="openstack/cinder-17c6-account-create-update-kt4nm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.357130 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01e51b04-eb2c-4102-9a86-d19ca8c566d6-operator-scripts\") pod \"cinder-db-create-jvlml\" (UID: \"01e51b04-eb2c-4102-9a86-d19ca8c566d6\") " pod="openstack/cinder-db-create-jvlml" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.413660 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2dmx\" (UniqueName: \"kubernetes.io/projected/01e51b04-eb2c-4102-9a86-d19ca8c566d6-kube-api-access-l2dmx\") pod \"cinder-db-create-jvlml\" (UID: \"01e51b04-eb2c-4102-9a86-d19ca8c566d6\") " pod="openstack/cinder-db-create-jvlml" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.416014 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-tb6zm"] Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.425951 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tb6zm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.439150 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.439524 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.439906 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jvlml" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.457888 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcd634cb-7e87-4e64-9eb7-894a15a047e2-operator-scripts\") pod \"neutron-db-create-tmqcq\" (UID: \"bcd634cb-7e87-4e64-9eb7-894a15a047e2\") " pod="openstack/neutron-db-create-tmqcq" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.457986 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrmz5\" (UniqueName: \"kubernetes.io/projected/21d83a41-8ec9-424e-be10-fac93f693310-kube-api-access-qrmz5\") pod \"cinder-17c6-account-create-update-kt4nm\" (UID: \"21d83a41-8ec9-424e-be10-fac93f693310\") " pod="openstack/cinder-17c6-account-create-update-kt4nm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.458011 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gddh\" (UniqueName: \"kubernetes.io/projected/bcd634cb-7e87-4e64-9eb7-894a15a047e2-kube-api-access-8gddh\") pod \"neutron-db-create-tmqcq\" (UID: \"bcd634cb-7e87-4e64-9eb7-894a15a047e2\") " pod="openstack/neutron-db-create-tmqcq" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.458036 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21d83a41-8ec9-424e-be10-fac93f693310-operator-scripts\") pod \"cinder-17c6-account-create-update-kt4nm\" (UID: \"21d83a41-8ec9-424e-be10-fac93f693310\") " pod="openstack/cinder-17c6-account-create-update-kt4nm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.458890 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21d83a41-8ec9-424e-be10-fac93f693310-operator-scripts\") pod \"cinder-17c6-account-create-update-kt4nm\" (UID: \"21d83a41-8ec9-424e-be10-fac93f693310\") " pod="openstack/cinder-17c6-account-create-update-kt4nm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.460247 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c91e-account-create-update-wxssk"] Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.461334 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c91e-account-create-update-wxssk" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.477956 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.478186 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zlbmq" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.478888 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.500412 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-tb6zm"] Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.514223 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrmz5\" (UniqueName: \"kubernetes.io/projected/21d83a41-8ec9-424e-be10-fac93f693310-kube-api-access-qrmz5\") pod \"cinder-17c6-account-create-update-kt4nm\" (UID: \"21d83a41-8ec9-424e-be10-fac93f693310\") " pod="openstack/cinder-17c6-account-create-update-kt4nm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.556726 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c91e-account-create-update-wxssk"] Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.558835 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gddh\" (UniqueName: \"kubernetes.io/projected/bcd634cb-7e87-4e64-9eb7-894a15a047e2-kube-api-access-8gddh\") pod \"neutron-db-create-tmqcq\" (UID: \"bcd634cb-7e87-4e64-9eb7-894a15a047e2\") " pod="openstack/neutron-db-create-tmqcq" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.558864 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d17c667c-7985-4132-88ee-f0e82ab7d25c-operator-scripts\") pod \"neutron-c91e-account-create-update-wxssk\" (UID: \"d17c667c-7985-4132-88ee-f0e82ab7d25c\") " pod="openstack/neutron-c91e-account-create-update-wxssk" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.558884 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4zkx\" (UniqueName: \"kubernetes.io/projected/d17c667c-7985-4132-88ee-f0e82ab7d25c-kube-api-access-d4zkx\") pod \"neutron-c91e-account-create-update-wxssk\" (UID: \"d17c667c-7985-4132-88ee-f0e82ab7d25c\") " pod="openstack/neutron-c91e-account-create-update-wxssk" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.558958 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwxjh\" (UniqueName: \"kubernetes.io/projected/72889b41-41e2-4d58-8e6d-f372f75793b7-kube-api-access-cwxjh\") pod \"keystone-db-sync-tb6zm\" (UID: \"72889b41-41e2-4d58-8e6d-f372f75793b7\") " pod="openstack/keystone-db-sync-tb6zm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.559005 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72889b41-41e2-4d58-8e6d-f372f75793b7-combined-ca-bundle\") pod \"keystone-db-sync-tb6zm\" (UID: \"72889b41-41e2-4d58-8e6d-f372f75793b7\") " pod="openstack/keystone-db-sync-tb6zm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.559028 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcd634cb-7e87-4e64-9eb7-894a15a047e2-operator-scripts\") pod \"neutron-db-create-tmqcq\" (UID: \"bcd634cb-7e87-4e64-9eb7-894a15a047e2\") " pod="openstack/neutron-db-create-tmqcq" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.559051 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72889b41-41e2-4d58-8e6d-f372f75793b7-config-data\") pod \"keystone-db-sync-tb6zm\" (UID: \"72889b41-41e2-4d58-8e6d-f372f75793b7\") " pod="openstack/keystone-db-sync-tb6zm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.560995 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcd634cb-7e87-4e64-9eb7-894a15a047e2-operator-scripts\") pod \"neutron-db-create-tmqcq\" (UID: \"bcd634cb-7e87-4e64-9eb7-894a15a047e2\") " pod="openstack/neutron-db-create-tmqcq" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.609184 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gddh\" (UniqueName: \"kubernetes.io/projected/bcd634cb-7e87-4e64-9eb7-894a15a047e2-kube-api-access-8gddh\") pod \"neutron-db-create-tmqcq\" (UID: \"bcd634cb-7e87-4e64-9eb7-894a15a047e2\") " pod="openstack/neutron-db-create-tmqcq" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.660643 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwxjh\" (UniqueName: \"kubernetes.io/projected/72889b41-41e2-4d58-8e6d-f372f75793b7-kube-api-access-cwxjh\") pod \"keystone-db-sync-tb6zm\" (UID: \"72889b41-41e2-4d58-8e6d-f372f75793b7\") " pod="openstack/keystone-db-sync-tb6zm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.660721 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72889b41-41e2-4d58-8e6d-f372f75793b7-combined-ca-bundle\") pod \"keystone-db-sync-tb6zm\" (UID: \"72889b41-41e2-4d58-8e6d-f372f75793b7\") " pod="openstack/keystone-db-sync-tb6zm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.660768 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72889b41-41e2-4d58-8e6d-f372f75793b7-config-data\") pod \"keystone-db-sync-tb6zm\" (UID: \"72889b41-41e2-4d58-8e6d-f372f75793b7\") " pod="openstack/keystone-db-sync-tb6zm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.660822 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d17c667c-7985-4132-88ee-f0e82ab7d25c-operator-scripts\") pod \"neutron-c91e-account-create-update-wxssk\" (UID: \"d17c667c-7985-4132-88ee-f0e82ab7d25c\") " pod="openstack/neutron-c91e-account-create-update-wxssk" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.660849 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4zkx\" (UniqueName: \"kubernetes.io/projected/d17c667c-7985-4132-88ee-f0e82ab7d25c-kube-api-access-d4zkx\") pod \"neutron-c91e-account-create-update-wxssk\" (UID: \"d17c667c-7985-4132-88ee-f0e82ab7d25c\") " pod="openstack/neutron-c91e-account-create-update-wxssk" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.663894 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d17c667c-7985-4132-88ee-f0e82ab7d25c-operator-scripts\") pod \"neutron-c91e-account-create-update-wxssk\" (UID: \"d17c667c-7985-4132-88ee-f0e82ab7d25c\") " pod="openstack/neutron-c91e-account-create-update-wxssk" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.666307 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72889b41-41e2-4d58-8e6d-f372f75793b7-config-data\") pod \"keystone-db-sync-tb6zm\" (UID: \"72889b41-41e2-4d58-8e6d-f372f75793b7\") " pod="openstack/keystone-db-sync-tb6zm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.668090 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72889b41-41e2-4d58-8e6d-f372f75793b7-combined-ca-bundle\") pod \"keystone-db-sync-tb6zm\" (UID: \"72889b41-41e2-4d58-8e6d-f372f75793b7\") " pod="openstack/keystone-db-sync-tb6zm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.693974 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4zkx\" (UniqueName: \"kubernetes.io/projected/d17c667c-7985-4132-88ee-f0e82ab7d25c-kube-api-access-d4zkx\") pod \"neutron-c91e-account-create-update-wxssk\" (UID: \"d17c667c-7985-4132-88ee-f0e82ab7d25c\") " pod="openstack/neutron-c91e-account-create-update-wxssk" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.694446 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwxjh\" (UniqueName: \"kubernetes.io/projected/72889b41-41e2-4d58-8e6d-f372f75793b7-kube-api-access-cwxjh\") pod \"keystone-db-sync-tb6zm\" (UID: \"72889b41-41e2-4d58-8e6d-f372f75793b7\") " pod="openstack/keystone-db-sync-tb6zm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.704901 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-17c6-account-create-update-kt4nm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.719023 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tmqcq" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.814734 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tb6zm" Dec 01 14:55:03 crc kubenswrapper[4810]: I1201 14:55:03.857929 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c91e-account-create-update-wxssk" Dec 01 14:55:04 crc kubenswrapper[4810]: I1201 14:55:04.013961 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-wt46c"] Dec 01 14:55:04 crc kubenswrapper[4810]: I1201 14:55:04.029646 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-ff71-account-create-update-c62lr"] Dec 01 14:55:04 crc kubenswrapper[4810]: W1201 14:55:04.062423 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0221e2e_817a_4fa9_9717_2bb622c966f0.slice/crio-2a7d286764a6894129b2bac252fd20c0f595581041fba830b4ce6a5d543424e0 WatchSource:0}: Error finding container 2a7d286764a6894129b2bac252fd20c0f595581041fba830b4ce6a5d543424e0: Status 404 returned error can't find the container with id 2a7d286764a6894129b2bac252fd20c0f595581041fba830b4ce6a5d543424e0 Dec 01 14:55:04 crc kubenswrapper[4810]: I1201 14:55:04.072804 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-wt46c" event={"ID":"b96848ac-9038-4172-a36b-fe0889e3f6b6","Type":"ContainerStarted","Data":"ba0ecbf055638c02d2602b0452ffdbd7a3f11ec1d1428f60e042e08e86c5a2c0"} Dec 01 14:55:04 crc kubenswrapper[4810]: I1201 14:55:04.083889 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="8e1ade99d7746e14166be3755e17c0d1329bcaa748d059179439c1c0ba6fc843" exitCode=0 Dec 01 14:55:04 crc kubenswrapper[4810]: I1201 14:55:04.083924 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"8e1ade99d7746e14166be3755e17c0d1329bcaa748d059179439c1c0ba6fc843"} Dec 01 14:55:04 crc kubenswrapper[4810]: I1201 14:55:04.083951 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"5eb750c936738b66d65230eb7997d67c73bd78542b11814d460ae2c7e8d2d0eb"} Dec 01 14:55:04 crc kubenswrapper[4810]: I1201 14:55:04.083970 4810 scope.go:117] "RemoveContainer" containerID="7d9c11d61c66e363668182137baa230beaee1296f3efd5bf440ccc7a13833e7c" Dec 01 14:55:04 crc kubenswrapper[4810]: I1201 14:55:04.175963 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-jvlml"] Dec 01 14:55:04 crc kubenswrapper[4810]: I1201 14:55:04.397294 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-17c6-account-create-update-kt4nm"] Dec 01 14:55:04 crc kubenswrapper[4810]: I1201 14:55:04.577379 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-tmqcq"] Dec 01 14:55:04 crc kubenswrapper[4810]: I1201 14:55:04.577411 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-tb6zm"] Dec 01 14:55:04 crc kubenswrapper[4810]: I1201 14:55:04.584171 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c91e-account-create-update-wxssk"] Dec 01 14:55:05 crc kubenswrapper[4810]: I1201 14:55:05.105332 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tb6zm" event={"ID":"72889b41-41e2-4d58-8e6d-f372f75793b7","Type":"ContainerStarted","Data":"c6cb987c8e16d944c4702c8401d3e5668494861a62c20941e781e53f7d749f15"} Dec 01 14:55:05 crc kubenswrapper[4810]: I1201 14:55:05.107671 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-17c6-account-create-update-kt4nm" event={"ID":"21d83a41-8ec9-424e-be10-fac93f693310","Type":"ContainerStarted","Data":"7a4e6dbf40a4598967e3a698bdc82e692daf7aab7fb17c1ade0ae7bcc69af608"} Dec 01 14:55:05 crc kubenswrapper[4810]: I1201 14:55:05.109022 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c91e-account-create-update-wxssk" event={"ID":"d17c667c-7985-4132-88ee-f0e82ab7d25c","Type":"ContainerStarted","Data":"992a617dfdc29e901d824ce9541c9518206ef294344d5bb0912daaeaec9b6246"} Dec 01 14:55:05 crc kubenswrapper[4810]: I1201 14:55:05.109825 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jvlml" event={"ID":"01e51b04-eb2c-4102-9a86-d19ca8c566d6","Type":"ContainerStarted","Data":"1f2f0b062e687c1bd26b3f53d055f07265df30970fb8fe66eeaf3f1f556f4bac"} Dec 01 14:55:05 crc kubenswrapper[4810]: I1201 14:55:05.110591 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ff71-account-create-update-c62lr" event={"ID":"f0221e2e-817a-4fa9-9717-2bb622c966f0","Type":"ContainerStarted","Data":"2a7d286764a6894129b2bac252fd20c0f595581041fba830b4ce6a5d543424e0"} Dec 01 14:55:05 crc kubenswrapper[4810]: I1201 14:55:05.111599 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tmqcq" event={"ID":"bcd634cb-7e87-4e64-9eb7-894a15a047e2","Type":"ContainerStarted","Data":"d8951ab77e926361b7507aed9f3d09b9f269b85cf85d246f13a16d178f425de6"} Dec 01 14:55:06 crc kubenswrapper[4810]: I1201 14:55:06.122413 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-wt46c" event={"ID":"b96848ac-9038-4172-a36b-fe0889e3f6b6","Type":"ContainerStarted","Data":"8483ad5843c1db556402790a55a9fc7d801088aacf9c6a0cb8f53cc0e5bd13b9"} Dec 01 14:55:06 crc kubenswrapper[4810]: I1201 14:55:06.124341 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ff71-account-create-update-c62lr" event={"ID":"f0221e2e-817a-4fa9-9717-2bb622c966f0","Type":"ContainerStarted","Data":"4733dd6747571f5dcdd5fb240a5a3752d303ca44a559a0da9797ccf77d32d51b"} Dec 01 14:55:07 crc kubenswrapper[4810]: I1201 14:55:07.136704 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jvlml" event={"ID":"01e51b04-eb2c-4102-9a86-d19ca8c566d6","Type":"ContainerStarted","Data":"baa21b2e3fe7443b704b8d1a9ef87496a23321efcab3a8630d626d9398fdef10"} Dec 01 14:55:08 crc kubenswrapper[4810]: I1201 14:55:08.149193 4810 generic.go:334] "Generic (PLEG): container finished" podID="21d83a41-8ec9-424e-be10-fac93f693310" containerID="6a11bca1051282741fc502516c5d8575301311334a580a73092fab993519d600" exitCode=0 Dec 01 14:55:08 crc kubenswrapper[4810]: I1201 14:55:08.149300 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-17c6-account-create-update-kt4nm" event={"ID":"21d83a41-8ec9-424e-be10-fac93f693310","Type":"ContainerDied","Data":"6a11bca1051282741fc502516c5d8575301311334a580a73092fab993519d600"} Dec 01 14:55:08 crc kubenswrapper[4810]: I1201 14:55:08.174093 4810 generic.go:334] "Generic (PLEG): container finished" podID="d17c667c-7985-4132-88ee-f0e82ab7d25c" containerID="200b93a61e8f7c29987fcdbc9f71328c297d619526d29ff4810c1a71a69f6e6c" exitCode=0 Dec 01 14:55:08 crc kubenswrapper[4810]: I1201 14:55:08.174186 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c91e-account-create-update-wxssk" event={"ID":"d17c667c-7985-4132-88ee-f0e82ab7d25c","Type":"ContainerDied","Data":"200b93a61e8f7c29987fcdbc9f71328c297d619526d29ff4810c1a71a69f6e6c"} Dec 01 14:55:08 crc kubenswrapper[4810]: I1201 14:55:08.176710 4810 generic.go:334] "Generic (PLEG): container finished" podID="01e51b04-eb2c-4102-9a86-d19ca8c566d6" containerID="baa21b2e3fe7443b704b8d1a9ef87496a23321efcab3a8630d626d9398fdef10" exitCode=0 Dec 01 14:55:08 crc kubenswrapper[4810]: I1201 14:55:08.176816 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jvlml" event={"ID":"01e51b04-eb2c-4102-9a86-d19ca8c566d6","Type":"ContainerDied","Data":"baa21b2e3fe7443b704b8d1a9ef87496a23321efcab3a8630d626d9398fdef10"} Dec 01 14:55:08 crc kubenswrapper[4810]: I1201 14:55:08.185745 4810 generic.go:334] "Generic (PLEG): container finished" podID="f0221e2e-817a-4fa9-9717-2bb622c966f0" containerID="4733dd6747571f5dcdd5fb240a5a3752d303ca44a559a0da9797ccf77d32d51b" exitCode=0 Dec 01 14:55:08 crc kubenswrapper[4810]: I1201 14:55:08.186092 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ff71-account-create-update-c62lr" event={"ID":"f0221e2e-817a-4fa9-9717-2bb622c966f0","Type":"ContainerDied","Data":"4733dd6747571f5dcdd5fb240a5a3752d303ca44a559a0da9797ccf77d32d51b"} Dec 01 14:55:08 crc kubenswrapper[4810]: I1201 14:55:08.193897 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerStarted","Data":"7b27d0e70deed824a6d31268fa11d712beb18bff269cd2c61fb15665851547b6"} Dec 01 14:55:08 crc kubenswrapper[4810]: I1201 14:55:08.194019 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerStarted","Data":"cc265ca64611ebb2ffafcd64f73b893f0707c1a67a0e439ab6b7c8ef0dec3227"} Dec 01 14:55:08 crc kubenswrapper[4810]: I1201 14:55:08.204812 4810 generic.go:334] "Generic (PLEG): container finished" podID="bcd634cb-7e87-4e64-9eb7-894a15a047e2" containerID="9e960dff62c094d4b5bb80a9748389f999d9e2812e3fcfb9bc2930885059fd81" exitCode=0 Dec 01 14:55:08 crc kubenswrapper[4810]: I1201 14:55:08.204894 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tmqcq" event={"ID":"bcd634cb-7e87-4e64-9eb7-894a15a047e2","Type":"ContainerDied","Data":"9e960dff62c094d4b5bb80a9748389f999d9e2812e3fcfb9bc2930885059fd81"} Dec 01 14:55:08 crc kubenswrapper[4810]: I1201 14:55:08.208225 4810 generic.go:334] "Generic (PLEG): container finished" podID="b96848ac-9038-4172-a36b-fe0889e3f6b6" containerID="8483ad5843c1db556402790a55a9fc7d801088aacf9c6a0cb8f53cc0e5bd13b9" exitCode=0 Dec 01 14:55:08 crc kubenswrapper[4810]: I1201 14:55:08.208286 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-wt46c" event={"ID":"b96848ac-9038-4172-a36b-fe0889e3f6b6","Type":"ContainerDied","Data":"8483ad5843c1db556402790a55a9fc7d801088aacf9c6a0cb8f53cc0e5bd13b9"} Dec 01 14:55:09 crc kubenswrapper[4810]: I1201 14:55:09.226804 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerStarted","Data":"7a9772a6d14d26dd0282c93cad34c0f2f79a0710fc08b93b93ed46c435efb103"} Dec 01 14:55:09 crc kubenswrapper[4810]: I1201 14:55:09.227284 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerStarted","Data":"8124aab34c8595c54770f6fe4ffe9d15bf793940c0b2cdffcb037b35b9a1e5ce"} Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.038333 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-17c6-account-create-update-kt4nm" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.045540 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ff71-account-create-update-c62lr" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.063689 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-wt46c" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.085735 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c91e-account-create-update-wxssk" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.117991 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrmz5\" (UniqueName: \"kubernetes.io/projected/21d83a41-8ec9-424e-be10-fac93f693310-kube-api-access-qrmz5\") pod \"21d83a41-8ec9-424e-be10-fac93f693310\" (UID: \"21d83a41-8ec9-424e-be10-fac93f693310\") " Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.118117 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b96848ac-9038-4172-a36b-fe0889e3f6b6-operator-scripts\") pod \"b96848ac-9038-4172-a36b-fe0889e3f6b6\" (UID: \"b96848ac-9038-4172-a36b-fe0889e3f6b6\") " Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.118172 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0221e2e-817a-4fa9-9717-2bb622c966f0-operator-scripts\") pod \"f0221e2e-817a-4fa9-9717-2bb622c966f0\" (UID: \"f0221e2e-817a-4fa9-9717-2bb622c966f0\") " Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.118191 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8nr7\" (UniqueName: \"kubernetes.io/projected/b96848ac-9038-4172-a36b-fe0889e3f6b6-kube-api-access-q8nr7\") pod \"b96848ac-9038-4172-a36b-fe0889e3f6b6\" (UID: \"b96848ac-9038-4172-a36b-fe0889e3f6b6\") " Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.118228 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm475\" (UniqueName: \"kubernetes.io/projected/f0221e2e-817a-4fa9-9717-2bb622c966f0-kube-api-access-wm475\") pod \"f0221e2e-817a-4fa9-9717-2bb622c966f0\" (UID: \"f0221e2e-817a-4fa9-9717-2bb622c966f0\") " Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.118271 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21d83a41-8ec9-424e-be10-fac93f693310-operator-scripts\") pod \"21d83a41-8ec9-424e-be10-fac93f693310\" (UID: \"21d83a41-8ec9-424e-be10-fac93f693310\") " Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.120114 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21d83a41-8ec9-424e-be10-fac93f693310-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "21d83a41-8ec9-424e-be10-fac93f693310" (UID: "21d83a41-8ec9-424e-be10-fac93f693310"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.120597 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0221e2e-817a-4fa9-9717-2bb622c966f0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f0221e2e-817a-4fa9-9717-2bb622c966f0" (UID: "f0221e2e-817a-4fa9-9717-2bb622c966f0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.123407 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jvlml" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.124650 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0221e2e-817a-4fa9-9717-2bb622c966f0-kube-api-access-wm475" (OuterVolumeSpecName: "kube-api-access-wm475") pod "f0221e2e-817a-4fa9-9717-2bb622c966f0" (UID: "f0221e2e-817a-4fa9-9717-2bb622c966f0"). InnerVolumeSpecName "kube-api-access-wm475". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.124712 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21d83a41-8ec9-424e-be10-fac93f693310-kube-api-access-qrmz5" (OuterVolumeSpecName: "kube-api-access-qrmz5") pod "21d83a41-8ec9-424e-be10-fac93f693310" (UID: "21d83a41-8ec9-424e-be10-fac93f693310"). InnerVolumeSpecName "kube-api-access-qrmz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.124752 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b96848ac-9038-4172-a36b-fe0889e3f6b6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b96848ac-9038-4172-a36b-fe0889e3f6b6" (UID: "b96848ac-9038-4172-a36b-fe0889e3f6b6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.132761 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b96848ac-9038-4172-a36b-fe0889e3f6b6-kube-api-access-q8nr7" (OuterVolumeSpecName: "kube-api-access-q8nr7") pod "b96848ac-9038-4172-a36b-fe0889e3f6b6" (UID: "b96848ac-9038-4172-a36b-fe0889e3f6b6"). InnerVolumeSpecName "kube-api-access-q8nr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.167099 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tmqcq" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.219326 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d17c667c-7985-4132-88ee-f0e82ab7d25c-operator-scripts\") pod \"d17c667c-7985-4132-88ee-f0e82ab7d25c\" (UID: \"d17c667c-7985-4132-88ee-f0e82ab7d25c\") " Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.219426 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gddh\" (UniqueName: \"kubernetes.io/projected/bcd634cb-7e87-4e64-9eb7-894a15a047e2-kube-api-access-8gddh\") pod \"bcd634cb-7e87-4e64-9eb7-894a15a047e2\" (UID: \"bcd634cb-7e87-4e64-9eb7-894a15a047e2\") " Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.219512 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcd634cb-7e87-4e64-9eb7-894a15a047e2-operator-scripts\") pod \"bcd634cb-7e87-4e64-9eb7-894a15a047e2\" (UID: \"bcd634cb-7e87-4e64-9eb7-894a15a047e2\") " Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.219594 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2dmx\" (UniqueName: \"kubernetes.io/projected/01e51b04-eb2c-4102-9a86-d19ca8c566d6-kube-api-access-l2dmx\") pod \"01e51b04-eb2c-4102-9a86-d19ca8c566d6\" (UID: \"01e51b04-eb2c-4102-9a86-d19ca8c566d6\") " Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.219620 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01e51b04-eb2c-4102-9a86-d19ca8c566d6-operator-scripts\") pod \"01e51b04-eb2c-4102-9a86-d19ca8c566d6\" (UID: \"01e51b04-eb2c-4102-9a86-d19ca8c566d6\") " Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.219663 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4zkx\" (UniqueName: \"kubernetes.io/projected/d17c667c-7985-4132-88ee-f0e82ab7d25c-kube-api-access-d4zkx\") pod \"d17c667c-7985-4132-88ee-f0e82ab7d25c\" (UID: \"d17c667c-7985-4132-88ee-f0e82ab7d25c\") " Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.220066 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b96848ac-9038-4172-a36b-fe0889e3f6b6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.220092 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0221e2e-817a-4fa9-9717-2bb622c966f0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.220102 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8nr7\" (UniqueName: \"kubernetes.io/projected/b96848ac-9038-4172-a36b-fe0889e3f6b6-kube-api-access-q8nr7\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.220112 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm475\" (UniqueName: \"kubernetes.io/projected/f0221e2e-817a-4fa9-9717-2bb622c966f0-kube-api-access-wm475\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.220120 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21d83a41-8ec9-424e-be10-fac93f693310-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.220130 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrmz5\" (UniqueName: \"kubernetes.io/projected/21d83a41-8ec9-424e-be10-fac93f693310-kube-api-access-qrmz5\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.220121 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcd634cb-7e87-4e64-9eb7-894a15a047e2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bcd634cb-7e87-4e64-9eb7-894a15a047e2" (UID: "bcd634cb-7e87-4e64-9eb7-894a15a047e2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.220206 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d17c667c-7985-4132-88ee-f0e82ab7d25c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d17c667c-7985-4132-88ee-f0e82ab7d25c" (UID: "d17c667c-7985-4132-88ee-f0e82ab7d25c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.220555 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01e51b04-eb2c-4102-9a86-d19ca8c566d6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "01e51b04-eb2c-4102-9a86-d19ca8c566d6" (UID: "01e51b04-eb2c-4102-9a86-d19ca8c566d6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.223655 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01e51b04-eb2c-4102-9a86-d19ca8c566d6-kube-api-access-l2dmx" (OuterVolumeSpecName: "kube-api-access-l2dmx") pod "01e51b04-eb2c-4102-9a86-d19ca8c566d6" (UID: "01e51b04-eb2c-4102-9a86-d19ca8c566d6"). InnerVolumeSpecName "kube-api-access-l2dmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.223654 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcd634cb-7e87-4e64-9eb7-894a15a047e2-kube-api-access-8gddh" (OuterVolumeSpecName: "kube-api-access-8gddh") pod "bcd634cb-7e87-4e64-9eb7-894a15a047e2" (UID: "bcd634cb-7e87-4e64-9eb7-894a15a047e2"). InnerVolumeSpecName "kube-api-access-8gddh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.223752 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d17c667c-7985-4132-88ee-f0e82ab7d25c-kube-api-access-d4zkx" (OuterVolumeSpecName: "kube-api-access-d4zkx") pod "d17c667c-7985-4132-88ee-f0e82ab7d25c" (UID: "d17c667c-7985-4132-88ee-f0e82ab7d25c"). InnerVolumeSpecName "kube-api-access-d4zkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.251055 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jvlml" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.251184 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jvlml" event={"ID":"01e51b04-eb2c-4102-9a86-d19ca8c566d6","Type":"ContainerDied","Data":"1f2f0b062e687c1bd26b3f53d055f07265df30970fb8fe66eeaf3f1f556f4bac"} Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.251230 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f2f0b062e687c1bd26b3f53d055f07265df30970fb8fe66eeaf3f1f556f4bac" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.254122 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ff71-account-create-update-c62lr" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.254329 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ff71-account-create-update-c62lr" event={"ID":"f0221e2e-817a-4fa9-9717-2bb622c966f0","Type":"ContainerDied","Data":"2a7d286764a6894129b2bac252fd20c0f595581041fba830b4ce6a5d543424e0"} Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.254369 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a7d286764a6894129b2bac252fd20c0f595581041fba830b4ce6a5d543424e0" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.259185 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tmqcq" event={"ID":"bcd634cb-7e87-4e64-9eb7-894a15a047e2","Type":"ContainerDied","Data":"d8951ab77e926361b7507aed9f3d09b9f269b85cf85d246f13a16d178f425de6"} Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.259212 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8951ab77e926361b7507aed9f3d09b9f269b85cf85d246f13a16d178f425de6" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.259265 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tmqcq" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.261656 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tb6zm" event={"ID":"72889b41-41e2-4d58-8e6d-f372f75793b7","Type":"ContainerStarted","Data":"28c16aba11f39ca65575b319d525083a1ad5c9dbebbd5e886ea98a39f8b9c23c"} Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.264247 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-wt46c" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.264258 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-wt46c" event={"ID":"b96848ac-9038-4172-a36b-fe0889e3f6b6","Type":"ContainerDied","Data":"ba0ecbf055638c02d2602b0452ffdbd7a3f11ec1d1428f60e042e08e86c5a2c0"} Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.264290 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba0ecbf055638c02d2602b0452ffdbd7a3f11ec1d1428f60e042e08e86c5a2c0" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.273650 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-17c6-account-create-update-kt4nm" event={"ID":"21d83a41-8ec9-424e-be10-fac93f693310","Type":"ContainerDied","Data":"7a4e6dbf40a4598967e3a698bdc82e692daf7aab7fb17c1ade0ae7bcc69af608"} Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.273691 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-17c6-account-create-update-kt4nm" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.273697 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a4e6dbf40a4598967e3a698bdc82e692daf7aab7fb17c1ade0ae7bcc69af608" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.284722 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c91e-account-create-update-wxssk" event={"ID":"d17c667c-7985-4132-88ee-f0e82ab7d25c","Type":"ContainerDied","Data":"992a617dfdc29e901d824ce9541c9518206ef294344d5bb0912daaeaec9b6246"} Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.284754 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="992a617dfdc29e901d824ce9541c9518206ef294344d5bb0912daaeaec9b6246" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.284813 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c91e-account-create-update-wxssk" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.286558 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-tb6zm" podStartSLOduration=1.965476045 podStartE2EDuration="9.286545966s" podCreationTimestamp="2025-12-01 14:55:03 +0000 UTC" firstStartedPulling="2025-12-01 14:55:04.5825779 +0000 UTC m=+1270.346087503" lastFinishedPulling="2025-12-01 14:55:11.903647821 +0000 UTC m=+1277.667157424" observedRunningTime="2025-12-01 14:55:12.280656806 +0000 UTC m=+1278.044166419" watchObservedRunningTime="2025-12-01 14:55:12.286545966 +0000 UTC m=+1278.050055569" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.321603 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d17c667c-7985-4132-88ee-f0e82ab7d25c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.321637 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gddh\" (UniqueName: \"kubernetes.io/projected/bcd634cb-7e87-4e64-9eb7-894a15a047e2-kube-api-access-8gddh\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.321651 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcd634cb-7e87-4e64-9eb7-894a15a047e2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.321667 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2dmx\" (UniqueName: \"kubernetes.io/projected/01e51b04-eb2c-4102-9a86-d19ca8c566d6-kube-api-access-l2dmx\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.321680 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01e51b04-eb2c-4102-9a86-d19ca8c566d6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:12 crc kubenswrapper[4810]: I1201 14:55:12.321693 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4zkx\" (UniqueName: \"kubernetes.io/projected/d17c667c-7985-4132-88ee-f0e82ab7d25c-kube-api-access-d4zkx\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:13 crc kubenswrapper[4810]: I1201 14:55:13.298768 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerStarted","Data":"0920962c564ec41c800c30de23663892437885ded40e3db2acbc5ff4420710d7"} Dec 01 14:55:13 crc kubenswrapper[4810]: I1201 14:55:13.299159 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerStarted","Data":"7c0fca921a04e4ac4e65934e2a8f95df9f334f4836f26bf901cf250082061924"} Dec 01 14:55:13 crc kubenswrapper[4810]: I1201 14:55:13.299173 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerStarted","Data":"7c19582265bd684bc9384473ffe961c1afdbc4a8cbb8d269b1882eb4dab174dc"} Dec 01 14:55:13 crc kubenswrapper[4810]: I1201 14:55:13.299184 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerStarted","Data":"184d4e1ea01b4832c5828fa2fe97577fb38a0e14ab4fe7d23a5339c79d094ef7"} Dec 01 14:55:13 crc kubenswrapper[4810]: I1201 14:55:13.301323 4810 generic.go:334] "Generic (PLEG): container finished" podID="986d954e-89dd-4878-bbdd-b93e6a0ca155" containerID="8ebd6744e6dfcf18f3444cf3cf726b084641649414a0ce03fbfd64313fce06e2" exitCode=0 Dec 01 14:55:13 crc kubenswrapper[4810]: I1201 14:55:13.302137 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-h4brs" event={"ID":"986d954e-89dd-4878-bbdd-b93e6a0ca155","Type":"ContainerDied","Data":"8ebd6744e6dfcf18f3444cf3cf726b084641649414a0ce03fbfd64313fce06e2"} Dec 01 14:55:14 crc kubenswrapper[4810]: I1201 14:55:14.640328 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-h4brs" Dec 01 14:55:14 crc kubenswrapper[4810]: I1201 14:55:14.763385 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/986d954e-89dd-4878-bbdd-b93e6a0ca155-config-data\") pod \"986d954e-89dd-4878-bbdd-b93e6a0ca155\" (UID: \"986d954e-89dd-4878-bbdd-b93e6a0ca155\") " Dec 01 14:55:14 crc kubenswrapper[4810]: I1201 14:55:14.763454 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whcww\" (UniqueName: \"kubernetes.io/projected/986d954e-89dd-4878-bbdd-b93e6a0ca155-kube-api-access-whcww\") pod \"986d954e-89dd-4878-bbdd-b93e6a0ca155\" (UID: \"986d954e-89dd-4878-bbdd-b93e6a0ca155\") " Dec 01 14:55:14 crc kubenswrapper[4810]: I1201 14:55:14.763583 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/986d954e-89dd-4878-bbdd-b93e6a0ca155-combined-ca-bundle\") pod \"986d954e-89dd-4878-bbdd-b93e6a0ca155\" (UID: \"986d954e-89dd-4878-bbdd-b93e6a0ca155\") " Dec 01 14:55:14 crc kubenswrapper[4810]: I1201 14:55:14.763715 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/986d954e-89dd-4878-bbdd-b93e6a0ca155-db-sync-config-data\") pod \"986d954e-89dd-4878-bbdd-b93e6a0ca155\" (UID: \"986d954e-89dd-4878-bbdd-b93e6a0ca155\") " Dec 01 14:55:14 crc kubenswrapper[4810]: I1201 14:55:14.772714 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/986d954e-89dd-4878-bbdd-b93e6a0ca155-kube-api-access-whcww" (OuterVolumeSpecName: "kube-api-access-whcww") pod "986d954e-89dd-4878-bbdd-b93e6a0ca155" (UID: "986d954e-89dd-4878-bbdd-b93e6a0ca155"). InnerVolumeSpecName "kube-api-access-whcww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:14 crc kubenswrapper[4810]: I1201 14:55:14.772832 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/986d954e-89dd-4878-bbdd-b93e6a0ca155-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "986d954e-89dd-4878-bbdd-b93e6a0ca155" (UID: "986d954e-89dd-4878-bbdd-b93e6a0ca155"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:14 crc kubenswrapper[4810]: I1201 14:55:14.792731 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/986d954e-89dd-4878-bbdd-b93e6a0ca155-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "986d954e-89dd-4878-bbdd-b93e6a0ca155" (UID: "986d954e-89dd-4878-bbdd-b93e6a0ca155"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:14 crc kubenswrapper[4810]: I1201 14:55:14.811972 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/986d954e-89dd-4878-bbdd-b93e6a0ca155-config-data" (OuterVolumeSpecName: "config-data") pod "986d954e-89dd-4878-bbdd-b93e6a0ca155" (UID: "986d954e-89dd-4878-bbdd-b93e6a0ca155"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:14 crc kubenswrapper[4810]: I1201 14:55:14.865414 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whcww\" (UniqueName: \"kubernetes.io/projected/986d954e-89dd-4878-bbdd-b93e6a0ca155-kube-api-access-whcww\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:14 crc kubenswrapper[4810]: I1201 14:55:14.865725 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/986d954e-89dd-4878-bbdd-b93e6a0ca155-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:14 crc kubenswrapper[4810]: I1201 14:55:14.865857 4810 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/986d954e-89dd-4878-bbdd-b93e6a0ca155-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:14 crc kubenswrapper[4810]: I1201 14:55:14.865953 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/986d954e-89dd-4878-bbdd-b93e6a0ca155-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.321690 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerStarted","Data":"18f2b5779be96d5d0ed49bece9ede2da2e81ab667fe59b1dcf9a253cd41860c6"} Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.322205 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerStarted","Data":"4d0480c8c52ca952777f48c3418a4710c30076e9fef37046f82aa4cf60de8531"} Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.322281 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerStarted","Data":"ed49f712640c3ec14f04807a032c7cf926007968ca0dcd16ace0f1dc97a84d7a"} Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.323608 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-h4brs" event={"ID":"986d954e-89dd-4878-bbdd-b93e6a0ca155","Type":"ContainerDied","Data":"4d22615abe344366b599db7410cafca71f369ce6428155ee9a3711714c028493"} Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.323702 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d22615abe344366b599db7410cafca71f369ce6428155ee9a3711714c028493" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.323809 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-h4brs" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.771075 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6df78bdcfc-d8b6v"] Dec 01 14:55:15 crc kubenswrapper[4810]: E1201 14:55:15.771793 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b96848ac-9038-4172-a36b-fe0889e3f6b6" containerName="mariadb-database-create" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.771813 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b96848ac-9038-4172-a36b-fe0889e3f6b6" containerName="mariadb-database-create" Dec 01 14:55:15 crc kubenswrapper[4810]: E1201 14:55:15.771841 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0221e2e-817a-4fa9-9717-2bb622c966f0" containerName="mariadb-account-create-update" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.771850 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0221e2e-817a-4fa9-9717-2bb622c966f0" containerName="mariadb-account-create-update" Dec 01 14:55:15 crc kubenswrapper[4810]: E1201 14:55:15.771863 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcd634cb-7e87-4e64-9eb7-894a15a047e2" containerName="mariadb-database-create" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.771870 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcd634cb-7e87-4e64-9eb7-894a15a047e2" containerName="mariadb-database-create" Dec 01 14:55:15 crc kubenswrapper[4810]: E1201 14:55:15.771893 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01e51b04-eb2c-4102-9a86-d19ca8c566d6" containerName="mariadb-database-create" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.771898 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="01e51b04-eb2c-4102-9a86-d19ca8c566d6" containerName="mariadb-database-create" Dec 01 14:55:15 crc kubenswrapper[4810]: E1201 14:55:15.771911 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21d83a41-8ec9-424e-be10-fac93f693310" containerName="mariadb-account-create-update" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.771917 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="21d83a41-8ec9-424e-be10-fac93f693310" containerName="mariadb-account-create-update" Dec 01 14:55:15 crc kubenswrapper[4810]: E1201 14:55:15.771926 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="986d954e-89dd-4878-bbdd-b93e6a0ca155" containerName="glance-db-sync" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.771931 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="986d954e-89dd-4878-bbdd-b93e6a0ca155" containerName="glance-db-sync" Dec 01 14:55:15 crc kubenswrapper[4810]: E1201 14:55:15.771943 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d17c667c-7985-4132-88ee-f0e82ab7d25c" containerName="mariadb-account-create-update" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.771949 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d17c667c-7985-4132-88ee-f0e82ab7d25c" containerName="mariadb-account-create-update" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.772103 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="21d83a41-8ec9-424e-be10-fac93f693310" containerName="mariadb-account-create-update" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.772126 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0221e2e-817a-4fa9-9717-2bb622c966f0" containerName="mariadb-account-create-update" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.772140 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="01e51b04-eb2c-4102-9a86-d19ca8c566d6" containerName="mariadb-database-create" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.772150 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b96848ac-9038-4172-a36b-fe0889e3f6b6" containerName="mariadb-database-create" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.772159 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d17c667c-7985-4132-88ee-f0e82ab7d25c" containerName="mariadb-account-create-update" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.772177 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcd634cb-7e87-4e64-9eb7-894a15a047e2" containerName="mariadb-database-create" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.772187 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="986d954e-89dd-4878-bbdd-b93e6a0ca155" containerName="glance-db-sync" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.772999 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.792933 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6df78bdcfc-d8b6v"] Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.886904 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-config\") pod \"dnsmasq-dns-6df78bdcfc-d8b6v\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.886966 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-ovsdbserver-nb\") pod \"dnsmasq-dns-6df78bdcfc-d8b6v\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.887031 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-ovsdbserver-sb\") pod \"dnsmasq-dns-6df78bdcfc-d8b6v\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.887060 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkdjj\" (UniqueName: \"kubernetes.io/projected/9339fc60-d07c-4538-903c-8ddbb89b1c23-kube-api-access-bkdjj\") pod \"dnsmasq-dns-6df78bdcfc-d8b6v\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.887151 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-dns-svc\") pod \"dnsmasq-dns-6df78bdcfc-d8b6v\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.988320 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-dns-svc\") pod \"dnsmasq-dns-6df78bdcfc-d8b6v\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.988399 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-config\") pod \"dnsmasq-dns-6df78bdcfc-d8b6v\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.988439 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-ovsdbserver-nb\") pod \"dnsmasq-dns-6df78bdcfc-d8b6v\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.988500 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-ovsdbserver-sb\") pod \"dnsmasq-dns-6df78bdcfc-d8b6v\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.988518 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkdjj\" (UniqueName: \"kubernetes.io/projected/9339fc60-d07c-4538-903c-8ddbb89b1c23-kube-api-access-bkdjj\") pod \"dnsmasq-dns-6df78bdcfc-d8b6v\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.989742 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-config\") pod \"dnsmasq-dns-6df78bdcfc-d8b6v\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.990574 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-ovsdbserver-sb\") pod \"dnsmasq-dns-6df78bdcfc-d8b6v\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.990911 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-dns-svc\") pod \"dnsmasq-dns-6df78bdcfc-d8b6v\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:15 crc kubenswrapper[4810]: I1201 14:55:15.991297 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-ovsdbserver-nb\") pod \"dnsmasq-dns-6df78bdcfc-d8b6v\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.008926 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkdjj\" (UniqueName: \"kubernetes.io/projected/9339fc60-d07c-4538-903c-8ddbb89b1c23-kube-api-access-bkdjj\") pod \"dnsmasq-dns-6df78bdcfc-d8b6v\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.104761 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.358804 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerStarted","Data":"24cce379393b557c9eef88db330dd60281782cf5c7918b19088e9f8504b09390"} Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.359296 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerStarted","Data":"bfb0384344df9db35a4ff5ee4331e251ef264c0c11ab485df0d97f60541193de"} Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.359307 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerStarted","Data":"10f0178b4b19110c943a68bcae88fac92f4d9354eaf37cbb5fce2607db759cb3"} Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.359317 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerStarted","Data":"3eeb13544ba7950f14cd2ed7a1ec0e2b77f5688fbb09bf134178e7165abccdf9"} Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.362128 4810 generic.go:334] "Generic (PLEG): container finished" podID="72889b41-41e2-4d58-8e6d-f372f75793b7" containerID="28c16aba11f39ca65575b319d525083a1ad5c9dbebbd5e886ea98a39f8b9c23c" exitCode=0 Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.362161 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tb6zm" event={"ID":"72889b41-41e2-4d58-8e6d-f372f75793b7","Type":"ContainerDied","Data":"28c16aba11f39ca65575b319d525083a1ad5c9dbebbd5e886ea98a39f8b9c23c"} Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.416994 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=35.816385228 podStartE2EDuration="48.41697173s" podCreationTimestamp="2025-12-01 14:54:28 +0000 UTC" firstStartedPulling="2025-12-01 14:55:01.827182262 +0000 UTC m=+1267.590691855" lastFinishedPulling="2025-12-01 14:55:14.427768754 +0000 UTC m=+1280.191278357" observedRunningTime="2025-12-01 14:55:16.394978423 +0000 UTC m=+1282.158488026" watchObservedRunningTime="2025-12-01 14:55:16.41697173 +0000 UTC m=+1282.180481333" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.548376 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6df78bdcfc-d8b6v"] Dec 01 14:55:16 crc kubenswrapper[4810]: W1201 14:55:16.572327 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9339fc60_d07c_4538_903c_8ddbb89b1c23.slice/crio-18e0838117cdfca1970e942846b629974f41efed7f5b1811b042f9c4a07a038f WatchSource:0}: Error finding container 18e0838117cdfca1970e942846b629974f41efed7f5b1811b042f9c4a07a038f: Status 404 returned error can't find the container with id 18e0838117cdfca1970e942846b629974f41efed7f5b1811b042f9c4a07a038f Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.645198 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6df78bdcfc-d8b6v"] Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.671736 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bfc9d5487-gdtxm"] Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.673019 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.686168 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.687834 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bfc9d5487-gdtxm"] Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.708675 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-config\") pod \"dnsmasq-dns-5bfc9d5487-gdtxm\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.708743 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-ovsdbserver-sb\") pod \"dnsmasq-dns-5bfc9d5487-gdtxm\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.708831 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-dns-swift-storage-0\") pod \"dnsmasq-dns-5bfc9d5487-gdtxm\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.708897 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-dns-svc\") pod \"dnsmasq-dns-5bfc9d5487-gdtxm\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.708938 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-ovsdbserver-nb\") pod \"dnsmasq-dns-5bfc9d5487-gdtxm\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.708994 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf6kz\" (UniqueName: \"kubernetes.io/projected/d248421e-ce7f-4415-a0f2-c7f629d05594-kube-api-access-qf6kz\") pod \"dnsmasq-dns-5bfc9d5487-gdtxm\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.810920 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf6kz\" (UniqueName: \"kubernetes.io/projected/d248421e-ce7f-4415-a0f2-c7f629d05594-kube-api-access-qf6kz\") pod \"dnsmasq-dns-5bfc9d5487-gdtxm\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.810991 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-config\") pod \"dnsmasq-dns-5bfc9d5487-gdtxm\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.811033 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-ovsdbserver-sb\") pod \"dnsmasq-dns-5bfc9d5487-gdtxm\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.811071 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-dns-swift-storage-0\") pod \"dnsmasq-dns-5bfc9d5487-gdtxm\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.811115 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-dns-svc\") pod \"dnsmasq-dns-5bfc9d5487-gdtxm\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.811144 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-ovsdbserver-nb\") pod \"dnsmasq-dns-5bfc9d5487-gdtxm\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.812683 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-config\") pod \"dnsmasq-dns-5bfc9d5487-gdtxm\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.812704 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-ovsdbserver-nb\") pod \"dnsmasq-dns-5bfc9d5487-gdtxm\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.812849 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-dns-swift-storage-0\") pod \"dnsmasq-dns-5bfc9d5487-gdtxm\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.812895 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-dns-svc\") pod \"dnsmasq-dns-5bfc9d5487-gdtxm\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.815037 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-ovsdbserver-sb\") pod \"dnsmasq-dns-5bfc9d5487-gdtxm\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:16 crc kubenswrapper[4810]: I1201 14:55:16.831536 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf6kz\" (UniqueName: \"kubernetes.io/projected/d248421e-ce7f-4415-a0f2-c7f629d05594-kube-api-access-qf6kz\") pod \"dnsmasq-dns-5bfc9d5487-gdtxm\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.031430 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.373185 4810 generic.go:334] "Generic (PLEG): container finished" podID="9339fc60-d07c-4538-903c-8ddbb89b1c23" containerID="22364a2e21bc822a6d690b8b6e17e27da35d99873e5ec0c3ac83a6b1f7d9e58d" exitCode=0 Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.373238 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" event={"ID":"9339fc60-d07c-4538-903c-8ddbb89b1c23","Type":"ContainerDied","Data":"22364a2e21bc822a6d690b8b6e17e27da35d99873e5ec0c3ac83a6b1f7d9e58d"} Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.373308 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" event={"ID":"9339fc60-d07c-4538-903c-8ddbb89b1c23","Type":"ContainerStarted","Data":"18e0838117cdfca1970e942846b629974f41efed7f5b1811b042f9c4a07a038f"} Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.535789 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bfc9d5487-gdtxm"] Dec 01 14:55:17 crc kubenswrapper[4810]: W1201 14:55:17.537233 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd248421e_ce7f_4415_a0f2_c7f629d05594.slice/crio-724f99a7a5ec30672892fe66f3905284bcec53fda161dfe4ae8493f588d5a2b8 WatchSource:0}: Error finding container 724f99a7a5ec30672892fe66f3905284bcec53fda161dfe4ae8493f588d5a2b8: Status 404 returned error can't find the container with id 724f99a7a5ec30672892fe66f3905284bcec53fda161dfe4ae8493f588d5a2b8 Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.747493 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tb6zm" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.778794 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.829983 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-config\") pod \"9339fc60-d07c-4538-903c-8ddbb89b1c23\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.830039 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwxjh\" (UniqueName: \"kubernetes.io/projected/72889b41-41e2-4d58-8e6d-f372f75793b7-kube-api-access-cwxjh\") pod \"72889b41-41e2-4d58-8e6d-f372f75793b7\" (UID: \"72889b41-41e2-4d58-8e6d-f372f75793b7\") " Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.830082 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-ovsdbserver-sb\") pod \"9339fc60-d07c-4538-903c-8ddbb89b1c23\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.830140 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72889b41-41e2-4d58-8e6d-f372f75793b7-config-data\") pod \"72889b41-41e2-4d58-8e6d-f372f75793b7\" (UID: \"72889b41-41e2-4d58-8e6d-f372f75793b7\") " Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.830165 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-dns-svc\") pod \"9339fc60-d07c-4538-903c-8ddbb89b1c23\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.830254 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkdjj\" (UniqueName: \"kubernetes.io/projected/9339fc60-d07c-4538-903c-8ddbb89b1c23-kube-api-access-bkdjj\") pod \"9339fc60-d07c-4538-903c-8ddbb89b1c23\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.830272 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72889b41-41e2-4d58-8e6d-f372f75793b7-combined-ca-bundle\") pod \"72889b41-41e2-4d58-8e6d-f372f75793b7\" (UID: \"72889b41-41e2-4d58-8e6d-f372f75793b7\") " Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.830304 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-ovsdbserver-nb\") pod \"9339fc60-d07c-4538-903c-8ddbb89b1c23\" (UID: \"9339fc60-d07c-4538-903c-8ddbb89b1c23\") " Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.835487 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72889b41-41e2-4d58-8e6d-f372f75793b7-kube-api-access-cwxjh" (OuterVolumeSpecName: "kube-api-access-cwxjh") pod "72889b41-41e2-4d58-8e6d-f372f75793b7" (UID: "72889b41-41e2-4d58-8e6d-f372f75793b7"). InnerVolumeSpecName "kube-api-access-cwxjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.839758 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9339fc60-d07c-4538-903c-8ddbb89b1c23-kube-api-access-bkdjj" (OuterVolumeSpecName: "kube-api-access-bkdjj") pod "9339fc60-d07c-4538-903c-8ddbb89b1c23" (UID: "9339fc60-d07c-4538-903c-8ddbb89b1c23"). InnerVolumeSpecName "kube-api-access-bkdjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.851931 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9339fc60-d07c-4538-903c-8ddbb89b1c23" (UID: "9339fc60-d07c-4538-903c-8ddbb89b1c23"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.856158 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-config" (OuterVolumeSpecName: "config") pod "9339fc60-d07c-4538-903c-8ddbb89b1c23" (UID: "9339fc60-d07c-4538-903c-8ddbb89b1c23"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.862400 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72889b41-41e2-4d58-8e6d-f372f75793b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72889b41-41e2-4d58-8e6d-f372f75793b7" (UID: "72889b41-41e2-4d58-8e6d-f372f75793b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.863031 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9339fc60-d07c-4538-903c-8ddbb89b1c23" (UID: "9339fc60-d07c-4538-903c-8ddbb89b1c23"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.863316 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9339fc60-d07c-4538-903c-8ddbb89b1c23" (UID: "9339fc60-d07c-4538-903c-8ddbb89b1c23"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.880550 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72889b41-41e2-4d58-8e6d-f372f75793b7-config-data" (OuterVolumeSpecName: "config-data") pod "72889b41-41e2-4d58-8e6d-f372f75793b7" (UID: "72889b41-41e2-4d58-8e6d-f372f75793b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.932647 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkdjj\" (UniqueName: \"kubernetes.io/projected/9339fc60-d07c-4538-903c-8ddbb89b1c23-kube-api-access-bkdjj\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.932680 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72889b41-41e2-4d58-8e6d-f372f75793b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.932689 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.932699 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.932707 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwxjh\" (UniqueName: \"kubernetes.io/projected/72889b41-41e2-4d58-8e6d-f372f75793b7-kube-api-access-cwxjh\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.932717 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.932756 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72889b41-41e2-4d58-8e6d-f372f75793b7-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:17 crc kubenswrapper[4810]: I1201 14:55:17.932766 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9339fc60-d07c-4538-903c-8ddbb89b1c23-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.383610 4810 generic.go:334] "Generic (PLEG): container finished" podID="d248421e-ce7f-4415-a0f2-c7f629d05594" containerID="612de26ea67fd116650b4e9a5ae8518b747e2b5e93d817cb0fc611a9875f91d8" exitCode=0 Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.383685 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" event={"ID":"d248421e-ce7f-4415-a0f2-c7f629d05594","Type":"ContainerDied","Data":"612de26ea67fd116650b4e9a5ae8518b747e2b5e93d817cb0fc611a9875f91d8"} Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.383713 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" event={"ID":"d248421e-ce7f-4415-a0f2-c7f629d05594","Type":"ContainerStarted","Data":"724f99a7a5ec30672892fe66f3905284bcec53fda161dfe4ae8493f588d5a2b8"} Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.385905 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" event={"ID":"9339fc60-d07c-4538-903c-8ddbb89b1c23","Type":"ContainerDied","Data":"18e0838117cdfca1970e942846b629974f41efed7f5b1811b042f9c4a07a038f"} Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.385951 4810 scope.go:117] "RemoveContainer" containerID="22364a2e21bc822a6d690b8b6e17e27da35d99873e5ec0c3ac83a6b1f7d9e58d" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.386116 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6df78bdcfc-d8b6v" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.390276 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-tb6zm" event={"ID":"72889b41-41e2-4d58-8e6d-f372f75793b7","Type":"ContainerDied","Data":"c6cb987c8e16d944c4702c8401d3e5668494861a62c20941e781e53f7d749f15"} Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.390378 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6cb987c8e16d944c4702c8401d3e5668494861a62c20941e781e53f7d749f15" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.390515 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-tb6zm" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.547769 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6df78bdcfc-d8b6v"] Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.587131 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6df78bdcfc-d8b6v"] Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.682258 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-kpc4p"] Dec 01 14:55:18 crc kubenswrapper[4810]: E1201 14:55:18.682664 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9339fc60-d07c-4538-903c-8ddbb89b1c23" containerName="init" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.682677 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9339fc60-d07c-4538-903c-8ddbb89b1c23" containerName="init" Dec 01 14:55:18 crc kubenswrapper[4810]: E1201 14:55:18.682693 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72889b41-41e2-4d58-8e6d-f372f75793b7" containerName="keystone-db-sync" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.682699 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="72889b41-41e2-4d58-8e6d-f372f75793b7" containerName="keystone-db-sync" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.682890 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="72889b41-41e2-4d58-8e6d-f372f75793b7" containerName="keystone-db-sync" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.682904 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9339fc60-d07c-4538-903c-8ddbb89b1c23" containerName="init" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.683419 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.687881 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.688150 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.688332 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.688839 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zlbmq" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.691274 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.699099 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bfc9d5487-gdtxm"] Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.708712 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-kpc4p"] Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.739536 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65c6dfc787-x6lgx"] Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.741065 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.746564 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-combined-ca-bundle\") pod \"keystone-bootstrap-kpc4p\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.746624 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-scripts\") pod \"keystone-bootstrap-kpc4p\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.746749 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-fernet-keys\") pod \"keystone-bootstrap-kpc4p\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.746793 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7vxv\" (UniqueName: \"kubernetes.io/projected/fe806b26-bc8c-4e16-ba35-ffa489befc9a-kube-api-access-j7vxv\") pod \"keystone-bootstrap-kpc4p\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.746831 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-config-data\") pod \"keystone-bootstrap-kpc4p\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.746992 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-credential-keys\") pod \"keystone-bootstrap-kpc4p\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.784825 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65c6dfc787-x6lgx"] Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.848241 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7vxv\" (UniqueName: \"kubernetes.io/projected/fe806b26-bc8c-4e16-ba35-ffa489befc9a-kube-api-access-j7vxv\") pod \"keystone-bootstrap-kpc4p\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.848305 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-config-data\") pod \"keystone-bootstrap-kpc4p\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.848355 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt924\" (UniqueName: \"kubernetes.io/projected/6ae6544f-7c47-4cf9-a00c-24459c58edc9-kube-api-access-dt924\") pod \"dnsmasq-dns-65c6dfc787-x6lgx\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.848377 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-dns-svc\") pod \"dnsmasq-dns-65c6dfc787-x6lgx\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.848395 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-ovsdbserver-nb\") pod \"dnsmasq-dns-65c6dfc787-x6lgx\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.848419 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-credential-keys\") pod \"keystone-bootstrap-kpc4p\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.848449 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-combined-ca-bundle\") pod \"keystone-bootstrap-kpc4p\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.848500 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-ovsdbserver-sb\") pod \"dnsmasq-dns-65c6dfc787-x6lgx\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.848521 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-scripts\") pod \"keystone-bootstrap-kpc4p\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.848599 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-config\") pod \"dnsmasq-dns-65c6dfc787-x6lgx\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.848661 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-dns-swift-storage-0\") pod \"dnsmasq-dns-65c6dfc787-x6lgx\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.848701 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-fernet-keys\") pod \"keystone-bootstrap-kpc4p\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.854867 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-credential-keys\") pod \"keystone-bootstrap-kpc4p\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.855207 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-fernet-keys\") pod \"keystone-bootstrap-kpc4p\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.855728 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-combined-ca-bundle\") pod \"keystone-bootstrap-kpc4p\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.861521 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-scripts\") pod \"keystone-bootstrap-kpc4p\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.864819 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-config-data\") pod \"keystone-bootstrap-kpc4p\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.869883 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-vlw7c"] Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.870989 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.876849 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-8q2gm" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.876902 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.877098 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.879360 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7vxv\" (UniqueName: \"kubernetes.io/projected/fe806b26-bc8c-4e16-ba35-ffa489befc9a-kube-api-access-j7vxv\") pod \"keystone-bootstrap-kpc4p\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.885612 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vlw7c"] Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.951232 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-dns-swift-storage-0\") pod \"dnsmasq-dns-65c6dfc787-x6lgx\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.950404 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-dns-swift-storage-0\") pod \"dnsmasq-dns-65c6dfc787-x6lgx\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.951314 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-scripts\") pod \"cinder-db-sync-vlw7c\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.951347 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-config-data\") pod \"cinder-db-sync-vlw7c\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.951365 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/09861728-4533-4ddf-a32d-3cb1b70fc325-etc-machine-id\") pod \"cinder-db-sync-vlw7c\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.951392 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngs2b\" (UniqueName: \"kubernetes.io/projected/09861728-4533-4ddf-a32d-3cb1b70fc325-kube-api-access-ngs2b\") pod \"cinder-db-sync-vlw7c\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.951415 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-combined-ca-bundle\") pod \"cinder-db-sync-vlw7c\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.951443 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt924\" (UniqueName: \"kubernetes.io/projected/6ae6544f-7c47-4cf9-a00c-24459c58edc9-kube-api-access-dt924\") pod \"dnsmasq-dns-65c6dfc787-x6lgx\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.951462 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-dns-svc\") pod \"dnsmasq-dns-65c6dfc787-x6lgx\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.951507 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-ovsdbserver-nb\") pod \"dnsmasq-dns-65c6dfc787-x6lgx\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.951543 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-ovsdbserver-sb\") pod \"dnsmasq-dns-65c6dfc787-x6lgx\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.951590 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-config\") pod \"dnsmasq-dns-65c6dfc787-x6lgx\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.951609 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-db-sync-config-data\") pod \"cinder-db-sync-vlw7c\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.952518 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-dns-svc\") pod \"dnsmasq-dns-65c6dfc787-x6lgx\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.953301 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-ovsdbserver-nb\") pod \"dnsmasq-dns-65c6dfc787-x6lgx\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.953894 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-ovsdbserver-sb\") pod \"dnsmasq-dns-65c6dfc787-x6lgx\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.954297 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-config\") pod \"dnsmasq-dns-65c6dfc787-x6lgx\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.980639 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.982561 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.988792 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.989039 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 14:55:18 crc kubenswrapper[4810]: I1201 14:55:18.996500 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt924\" (UniqueName: \"kubernetes.io/projected/6ae6544f-7c47-4cf9-a00c-24459c58edc9-kube-api-access-dt924\") pod \"dnsmasq-dns-65c6dfc787-x6lgx\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.005859 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.011636 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.013196 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-b6vj7"] Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.014249 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-b6vj7" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.028784 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.028912 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.029063 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-22xgp" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.041495 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-b6vj7"] Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.053021 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/073cab97-c80f-4905-92df-82b2537ed262-run-httpd\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.053117 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/077fd1b8-6283-4a1b-b159-437897afa556-combined-ca-bundle\") pod \"neutron-db-sync-b6vj7\" (UID: \"077fd1b8-6283-4a1b-b159-437897afa556\") " pod="openstack/neutron-db-sync-b6vj7" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.053158 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-config-data\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.053185 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-db-sync-config-data\") pod \"cinder-db-sync-vlw7c\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.053248 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-scripts\") pod \"cinder-db-sync-vlw7c\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.053268 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfp2z\" (UniqueName: \"kubernetes.io/projected/073cab97-c80f-4905-92df-82b2537ed262-kube-api-access-vfp2z\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.053307 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/077fd1b8-6283-4a1b-b159-437897afa556-config\") pod \"neutron-db-sync-b6vj7\" (UID: \"077fd1b8-6283-4a1b-b159-437897afa556\") " pod="openstack/neutron-db-sync-b6vj7" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.053326 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-config-data\") pod \"cinder-db-sync-vlw7c\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.053344 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/09861728-4533-4ddf-a32d-3cb1b70fc325-etc-machine-id\") pod \"cinder-db-sync-vlw7c\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.053377 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/073cab97-c80f-4905-92df-82b2537ed262-log-httpd\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.053401 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.053419 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngs2b\" (UniqueName: \"kubernetes.io/projected/09861728-4533-4ddf-a32d-3cb1b70fc325-kube-api-access-ngs2b\") pod \"cinder-db-sync-vlw7c\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.053454 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ln4z\" (UniqueName: \"kubernetes.io/projected/077fd1b8-6283-4a1b-b159-437897afa556-kube-api-access-7ln4z\") pod \"neutron-db-sync-b6vj7\" (UID: \"077fd1b8-6283-4a1b-b159-437897afa556\") " pod="openstack/neutron-db-sync-b6vj7" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.053489 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-combined-ca-bundle\") pod \"cinder-db-sync-vlw7c\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.053538 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.053559 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-scripts\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.058456 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/09861728-4533-4ddf-a32d-3cb1b70fc325-etc-machine-id\") pod \"cinder-db-sync-vlw7c\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.058535 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-zwn5q"] Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.059116 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-scripts\") pod \"cinder-db-sync-vlw7c\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.060281 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zwn5q" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.060729 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-db-sync-config-data\") pod \"cinder-db-sync-vlw7c\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.063630 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.063919 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-ppms8" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.067867 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-combined-ca-bundle\") pod \"cinder-db-sync-vlw7c\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.071922 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-config-data\") pod \"cinder-db-sync-vlw7c\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.072285 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.096610 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngs2b\" (UniqueName: \"kubernetes.io/projected/09861728-4533-4ddf-a32d-3cb1b70fc325-kube-api-access-ngs2b\") pod \"cinder-db-sync-vlw7c\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.156768 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-zwn5q"] Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.157817 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/073cab97-c80f-4905-92df-82b2537ed262-run-httpd\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.157866 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/077fd1b8-6283-4a1b-b159-437897afa556-combined-ca-bundle\") pod \"neutron-db-sync-b6vj7\" (UID: \"077fd1b8-6283-4a1b-b159-437897afa556\") " pod="openstack/neutron-db-sync-b6vj7" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.157899 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef385349-2803-45cf-9307-8657305dc6cf-combined-ca-bundle\") pod \"barbican-db-sync-zwn5q\" (UID: \"ef385349-2803-45cf-9307-8657305dc6cf\") " pod="openstack/barbican-db-sync-zwn5q" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.157946 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-config-data\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.158020 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef385349-2803-45cf-9307-8657305dc6cf-db-sync-config-data\") pod \"barbican-db-sync-zwn5q\" (UID: \"ef385349-2803-45cf-9307-8657305dc6cf\") " pod="openstack/barbican-db-sync-zwn5q" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.158053 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfp2z\" (UniqueName: \"kubernetes.io/projected/073cab97-c80f-4905-92df-82b2537ed262-kube-api-access-vfp2z\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.158084 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/077fd1b8-6283-4a1b-b159-437897afa556-config\") pod \"neutron-db-sync-b6vj7\" (UID: \"077fd1b8-6283-4a1b-b159-437897afa556\") " pod="openstack/neutron-db-sync-b6vj7" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.158120 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/073cab97-c80f-4905-92df-82b2537ed262-log-httpd\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.158151 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcv9v\" (UniqueName: \"kubernetes.io/projected/ef385349-2803-45cf-9307-8657305dc6cf-kube-api-access-rcv9v\") pod \"barbican-db-sync-zwn5q\" (UID: \"ef385349-2803-45cf-9307-8657305dc6cf\") " pod="openstack/barbican-db-sync-zwn5q" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.158177 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.158214 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ln4z\" (UniqueName: \"kubernetes.io/projected/077fd1b8-6283-4a1b-b159-437897afa556-kube-api-access-7ln4z\") pod \"neutron-db-sync-b6vj7\" (UID: \"077fd1b8-6283-4a1b-b159-437897afa556\") " pod="openstack/neutron-db-sync-b6vj7" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.158269 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.158307 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-scripts\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.158902 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/073cab97-c80f-4905-92df-82b2537ed262-run-httpd\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.169025 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/073cab97-c80f-4905-92df-82b2537ed262-log-httpd\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.173223 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/077fd1b8-6283-4a1b-b159-437897afa556-config\") pod \"neutron-db-sync-b6vj7\" (UID: \"077fd1b8-6283-4a1b-b159-437897afa556\") " pod="openstack/neutron-db-sync-b6vj7" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.173337 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-scripts\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.173782 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/077fd1b8-6283-4a1b-b159-437897afa556-combined-ca-bundle\") pod \"neutron-db-sync-b6vj7\" (UID: \"077fd1b8-6283-4a1b-b159-437897afa556\") " pod="openstack/neutron-db-sync-b6vj7" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.185945 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-config-data\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.190359 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65c6dfc787-x6lgx"] Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.190399 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfp2z\" (UniqueName: \"kubernetes.io/projected/073cab97-c80f-4905-92df-82b2537ed262-kube-api-access-vfp2z\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.199053 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.201071 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-l4gdq"] Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.202043 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ln4z\" (UniqueName: \"kubernetes.io/projected/077fd1b8-6283-4a1b-b159-437897afa556-kube-api-access-7ln4z\") pod \"neutron-db-sync-b6vj7\" (UID: \"077fd1b8-6283-4a1b-b159-437897afa556\") " pod="openstack/neutron-db-sync-b6vj7" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.202228 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.202818 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.205328 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-hgvgs" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.207096 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.213259 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.219095 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f645789c-lshtm"] Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.220401 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.238668 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-l4gdq"] Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.255160 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f645789c-lshtm"] Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.266722 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51938ba3-de39-414b-9777-fdc90cc14a97-scripts\") pod \"placement-db-sync-l4gdq\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.286379 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.291183 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef385349-2803-45cf-9307-8657305dc6cf-db-sync-config-data\") pod \"barbican-db-sync-zwn5q\" (UID: \"ef385349-2803-45cf-9307-8657305dc6cf\") " pod="openstack/barbican-db-sync-zwn5q" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.291235 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-dns-svc\") pod \"dnsmasq-dns-55f645789c-lshtm\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.291297 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51938ba3-de39-414b-9777-fdc90cc14a97-combined-ca-bundle\") pod \"placement-db-sync-l4gdq\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.291351 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-dns-swift-storage-0\") pod \"dnsmasq-dns-55f645789c-lshtm\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.291382 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcv9v\" (UniqueName: \"kubernetes.io/projected/ef385349-2803-45cf-9307-8657305dc6cf-kube-api-access-rcv9v\") pod \"barbican-db-sync-zwn5q\" (UID: \"ef385349-2803-45cf-9307-8657305dc6cf\") " pod="openstack/barbican-db-sync-zwn5q" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.291422 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwl68\" (UniqueName: \"kubernetes.io/projected/59e1bba9-e680-453b-bb4d-9f7f202551eb-kube-api-access-hwl68\") pod \"dnsmasq-dns-55f645789c-lshtm\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.291442 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-ovsdbserver-sb\") pod \"dnsmasq-dns-55f645789c-lshtm\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.291465 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51938ba3-de39-414b-9777-fdc90cc14a97-config-data\") pod \"placement-db-sync-l4gdq\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.291691 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-ovsdbserver-nb\") pod \"dnsmasq-dns-55f645789c-lshtm\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.291746 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51938ba3-de39-414b-9777-fdc90cc14a97-logs\") pod \"placement-db-sync-l4gdq\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.292083 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnzbj\" (UniqueName: \"kubernetes.io/projected/51938ba3-de39-414b-9777-fdc90cc14a97-kube-api-access-bnzbj\") pod \"placement-db-sync-l4gdq\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.292218 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef385349-2803-45cf-9307-8657305dc6cf-combined-ca-bundle\") pod \"barbican-db-sync-zwn5q\" (UID: \"ef385349-2803-45cf-9307-8657305dc6cf\") " pod="openstack/barbican-db-sync-zwn5q" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.292299 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-config\") pod \"dnsmasq-dns-55f645789c-lshtm\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.300006 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef385349-2803-45cf-9307-8657305dc6cf-combined-ca-bundle\") pod \"barbican-db-sync-zwn5q\" (UID: \"ef385349-2803-45cf-9307-8657305dc6cf\") " pod="openstack/barbican-db-sync-zwn5q" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.307398 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.313233 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef385349-2803-45cf-9307-8657305dc6cf-db-sync-config-data\") pod \"barbican-db-sync-zwn5q\" (UID: \"ef385349-2803-45cf-9307-8657305dc6cf\") " pod="openstack/barbican-db-sync-zwn5q" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.342073 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcv9v\" (UniqueName: \"kubernetes.io/projected/ef385349-2803-45cf-9307-8657305dc6cf-kube-api-access-rcv9v\") pod \"barbican-db-sync-zwn5q\" (UID: \"ef385349-2803-45cf-9307-8657305dc6cf\") " pod="openstack/barbican-db-sync-zwn5q" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.345089 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-b6vj7" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.397074 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-config\") pod \"dnsmasq-dns-55f645789c-lshtm\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.397164 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51938ba3-de39-414b-9777-fdc90cc14a97-scripts\") pod \"placement-db-sync-l4gdq\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.397347 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-dns-svc\") pod \"dnsmasq-dns-55f645789c-lshtm\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.397425 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51938ba3-de39-414b-9777-fdc90cc14a97-combined-ca-bundle\") pod \"placement-db-sync-l4gdq\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.397542 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-dns-swift-storage-0\") pod \"dnsmasq-dns-55f645789c-lshtm\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.397690 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwl68\" (UniqueName: \"kubernetes.io/projected/59e1bba9-e680-453b-bb4d-9f7f202551eb-kube-api-access-hwl68\") pod \"dnsmasq-dns-55f645789c-lshtm\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.397786 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-ovsdbserver-sb\") pod \"dnsmasq-dns-55f645789c-lshtm\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.397809 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51938ba3-de39-414b-9777-fdc90cc14a97-config-data\") pod \"placement-db-sync-l4gdq\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.397878 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-ovsdbserver-nb\") pod \"dnsmasq-dns-55f645789c-lshtm\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.397907 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51938ba3-de39-414b-9777-fdc90cc14a97-logs\") pod \"placement-db-sync-l4gdq\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.397945 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnzbj\" (UniqueName: \"kubernetes.io/projected/51938ba3-de39-414b-9777-fdc90cc14a97-kube-api-access-bnzbj\") pod \"placement-db-sync-l4gdq\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.399607 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-dns-swift-storage-0\") pod \"dnsmasq-dns-55f645789c-lshtm\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.400182 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-config\") pod \"dnsmasq-dns-55f645789c-lshtm\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.401724 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-ovsdbserver-sb\") pod \"dnsmasq-dns-55f645789c-lshtm\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.403187 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-ovsdbserver-nb\") pod \"dnsmasq-dns-55f645789c-lshtm\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.403681 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51938ba3-de39-414b-9777-fdc90cc14a97-scripts\") pod \"placement-db-sync-l4gdq\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.403820 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51938ba3-de39-414b-9777-fdc90cc14a97-logs\") pod \"placement-db-sync-l4gdq\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.404054 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-dns-svc\") pod \"dnsmasq-dns-55f645789c-lshtm\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.409207 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51938ba3-de39-414b-9777-fdc90cc14a97-combined-ca-bundle\") pod \"placement-db-sync-l4gdq\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.413042 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zwn5q" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.418199 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51938ba3-de39-414b-9777-fdc90cc14a97-config-data\") pod \"placement-db-sync-l4gdq\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.420120 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnzbj\" (UniqueName: \"kubernetes.io/projected/51938ba3-de39-414b-9777-fdc90cc14a97-kube-api-access-bnzbj\") pod \"placement-db-sync-l4gdq\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.420319 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwl68\" (UniqueName: \"kubernetes.io/projected/59e1bba9-e680-453b-bb4d-9f7f202551eb-kube-api-access-hwl68\") pod \"dnsmasq-dns-55f645789c-lshtm\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.433663 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" podUID="d248421e-ce7f-4415-a0f2-c7f629d05594" containerName="dnsmasq-dns" containerID="cri-o://a560c35cfbd7517cb3e8612c0a71daadc1e3a456fa9aa10b671e5ae3248585b0" gracePeriod=10 Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.433805 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" event={"ID":"d248421e-ce7f-4415-a0f2-c7f629d05594","Type":"ContainerStarted","Data":"a560c35cfbd7517cb3e8612c0a71daadc1e3a456fa9aa10b671e5ae3248585b0"} Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.434200 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.472222 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" podStartSLOduration=3.472198386 podStartE2EDuration="3.472198386s" podCreationTimestamp="2025-12-01 14:55:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:55:19.46164651 +0000 UTC m=+1285.225156113" watchObservedRunningTime="2025-12-01 14:55:19.472198386 +0000 UTC m=+1285.235707989" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.533755 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.542923 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.684380 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65c6dfc787-x6lgx"] Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.701442 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-kpc4p"] Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.799683 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.801080 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.806448 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.806688 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-6kdsr" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.809177 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.811191 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.837057 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.913519 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.913966 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.914015 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-config-data\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.914066 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgzb4\" (UniqueName: \"kubernetes.io/projected/535446c9-3136-49fe-9b67-354c2ae9be55-kube-api-access-fgzb4\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.914102 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-scripts\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.914122 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/535446c9-3136-49fe-9b67-354c2ae9be55-logs\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.914148 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/535446c9-3136-49fe-9b67-354c2ae9be55-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.914171 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.929554 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.931903 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.936340 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.942320 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 14:55:19 crc kubenswrapper[4810]: I1201 14:55:19.960330 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.017532 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.017600 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.017670 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.017719 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgzb4\" (UniqueName: \"kubernetes.io/projected/535446c9-3136-49fe-9b67-354c2ae9be55-kube-api-access-fgzb4\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.017792 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.017841 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-scripts\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.017972 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.018103 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/535446c9-3136-49fe-9b67-354c2ae9be55-logs\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.018242 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/535446c9-3136-49fe-9b67-354c2ae9be55-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.018344 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.018575 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.018640 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.018806 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/535446c9-3136-49fe-9b67-354c2ae9be55-logs\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.018839 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzxsm\" (UniqueName: \"kubernetes.io/projected/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-kube-api-access-lzxsm\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.018877 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/535446c9-3136-49fe-9b67-354c2ae9be55-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.018901 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-config-data\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.019004 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.019050 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-logs\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.019621 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.025095 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.026113 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.027229 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-config-data\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.028495 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-scripts\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.037514 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgzb4\" (UniqueName: \"kubernetes.io/projected/535446c9-3136-49fe-9b67-354c2ae9be55-kube-api-access-fgzb4\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.045498 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.122153 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzxsm\" (UniqueName: \"kubernetes.io/projected/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-kube-api-access-lzxsm\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.122237 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.122261 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-logs\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.122286 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.122315 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.122351 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.122385 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.122425 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.122834 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.129370 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.129797 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-logs\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.130083 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.131438 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.135021 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.143435 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.148324 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzxsm\" (UniqueName: \"kubernetes.io/projected/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-kube-api-access-lzxsm\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.152852 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.197455 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.250809 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.276821 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.327317 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-ovsdbserver-nb\") pod \"d248421e-ce7f-4415-a0f2-c7f629d05594\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.327384 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-dns-svc\") pod \"d248421e-ce7f-4415-a0f2-c7f629d05594\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.327405 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-ovsdbserver-sb\") pod \"d248421e-ce7f-4415-a0f2-c7f629d05594\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.327433 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qf6kz\" (UniqueName: \"kubernetes.io/projected/d248421e-ce7f-4415-a0f2-c7f629d05594-kube-api-access-qf6kz\") pod \"d248421e-ce7f-4415-a0f2-c7f629d05594\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.327478 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-config\") pod \"d248421e-ce7f-4415-a0f2-c7f629d05594\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.327517 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-dns-swift-storage-0\") pod \"d248421e-ce7f-4415-a0f2-c7f629d05594\" (UID: \"d248421e-ce7f-4415-a0f2-c7f629d05594\") " Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.335087 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d248421e-ce7f-4415-a0f2-c7f629d05594-kube-api-access-qf6kz" (OuterVolumeSpecName: "kube-api-access-qf6kz") pod "d248421e-ce7f-4415-a0f2-c7f629d05594" (UID: "d248421e-ce7f-4415-a0f2-c7f629d05594"). InnerVolumeSpecName "kube-api-access-qf6kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.388593 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d248421e-ce7f-4415-a0f2-c7f629d05594" (UID: "d248421e-ce7f-4415-a0f2-c7f629d05594"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.417100 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-b6vj7"] Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.428196 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vlw7c"] Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.429151 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.429172 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qf6kz\" (UniqueName: \"kubernetes.io/projected/d248421e-ce7f-4415-a0f2-c7f629d05594-kube-api-access-qf6kz\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.435099 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d248421e-ce7f-4415-a0f2-c7f629d05594" (UID: "d248421e-ce7f-4415-a0f2-c7f629d05594"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:20 crc kubenswrapper[4810]: W1201 14:55:20.450078 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod077fd1b8_6283_4a1b_b159_437897afa556.slice/crio-b12b30254bfa5200bd9b31b78fe1e9a77d98d3731e8d0cebc9eb97ba2600041f WatchSource:0}: Error finding container b12b30254bfa5200bd9b31b78fe1e9a77d98d3731e8d0cebc9eb97ba2600041f: Status 404 returned error can't find the container with id b12b30254bfa5200bd9b31b78fe1e9a77d98d3731e8d0cebc9eb97ba2600041f Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.458588 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-zwn5q"] Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.470987 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d248421e-ce7f-4415-a0f2-c7f629d05594" (UID: "d248421e-ce7f-4415-a0f2-c7f629d05594"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.473162 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.480321 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-l4gdq"] Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.485410 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-config" (OuterVolumeSpecName: "config") pod "d248421e-ce7f-4415-a0f2-c7f629d05594" (UID: "d248421e-ce7f-4415-a0f2-c7f629d05594"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:20 crc kubenswrapper[4810]: W1201 14:55:20.515904 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod073cab97_c80f_4905_92df_82b2537ed262.slice/crio-3630a694a7b3b317b18e38d6907ec6c0512366643411e5d75feb5115e7d5f499 WatchSource:0}: Error finding container 3630a694a7b3b317b18e38d6907ec6c0512366643411e5d75feb5115e7d5f499: Status 404 returned error can't find the container with id 3630a694a7b3b317b18e38d6907ec6c0512366643411e5d75feb5115e7d5f499 Dec 01 14:55:20 crc kubenswrapper[4810]: W1201 14:55:20.519652 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef385349_2803_45cf_9307_8657305dc6cf.slice/crio-f0ad77614f91f3187631da4101355ee3cc53c89228e5f70fc1166fad86a7f416 WatchSource:0}: Error finding container f0ad77614f91f3187631da4101355ee3cc53c89228e5f70fc1166fad86a7f416: Status 404 returned error can't find the container with id f0ad77614f91f3187631da4101355ee3cc53c89228e5f70fc1166fad86a7f416 Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.521253 4810 generic.go:334] "Generic (PLEG): container finished" podID="d248421e-ce7f-4415-a0f2-c7f629d05594" containerID="a560c35cfbd7517cb3e8612c0a71daadc1e3a456fa9aa10b671e5ae3248585b0" exitCode=0 Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.521343 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.522131 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9339fc60-d07c-4538-903c-8ddbb89b1c23" path="/var/lib/kubelet/pods/9339fc60-d07c-4538-903c-8ddbb89b1c23/volumes" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.530308 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.530335 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.530344 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:20 crc kubenswrapper[4810]: W1201 14:55:20.532260 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51938ba3_de39_414b_9777_fdc90cc14a97.slice/crio-b490863d702616d194a2068a92afead078ac76cd4000ddeaa65b07572b8e0ccc WatchSource:0}: Error finding container b490863d702616d194a2068a92afead078ac76cd4000ddeaa65b07572b8e0ccc: Status 404 returned error can't find the container with id b490863d702616d194a2068a92afead078ac76cd4000ddeaa65b07572b8e0ccc Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.535066 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d248421e-ce7f-4415-a0f2-c7f629d05594" (UID: "d248421e-ce7f-4415-a0f2-c7f629d05594"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.536346 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-kpc4p" podStartSLOduration=2.536330504 podStartE2EDuration="2.536330504s" podCreationTimestamp="2025-12-01 14:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:55:20.533117107 +0000 UTC m=+1286.296626710" watchObservedRunningTime="2025-12-01 14:55:20.536330504 +0000 UTC m=+1286.299840107" Dec 01 14:55:20 crc kubenswrapper[4810]: W1201 14:55:20.569154 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59e1bba9_e680_453b_bb4d_9f7f202551eb.slice/crio-8ea87e13c69593db6ceeac2aa9dbf2944c96f815b6b0ad3fbafd9cc86bbed27a WatchSource:0}: Error finding container 8ea87e13c69593db6ceeac2aa9dbf2944c96f815b6b0ad3fbafd9cc86bbed27a: Status 404 returned error can't find the container with id 8ea87e13c69593db6ceeac2aa9dbf2944c96f815b6b0ad3fbafd9cc86bbed27a Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.649558 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d248421e-ce7f-4415-a0f2-c7f629d05594-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.748904 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kpc4p" event={"ID":"fe806b26-bc8c-4e16-ba35-ffa489befc9a","Type":"ContainerStarted","Data":"e16c83319d033bff450700a0e11c3bd1ea8dc45e44e6d30b69b942faeaa93a51"} Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.749284 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kpc4p" event={"ID":"fe806b26-bc8c-4e16-ba35-ffa489befc9a","Type":"ContainerStarted","Data":"aadf2421e58563ff05f5600236881798461ea190f039cd8e3555c85bf980bbdc"} Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.749299 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" event={"ID":"d248421e-ce7f-4415-a0f2-c7f629d05594","Type":"ContainerDied","Data":"a560c35cfbd7517cb3e8612c0a71daadc1e3a456fa9aa10b671e5ae3248585b0"} Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.749318 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bfc9d5487-gdtxm" event={"ID":"d248421e-ce7f-4415-a0f2-c7f629d05594","Type":"ContainerDied","Data":"724f99a7a5ec30672892fe66f3905284bcec53fda161dfe4ae8493f588d5a2b8"} Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.749334 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f645789c-lshtm"] Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.749356 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" event={"ID":"6ae6544f-7c47-4cf9-a00c-24459c58edc9","Type":"ContainerStarted","Data":"cbdd980252382bc45eecdcc58e2d334cb4b90494cb2a003990f4d2216e549a12"} Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.749369 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" event={"ID":"6ae6544f-7c47-4cf9-a00c-24459c58edc9","Type":"ContainerStarted","Data":"769cebce6e5ccc3ba99f449909bd8977f155db22d2722aae1f71dbbc4709e96c"} Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.749392 4810 scope.go:117] "RemoveContainer" containerID="a560c35cfbd7517cb3e8612c0a71daadc1e3a456fa9aa10b671e5ae3248585b0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.841826 4810 scope.go:117] "RemoveContainer" containerID="612de26ea67fd116650b4e9a5ae8518b747e2b5e93d817cb0fc611a9875f91d8" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.864664 4810 scope.go:117] "RemoveContainer" containerID="a560c35cfbd7517cb3e8612c0a71daadc1e3a456fa9aa10b671e5ae3248585b0" Dec 01 14:55:20 crc kubenswrapper[4810]: E1201 14:55:20.865251 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a560c35cfbd7517cb3e8612c0a71daadc1e3a456fa9aa10b671e5ae3248585b0\": container with ID starting with a560c35cfbd7517cb3e8612c0a71daadc1e3a456fa9aa10b671e5ae3248585b0 not found: ID does not exist" containerID="a560c35cfbd7517cb3e8612c0a71daadc1e3a456fa9aa10b671e5ae3248585b0" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.865310 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a560c35cfbd7517cb3e8612c0a71daadc1e3a456fa9aa10b671e5ae3248585b0"} err="failed to get container status \"a560c35cfbd7517cb3e8612c0a71daadc1e3a456fa9aa10b671e5ae3248585b0\": rpc error: code = NotFound desc = could not find container \"a560c35cfbd7517cb3e8612c0a71daadc1e3a456fa9aa10b671e5ae3248585b0\": container with ID starting with a560c35cfbd7517cb3e8612c0a71daadc1e3a456fa9aa10b671e5ae3248585b0 not found: ID does not exist" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.865344 4810 scope.go:117] "RemoveContainer" containerID="612de26ea67fd116650b4e9a5ae8518b747e2b5e93d817cb0fc611a9875f91d8" Dec 01 14:55:20 crc kubenswrapper[4810]: E1201 14:55:20.865746 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"612de26ea67fd116650b4e9a5ae8518b747e2b5e93d817cb0fc611a9875f91d8\": container with ID starting with 612de26ea67fd116650b4e9a5ae8518b747e2b5e93d817cb0fc611a9875f91d8 not found: ID does not exist" containerID="612de26ea67fd116650b4e9a5ae8518b747e2b5e93d817cb0fc611a9875f91d8" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.865778 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"612de26ea67fd116650b4e9a5ae8518b747e2b5e93d817cb0fc611a9875f91d8"} err="failed to get container status \"612de26ea67fd116650b4e9a5ae8518b747e2b5e93d817cb0fc611a9875f91d8\": rpc error: code = NotFound desc = could not find container \"612de26ea67fd116650b4e9a5ae8518b747e2b5e93d817cb0fc611a9875f91d8\": container with ID starting with 612de26ea67fd116650b4e9a5ae8518b747e2b5e93d817cb0fc611a9875f91d8 not found: ID does not exist" Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.902101 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bfc9d5487-gdtxm"] Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.921326 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bfc9d5487-gdtxm"] Dec 01 14:55:20 crc kubenswrapper[4810]: I1201 14:55:20.973402 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.094053 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.112545 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.159354 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-ovsdbserver-sb\") pod \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.159559 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-dns-svc\") pod \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.159608 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-config\") pod \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.159636 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-ovsdbserver-nb\") pod \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.159801 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-dns-swift-storage-0\") pod \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.159844 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dt924\" (UniqueName: \"kubernetes.io/projected/6ae6544f-7c47-4cf9-a00c-24459c58edc9-kube-api-access-dt924\") pod \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\" (UID: \"6ae6544f-7c47-4cf9-a00c-24459c58edc9\") " Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.166358 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ae6544f-7c47-4cf9-a00c-24459c58edc9-kube-api-access-dt924" (OuterVolumeSpecName: "kube-api-access-dt924") pod "6ae6544f-7c47-4cf9-a00c-24459c58edc9" (UID: "6ae6544f-7c47-4cf9-a00c-24459c58edc9"). InnerVolumeSpecName "kube-api-access-dt924". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.186946 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6ae6544f-7c47-4cf9-a00c-24459c58edc9" (UID: "6ae6544f-7c47-4cf9-a00c-24459c58edc9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.189728 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6ae6544f-7c47-4cf9-a00c-24459c58edc9" (UID: "6ae6544f-7c47-4cf9-a00c-24459c58edc9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.191664 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6ae6544f-7c47-4cf9-a00c-24459c58edc9" (UID: "6ae6544f-7c47-4cf9-a00c-24459c58edc9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.200023 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6ae6544f-7c47-4cf9-a00c-24459c58edc9" (UID: "6ae6544f-7c47-4cf9-a00c-24459c58edc9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.210049 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-config" (OuterVolumeSpecName: "config") pod "6ae6544f-7c47-4cf9-a00c-24459c58edc9" (UID: "6ae6544f-7c47-4cf9-a00c-24459c58edc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.262706 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.262770 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.262786 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.262800 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.262812 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dt924\" (UniqueName: \"kubernetes.io/projected/6ae6544f-7c47-4cf9-a00c-24459c58edc9-kube-api-access-dt924\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.262824 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ae6544f-7c47-4cf9-a00c-24459c58edc9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.541577 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817","Type":"ContainerStarted","Data":"218d754d49428c5dc3fc08b51c28db879a8bbd6ba74efa62323c5b7b0c3472c4"} Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.567555 4810 generic.go:334] "Generic (PLEG): container finished" podID="6ae6544f-7c47-4cf9-a00c-24459c58edc9" containerID="cbdd980252382bc45eecdcc58e2d334cb4b90494cb2a003990f4d2216e549a12" exitCode=0 Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.567715 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.568148 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" event={"ID":"6ae6544f-7c47-4cf9-a00c-24459c58edc9","Type":"ContainerDied","Data":"cbdd980252382bc45eecdcc58e2d334cb4b90494cb2a003990f4d2216e549a12"} Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.568185 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c6dfc787-x6lgx" event={"ID":"6ae6544f-7c47-4cf9-a00c-24459c58edc9","Type":"ContainerDied","Data":"769cebce6e5ccc3ba99f449909bd8977f155db22d2722aae1f71dbbc4709e96c"} Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.568205 4810 scope.go:117] "RemoveContainer" containerID="cbdd980252382bc45eecdcc58e2d334cb4b90494cb2a003990f4d2216e549a12" Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.585572 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-b6vj7" event={"ID":"077fd1b8-6283-4a1b-b159-437897afa556","Type":"ContainerStarted","Data":"d9a642dd5e5161d1c3547e4c6fe4ee983dcde299f557d63adfeffa793f5b30c3"} Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.585616 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-b6vj7" event={"ID":"077fd1b8-6283-4a1b-b159-437897afa556","Type":"ContainerStarted","Data":"b12b30254bfa5200bd9b31b78fe1e9a77d98d3731e8d0cebc9eb97ba2600041f"} Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.587737 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"073cab97-c80f-4905-92df-82b2537ed262","Type":"ContainerStarted","Data":"3630a694a7b3b317b18e38d6907ec6c0512366643411e5d75feb5115e7d5f499"} Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.590627 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-l4gdq" event={"ID":"51938ba3-de39-414b-9777-fdc90cc14a97","Type":"ContainerStarted","Data":"b490863d702616d194a2068a92afead078ac76cd4000ddeaa65b07572b8e0ccc"} Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.593100 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vlw7c" event={"ID":"09861728-4533-4ddf-a32d-3cb1b70fc325","Type":"ContainerStarted","Data":"58a06685d2101791f5bea371d664d3b80d561952217772fa304642ba18a8665d"} Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.594263 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zwn5q" event={"ID":"ef385349-2803-45cf-9307-8657305dc6cf","Type":"ContainerStarted","Data":"f0ad77614f91f3187631da4101355ee3cc53c89228e5f70fc1166fad86a7f416"} Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.596282 4810 generic.go:334] "Generic (PLEG): container finished" podID="59e1bba9-e680-453b-bb4d-9f7f202551eb" containerID="655da5f31bd45ec5aae36c4aec6d16ce77eb5e9fdbbd4c24b84f240e54627620" exitCode=0 Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.596322 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f645789c-lshtm" event={"ID":"59e1bba9-e680-453b-bb4d-9f7f202551eb","Type":"ContainerDied","Data":"655da5f31bd45ec5aae36c4aec6d16ce77eb5e9fdbbd4c24b84f240e54627620"} Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.596338 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f645789c-lshtm" event={"ID":"59e1bba9-e680-453b-bb4d-9f7f202551eb","Type":"ContainerStarted","Data":"8ea87e13c69593db6ceeac2aa9dbf2944c96f815b6b0ad3fbafd9cc86bbed27a"} Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.600418 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"535446c9-3136-49fe-9b67-354c2ae9be55","Type":"ContainerStarted","Data":"e058ff9711b586501f29dfd0fe2ab3928b9ab48eb1b2f3e66fcc74cce1856937"} Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.634361 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-b6vj7" podStartSLOduration=3.634337803 podStartE2EDuration="3.634337803s" podCreationTimestamp="2025-12-01 14:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:55:21.620301521 +0000 UTC m=+1287.383811124" watchObservedRunningTime="2025-12-01 14:55:21.634337803 +0000 UTC m=+1287.397847406" Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.637127 4810 scope.go:117] "RemoveContainer" containerID="cbdd980252382bc45eecdcc58e2d334cb4b90494cb2a003990f4d2216e549a12" Dec 01 14:55:21 crc kubenswrapper[4810]: E1201 14:55:21.638142 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbdd980252382bc45eecdcc58e2d334cb4b90494cb2a003990f4d2216e549a12\": container with ID starting with cbdd980252382bc45eecdcc58e2d334cb4b90494cb2a003990f4d2216e549a12 not found: ID does not exist" containerID="cbdd980252382bc45eecdcc58e2d334cb4b90494cb2a003990f4d2216e549a12" Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.638193 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbdd980252382bc45eecdcc58e2d334cb4b90494cb2a003990f4d2216e549a12"} err="failed to get container status \"cbdd980252382bc45eecdcc58e2d334cb4b90494cb2a003990f4d2216e549a12\": rpc error: code = NotFound desc = could not find container \"cbdd980252382bc45eecdcc58e2d334cb4b90494cb2a003990f4d2216e549a12\": container with ID starting with cbdd980252382bc45eecdcc58e2d334cb4b90494cb2a003990f4d2216e549a12 not found: ID does not exist" Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.792042 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65c6dfc787-x6lgx"] Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.801836 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65c6dfc787-x6lgx"] Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.813045 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.880212 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 14:55:21 crc kubenswrapper[4810]: I1201 14:55:21.914544 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:55:22 crc kubenswrapper[4810]: I1201 14:55:22.520182 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ae6544f-7c47-4cf9-a00c-24459c58edc9" path="/var/lib/kubelet/pods/6ae6544f-7c47-4cf9-a00c-24459c58edc9/volumes" Dec 01 14:55:22 crc kubenswrapper[4810]: I1201 14:55:22.521585 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d248421e-ce7f-4415-a0f2-c7f629d05594" path="/var/lib/kubelet/pods/d248421e-ce7f-4415-a0f2-c7f629d05594/volumes" Dec 01 14:55:22 crc kubenswrapper[4810]: I1201 14:55:22.633181 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817","Type":"ContainerStarted","Data":"8cb6af2c29480513ff7981f7f9f023b7a4bc21f738b0305082e69934bfea96c3"} Dec 01 14:55:22 crc kubenswrapper[4810]: I1201 14:55:22.645802 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f645789c-lshtm" event={"ID":"59e1bba9-e680-453b-bb4d-9f7f202551eb","Type":"ContainerStarted","Data":"e0deec04a2b1f5902cc569561667e0c9f848a4cc54bd1aa77303e35ec109d9b5"} Dec 01 14:55:22 crc kubenswrapper[4810]: I1201 14:55:22.645962 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:22 crc kubenswrapper[4810]: I1201 14:55:22.651666 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"535446c9-3136-49fe-9b67-354c2ae9be55","Type":"ContainerStarted","Data":"731da099a1d7826afc9690b79541d3c9d0a0f2dab30d6ef44afcede09a666794"} Dec 01 14:55:22 crc kubenswrapper[4810]: I1201 14:55:22.690494 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f645789c-lshtm" podStartSLOduration=3.690459343 podStartE2EDuration="3.690459343s" podCreationTimestamp="2025-12-01 14:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:55:22.668537527 +0000 UTC m=+1288.432047150" watchObservedRunningTime="2025-12-01 14:55:22.690459343 +0000 UTC m=+1288.453968946" Dec 01 14:55:23 crc kubenswrapper[4810]: I1201 14:55:23.665866 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"535446c9-3136-49fe-9b67-354c2ae9be55","Type":"ContainerStarted","Data":"ab13c39f4c9a577a70986369c13cabc96a35e93d0e337c0f4eb272c273de5d09"} Dec 01 14:55:23 crc kubenswrapper[4810]: I1201 14:55:23.665954 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="535446c9-3136-49fe-9b67-354c2ae9be55" containerName="glance-log" containerID="cri-o://731da099a1d7826afc9690b79541d3c9d0a0f2dab30d6ef44afcede09a666794" gracePeriod=30 Dec 01 14:55:23 crc kubenswrapper[4810]: I1201 14:55:23.666058 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="535446c9-3136-49fe-9b67-354c2ae9be55" containerName="glance-httpd" containerID="cri-o://ab13c39f4c9a577a70986369c13cabc96a35e93d0e337c0f4eb272c273de5d09" gracePeriod=30 Dec 01 14:55:23 crc kubenswrapper[4810]: I1201 14:55:23.669082 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" containerName="glance-log" containerID="cri-o://8cb6af2c29480513ff7981f7f9f023b7a4bc21f738b0305082e69934bfea96c3" gracePeriod=30 Dec 01 14:55:23 crc kubenswrapper[4810]: I1201 14:55:23.669240 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817","Type":"ContainerStarted","Data":"1b63d6fe729ca9dcc82b75dcf4c9bb4d47b7cd2d50638513fa29253a470549d4"} Dec 01 14:55:23 crc kubenswrapper[4810]: I1201 14:55:23.669295 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" containerName="glance-httpd" containerID="cri-o://1b63d6fe729ca9dcc82b75dcf4c9bb4d47b7cd2d50638513fa29253a470549d4" gracePeriod=30 Dec 01 14:55:23 crc kubenswrapper[4810]: I1201 14:55:23.698912 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.698890117 podStartE2EDuration="5.698890117s" podCreationTimestamp="2025-12-01 14:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:55:23.688624708 +0000 UTC m=+1289.452134311" watchObservedRunningTime="2025-12-01 14:55:23.698890117 +0000 UTC m=+1289.462399720" Dec 01 14:55:23 crc kubenswrapper[4810]: I1201 14:55:23.709303 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.70928334 podStartE2EDuration="5.70928334s" podCreationTimestamp="2025-12-01 14:55:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:55:23.709243069 +0000 UTC m=+1289.472752682" watchObservedRunningTime="2025-12-01 14:55:23.70928334 +0000 UTC m=+1289.472792943" Dec 01 14:55:24 crc kubenswrapper[4810]: I1201 14:55:24.690320 4810 generic.go:334] "Generic (PLEG): container finished" podID="fe806b26-bc8c-4e16-ba35-ffa489befc9a" containerID="e16c83319d033bff450700a0e11c3bd1ea8dc45e44e6d30b69b942faeaa93a51" exitCode=0 Dec 01 14:55:24 crc kubenswrapper[4810]: I1201 14:55:24.690689 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kpc4p" event={"ID":"fe806b26-bc8c-4e16-ba35-ffa489befc9a","Type":"ContainerDied","Data":"e16c83319d033bff450700a0e11c3bd1ea8dc45e44e6d30b69b942faeaa93a51"} Dec 01 14:55:24 crc kubenswrapper[4810]: I1201 14:55:24.700233 4810 generic.go:334] "Generic (PLEG): container finished" podID="535446c9-3136-49fe-9b67-354c2ae9be55" containerID="ab13c39f4c9a577a70986369c13cabc96a35e93d0e337c0f4eb272c273de5d09" exitCode=0 Dec 01 14:55:24 crc kubenswrapper[4810]: I1201 14:55:24.700281 4810 generic.go:334] "Generic (PLEG): container finished" podID="535446c9-3136-49fe-9b67-354c2ae9be55" containerID="731da099a1d7826afc9690b79541d3c9d0a0f2dab30d6ef44afcede09a666794" exitCode=143 Dec 01 14:55:24 crc kubenswrapper[4810]: I1201 14:55:24.700332 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"535446c9-3136-49fe-9b67-354c2ae9be55","Type":"ContainerDied","Data":"ab13c39f4c9a577a70986369c13cabc96a35e93d0e337c0f4eb272c273de5d09"} Dec 01 14:55:24 crc kubenswrapper[4810]: I1201 14:55:24.700359 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"535446c9-3136-49fe-9b67-354c2ae9be55","Type":"ContainerDied","Data":"731da099a1d7826afc9690b79541d3c9d0a0f2dab30d6ef44afcede09a666794"} Dec 01 14:55:24 crc kubenswrapper[4810]: I1201 14:55:24.704227 4810 generic.go:334] "Generic (PLEG): container finished" podID="ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" containerID="1b63d6fe729ca9dcc82b75dcf4c9bb4d47b7cd2d50638513fa29253a470549d4" exitCode=0 Dec 01 14:55:24 crc kubenswrapper[4810]: I1201 14:55:24.704253 4810 generic.go:334] "Generic (PLEG): container finished" podID="ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" containerID="8cb6af2c29480513ff7981f7f9f023b7a4bc21f738b0305082e69934bfea96c3" exitCode=143 Dec 01 14:55:24 crc kubenswrapper[4810]: I1201 14:55:24.704274 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817","Type":"ContainerDied","Data":"1b63d6fe729ca9dcc82b75dcf4c9bb4d47b7cd2d50638513fa29253a470549d4"} Dec 01 14:55:24 crc kubenswrapper[4810]: I1201 14:55:24.704306 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817","Type":"ContainerDied","Data":"8cb6af2c29480513ff7981f7f9f023b7a4bc21f738b0305082e69934bfea96c3"} Dec 01 14:55:29 crc kubenswrapper[4810]: I1201 14:55:29.544686 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:55:29 crc kubenswrapper[4810]: I1201 14:55:29.644956 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f6d79597f-wvf92"] Dec 01 14:55:29 crc kubenswrapper[4810]: I1201 14:55:29.645440 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" podUID="1aec0ffb-75a0-43e6-a721-4010c6ee0f81" containerName="dnsmasq-dns" containerID="cri-o://d0da55680ea404f4e37e5d9ef3fc938bbec4754bf125f24758a733873b008d4e" gracePeriod=10 Dec 01 14:55:30 crc kubenswrapper[4810]: I1201 14:55:30.816951 4810 generic.go:334] "Generic (PLEG): container finished" podID="1aec0ffb-75a0-43e6-a721-4010c6ee0f81" containerID="d0da55680ea404f4e37e5d9ef3fc938bbec4754bf125f24758a733873b008d4e" exitCode=0 Dec 01 14:55:30 crc kubenswrapper[4810]: I1201 14:55:30.817219 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" event={"ID":"1aec0ffb-75a0-43e6-a721-4010c6ee0f81","Type":"ContainerDied","Data":"d0da55680ea404f4e37e5d9ef3fc938bbec4754bf125f24758a733873b008d4e"} Dec 01 14:55:38 crc kubenswrapper[4810]: I1201 14:55:38.411670 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" podUID="1aec0ffb-75a0-43e6-a721-4010c6ee0f81" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: i/o timeout" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.117830 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.158790 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-config-data\") pod \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.158847 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7vxv\" (UniqueName: \"kubernetes.io/projected/fe806b26-bc8c-4e16-ba35-ffa489befc9a-kube-api-access-j7vxv\") pod \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.158892 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-combined-ca-bundle\") pod \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.158918 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-fernet-keys\") pod \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.158958 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-scripts\") pod \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.158981 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-credential-keys\") pod \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\" (UID: \"fe806b26-bc8c-4e16-ba35-ffa489befc9a\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.165165 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "fe806b26-bc8c-4e16-ba35-ffa489befc9a" (UID: "fe806b26-bc8c-4e16-ba35-ffa489befc9a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.165194 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "fe806b26-bc8c-4e16-ba35-ffa489befc9a" (UID: "fe806b26-bc8c-4e16-ba35-ffa489befc9a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.176059 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-scripts" (OuterVolumeSpecName: "scripts") pod "fe806b26-bc8c-4e16-ba35-ffa489befc9a" (UID: "fe806b26-bc8c-4e16-ba35-ffa489befc9a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.187248 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe806b26-bc8c-4e16-ba35-ffa489befc9a" (UID: "fe806b26-bc8c-4e16-ba35-ffa489befc9a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.190539 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-config-data" (OuterVolumeSpecName: "config-data") pod "fe806b26-bc8c-4e16-ba35-ffa489befc9a" (UID: "fe806b26-bc8c-4e16-ba35-ffa489befc9a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.192015 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe806b26-bc8c-4e16-ba35-ffa489befc9a-kube-api-access-j7vxv" (OuterVolumeSpecName: "kube-api-access-j7vxv") pod "fe806b26-bc8c-4e16-ba35-ffa489befc9a" (UID: "fe806b26-bc8c-4e16-ba35-ffa489befc9a"). InnerVolumeSpecName "kube-api-access-j7vxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.261732 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.261770 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7vxv\" (UniqueName: \"kubernetes.io/projected/fe806b26-bc8c-4e16-ba35-ffa489befc9a-kube-api-access-j7vxv\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.261781 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.261790 4810 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.261799 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.261807 4810 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fe806b26-bc8c-4e16-ba35-ffa489befc9a-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: E1201 14:55:40.409702 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:2051e26a441f1ce22aeb8daa0137559d89bded994db8141c11dd580ae6d07a23" Dec 01 14:55:40 crc kubenswrapper[4810]: E1201 14:55:40.409916 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:2051e26a441f1ce22aeb8daa0137559d89bded994db8141c11dd580ae6d07a23,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5bdh89h686h56dh5d8hbch5b4h85h67bh6fh676hbchbdh6chc7hd9h557h65fhf5h5dfh84h9bh555h5d6h5c7h697h687hdfh54h67ch85h5dfq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vfp2z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(073cab97-c80f-4905-92df-82b2537ed262): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.466886 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.474349 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.480564 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575021 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzxsm\" (UniqueName: \"kubernetes.io/projected/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-kube-api-access-lzxsm\") pod \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575091 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-ovsdbserver-sb\") pod \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575131 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575179 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-config-data\") pod \"535446c9-3136-49fe-9b67-354c2ae9be55\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575211 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-config-data\") pod \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575238 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/535446c9-3136-49fe-9b67-354c2ae9be55-httpd-run\") pod \"535446c9-3136-49fe-9b67-354c2ae9be55\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575285 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-combined-ca-bundle\") pod \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575327 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgzb4\" (UniqueName: \"kubernetes.io/projected/535446c9-3136-49fe-9b67-354c2ae9be55-kube-api-access-fgzb4\") pod \"535446c9-3136-49fe-9b67-354c2ae9be55\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575347 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-dns-svc\") pod \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575405 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-public-tls-certs\") pod \"535446c9-3136-49fe-9b67-354c2ae9be55\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575446 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/535446c9-3136-49fe-9b67-354c2ae9be55-logs\") pod \"535446c9-3136-49fe-9b67-354c2ae9be55\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575512 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-logs\") pod \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575544 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-scripts\") pod \"535446c9-3136-49fe-9b67-354c2ae9be55\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575569 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42ztn\" (UniqueName: \"kubernetes.io/projected/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-kube-api-access-42ztn\") pod \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575600 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-httpd-run\") pod \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575625 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-scripts\") pod \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575665 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-combined-ca-bundle\") pod \"535446c9-3136-49fe-9b67-354c2ae9be55\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575720 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-internal-tls-certs\") pod \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\" (UID: \"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575746 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-ovsdbserver-nb\") pod \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575773 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-config\") pod \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\" (UID: \"1aec0ffb-75a0-43e6-a721-4010c6ee0f81\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.575800 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"535446c9-3136-49fe-9b67-354c2ae9be55\" (UID: \"535446c9-3136-49fe-9b67-354c2ae9be55\") " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.581665 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-kube-api-access-lzxsm" (OuterVolumeSpecName: "kube-api-access-lzxsm") pod "ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" (UID: "ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817"). InnerVolumeSpecName "kube-api-access-lzxsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.583268 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/535446c9-3136-49fe-9b67-354c2ae9be55-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "535446c9-3136-49fe-9b67-354c2ae9be55" (UID: "535446c9-3136-49fe-9b67-354c2ae9be55"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.584229 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-scripts" (OuterVolumeSpecName: "scripts") pod "ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" (UID: "ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.585163 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" (UID: "ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.587974 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" (UID: "ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.591683 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "535446c9-3136-49fe-9b67-354c2ae9be55" (UID: "535446c9-3136-49fe-9b67-354c2ae9be55"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.595182 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/535446c9-3136-49fe-9b67-354c2ae9be55-logs" (OuterVolumeSpecName: "logs") pod "535446c9-3136-49fe-9b67-354c2ae9be55" (UID: "535446c9-3136-49fe-9b67-354c2ae9be55"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.595238 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-logs" (OuterVolumeSpecName: "logs") pod "ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" (UID: "ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.597633 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-scripts" (OuterVolumeSpecName: "scripts") pod "535446c9-3136-49fe-9b67-354c2ae9be55" (UID: "535446c9-3136-49fe-9b67-354c2ae9be55"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.598615 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/535446c9-3136-49fe-9b67-354c2ae9be55-kube-api-access-fgzb4" (OuterVolumeSpecName: "kube-api-access-fgzb4") pod "535446c9-3136-49fe-9b67-354c2ae9be55" (UID: "535446c9-3136-49fe-9b67-354c2ae9be55"). InnerVolumeSpecName "kube-api-access-fgzb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.608645 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-kube-api-access-42ztn" (OuterVolumeSpecName: "kube-api-access-42ztn") pod "1aec0ffb-75a0-43e6-a721-4010c6ee0f81" (UID: "1aec0ffb-75a0-43e6-a721-4010c6ee0f81"). InnerVolumeSpecName "kube-api-access-42ztn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.614511 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "535446c9-3136-49fe-9b67-354c2ae9be55" (UID: "535446c9-3136-49fe-9b67-354c2ae9be55"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.641758 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" (UID: "ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.645865 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1aec0ffb-75a0-43e6-a721-4010c6ee0f81" (UID: "1aec0ffb-75a0-43e6-a721-4010c6ee0f81"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.653039 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" (UID: "ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.653133 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-config-data" (OuterVolumeSpecName: "config-data") pod "535446c9-3136-49fe-9b67-354c2ae9be55" (UID: "535446c9-3136-49fe-9b67-354c2ae9be55"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.659341 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1aec0ffb-75a0-43e6-a721-4010c6ee0f81" (UID: "1aec0ffb-75a0-43e6-a721-4010c6ee0f81"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.663293 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-config-data" (OuterVolumeSpecName: "config-data") pod "ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" (UID: "ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.665702 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-config" (OuterVolumeSpecName: "config") pod "1aec0ffb-75a0-43e6-a721-4010c6ee0f81" (UID: "1aec0ffb-75a0-43e6-a721-4010c6ee0f81"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.673056 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1aec0ffb-75a0-43e6-a721-4010c6ee0f81" (UID: "1aec0ffb-75a0-43e6-a721-4010c6ee0f81"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677431 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677483 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677517 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677531 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzxsm\" (UniqueName: \"kubernetes.io/projected/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-kube-api-access-lzxsm\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677546 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677564 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677616 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677629 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677640 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/535446c9-3136-49fe-9b67-354c2ae9be55-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677651 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677662 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677673 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgzb4\" (UniqueName: \"kubernetes.io/projected/535446c9-3136-49fe-9b67-354c2ae9be55-kube-api-access-fgzb4\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677683 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/535446c9-3136-49fe-9b67-354c2ae9be55-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677694 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677707 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42ztn\" (UniqueName: \"kubernetes.io/projected/1aec0ffb-75a0-43e6-a721-4010c6ee0f81-kube-api-access-42ztn\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677719 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677729 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677740 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677750 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.677761 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.680870 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "535446c9-3136-49fe-9b67-354c2ae9be55" (UID: "535446c9-3136-49fe-9b67-354c2ae9be55"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.694143 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.696890 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.778935 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.778983 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.778993 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/535446c9-3136-49fe-9b67-354c2ae9be55-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.921696 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"535446c9-3136-49fe-9b67-354c2ae9be55","Type":"ContainerDied","Data":"e058ff9711b586501f29dfd0fe2ab3928b9ab48eb1b2f3e66fcc74cce1856937"} Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.921711 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.921744 4810 scope.go:117] "RemoveContainer" containerID="ab13c39f4c9a577a70986369c13cabc96a35e93d0e337c0f4eb272c273de5d09" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.925784 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.925778 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817","Type":"ContainerDied","Data":"218d754d49428c5dc3fc08b51c28db879a8bbd6ba74efa62323c5b7b0c3472c4"} Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.929716 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" event={"ID":"1aec0ffb-75a0-43e6-a721-4010c6ee0f81","Type":"ContainerDied","Data":"1689fa3a24b341f5c3eca87491a11d142e9373c69b475f908affd61a79a767b6"} Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.929788 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.933891 4810 generic.go:334] "Generic (PLEG): container finished" podID="077fd1b8-6283-4a1b-b159-437897afa556" containerID="d9a642dd5e5161d1c3547e4c6fe4ee983dcde299f557d63adfeffa793f5b30c3" exitCode=0 Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.933963 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-b6vj7" event={"ID":"077fd1b8-6283-4a1b-b159-437897afa556","Type":"ContainerDied","Data":"d9a642dd5e5161d1c3547e4c6fe4ee983dcde299f557d63adfeffa793f5b30c3"} Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.935917 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kpc4p" event={"ID":"fe806b26-bc8c-4e16-ba35-ffa489befc9a","Type":"ContainerDied","Data":"aadf2421e58563ff05f5600236881798461ea190f039cd8e3555c85bf980bbdc"} Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.935961 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aadf2421e58563ff05f5600236881798461ea190f039cd8e3555c85bf980bbdc" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.935963 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kpc4p" Dec 01 14:55:40 crc kubenswrapper[4810]: I1201 14:55:40.985975 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.007537 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.015846 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f6d79597f-wvf92"] Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.026935 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f6d79597f-wvf92"] Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.033701 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 14:55:41 crc kubenswrapper[4810]: E1201 14:55:41.034097 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aec0ffb-75a0-43e6-a721-4010c6ee0f81" containerName="init" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.034110 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aec0ffb-75a0-43e6-a721-4010c6ee0f81" containerName="init" Dec 01 14:55:41 crc kubenswrapper[4810]: E1201 14:55:41.034121 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="535446c9-3136-49fe-9b67-354c2ae9be55" containerName="glance-httpd" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.034127 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="535446c9-3136-49fe-9b67-354c2ae9be55" containerName="glance-httpd" Dec 01 14:55:41 crc kubenswrapper[4810]: E1201 14:55:41.034145 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d248421e-ce7f-4415-a0f2-c7f629d05594" containerName="dnsmasq-dns" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.034150 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d248421e-ce7f-4415-a0f2-c7f629d05594" containerName="dnsmasq-dns" Dec 01 14:55:41 crc kubenswrapper[4810]: E1201 14:55:41.034161 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe806b26-bc8c-4e16-ba35-ffa489befc9a" containerName="keystone-bootstrap" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.034167 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe806b26-bc8c-4e16-ba35-ffa489befc9a" containerName="keystone-bootstrap" Dec 01 14:55:41 crc kubenswrapper[4810]: E1201 14:55:41.034175 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aec0ffb-75a0-43e6-a721-4010c6ee0f81" containerName="dnsmasq-dns" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.034181 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aec0ffb-75a0-43e6-a721-4010c6ee0f81" containerName="dnsmasq-dns" Dec 01 14:55:41 crc kubenswrapper[4810]: E1201 14:55:41.034198 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d248421e-ce7f-4415-a0f2-c7f629d05594" containerName="init" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.034204 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d248421e-ce7f-4415-a0f2-c7f629d05594" containerName="init" Dec 01 14:55:41 crc kubenswrapper[4810]: E1201 14:55:41.034215 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ae6544f-7c47-4cf9-a00c-24459c58edc9" containerName="init" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.034221 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ae6544f-7c47-4cf9-a00c-24459c58edc9" containerName="init" Dec 01 14:55:41 crc kubenswrapper[4810]: E1201 14:55:41.034228 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="535446c9-3136-49fe-9b67-354c2ae9be55" containerName="glance-log" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.034233 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="535446c9-3136-49fe-9b67-354c2ae9be55" containerName="glance-log" Dec 01 14:55:41 crc kubenswrapper[4810]: E1201 14:55:41.034243 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" containerName="glance-log" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.034248 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" containerName="glance-log" Dec 01 14:55:41 crc kubenswrapper[4810]: E1201 14:55:41.034260 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" containerName="glance-httpd" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.034266 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" containerName="glance-httpd" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.034502 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" containerName="glance-httpd" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.034519 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ae6544f-7c47-4cf9-a00c-24459c58edc9" containerName="init" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.034537 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="535446c9-3136-49fe-9b67-354c2ae9be55" containerName="glance-log" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.034547 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d248421e-ce7f-4415-a0f2-c7f629d05594" containerName="dnsmasq-dns" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.034557 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="535446c9-3136-49fe-9b67-354c2ae9be55" containerName="glance-httpd" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.034566 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe806b26-bc8c-4e16-ba35-ffa489befc9a" containerName="keystone-bootstrap" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.034573 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aec0ffb-75a0-43e6-a721-4010c6ee0f81" containerName="dnsmasq-dns" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.034583 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" containerName="glance-log" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.035593 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.040873 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-6kdsr" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.041179 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.041769 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.041929 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.042066 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.049549 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.057932 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.065308 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.066941 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.075171 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.075903 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.089772 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.187437 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.187516 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swxpr\" (UniqueName: \"kubernetes.io/projected/3039d59e-4377-411d-98e7-9fd84a2d93b6-kube-api-access-swxpr\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.187555 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-scripts\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.187579 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d18c05bc-3e20-4305-b8bf-d45e6ff46308-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.187621 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.187674 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.187713 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.187740 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3039d59e-4377-411d-98e7-9fd84a2d93b6-logs\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.187763 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.187783 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.187802 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d18c05bc-3e20-4305-b8bf-d45e6ff46308-logs\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.187816 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3039d59e-4377-411d-98e7-9fd84a2d93b6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.187833 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.187855 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-config-data\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.187876 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs2g9\" (UniqueName: \"kubernetes.io/projected/d18c05bc-3e20-4305-b8bf-d45e6ff46308-kube-api-access-rs2g9\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.187895 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.289973 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs2g9\" (UniqueName: \"kubernetes.io/projected/d18c05bc-3e20-4305-b8bf-d45e6ff46308-kube-api-access-rs2g9\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.290047 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.290090 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.290127 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swxpr\" (UniqueName: \"kubernetes.io/projected/3039d59e-4377-411d-98e7-9fd84a2d93b6-kube-api-access-swxpr\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.290171 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-scripts\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.290221 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d18c05bc-3e20-4305-b8bf-d45e6ff46308-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.290282 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.290320 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.290356 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.290373 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3039d59e-4377-411d-98e7-9fd84a2d93b6-logs\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.290388 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.290406 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.290422 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d18c05bc-3e20-4305-b8bf-d45e6ff46308-logs\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.290436 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3039d59e-4377-411d-98e7-9fd84a2d93b6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.290455 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.290492 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-config-data\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.296877 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d18c05bc-3e20-4305-b8bf-d45e6ff46308-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.301534 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.302011 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.308240 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.310367 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d18c05bc-3e20-4305-b8bf-d45e6ff46308-logs\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.310864 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3039d59e-4377-411d-98e7-9fd84a2d93b6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.311376 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3039d59e-4377-411d-98e7-9fd84a2d93b6-logs\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.311727 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.312030 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-config-data\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.320869 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-scripts\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.332155 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.345667 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-kpc4p"] Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.356101 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-kpc4p"] Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.359363 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.359908 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.363048 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.375904 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swxpr\" (UniqueName: \"kubernetes.io/projected/3039d59e-4377-411d-98e7-9fd84a2d93b6-kube-api-access-swxpr\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.399833 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs2g9\" (UniqueName: \"kubernetes.io/projected/d18c05bc-3e20-4305-b8bf-d45e6ff46308-kube-api-access-rs2g9\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.407682 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.463036 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.468927 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-v9d27"] Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.470069 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.479779 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-v9d27"] Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.480363 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.480737 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.483007 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.483233 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zlbmq" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.484222 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.484422 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.599068 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74n65\" (UniqueName: \"kubernetes.io/projected/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-kube-api-access-74n65\") pod \"keystone-bootstrap-v9d27\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.599424 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-fernet-keys\") pod \"keystone-bootstrap-v9d27\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.599449 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-scripts\") pod \"keystone-bootstrap-v9d27\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.599499 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-combined-ca-bundle\") pod \"keystone-bootstrap-v9d27\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.599551 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-credential-keys\") pod \"keystone-bootstrap-v9d27\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.599571 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-config-data\") pod \"keystone-bootstrap-v9d27\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.700500 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74n65\" (UniqueName: \"kubernetes.io/projected/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-kube-api-access-74n65\") pod \"keystone-bootstrap-v9d27\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.700563 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-fernet-keys\") pod \"keystone-bootstrap-v9d27\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.700585 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-scripts\") pod \"keystone-bootstrap-v9d27\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.700612 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-combined-ca-bundle\") pod \"keystone-bootstrap-v9d27\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.700664 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-credential-keys\") pod \"keystone-bootstrap-v9d27\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.700685 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-config-data\") pod \"keystone-bootstrap-v9d27\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.709989 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-config-data\") pod \"keystone-bootstrap-v9d27\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.710278 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-scripts\") pod \"keystone-bootstrap-v9d27\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.710591 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-credential-keys\") pod \"keystone-bootstrap-v9d27\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.711594 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-fernet-keys\") pod \"keystone-bootstrap-v9d27\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.716833 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-combined-ca-bundle\") pod \"keystone-bootstrap-v9d27\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.718327 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74n65\" (UniqueName: \"kubernetes.io/projected/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-kube-api-access-74n65\") pod \"keystone-bootstrap-v9d27\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.736949 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 14:55:41 crc kubenswrapper[4810]: I1201 14:55:41.823754 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.318167 4810 scope.go:117] "RemoveContainer" containerID="731da099a1d7826afc9690b79541d3c9d0a0f2dab30d6ef44afcede09a666794" Dec 01 14:55:42 crc kubenswrapper[4810]: E1201 14:55:42.352083 4810 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b5266c9a26766fce2b92f95dff52d362a760f7baf1474cdcb33bd68570e096c0" Dec 01 14:55:42 crc kubenswrapper[4810]: E1201 14:55:42.352243 4810 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b5266c9a26766fce2b92f95dff52d362a760f7baf1474cdcb33bd68570e096c0,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ngs2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-vlw7c_openstack(09861728-4533-4ddf-a32d-3cb1b70fc325): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 14:55:42 crc kubenswrapper[4810]: E1201 14:55:42.353425 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-vlw7c" podUID="09861728-4533-4ddf-a32d-3cb1b70fc325" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.509110 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1aec0ffb-75a0-43e6-a721-4010c6ee0f81" path="/var/lib/kubelet/pods/1aec0ffb-75a0-43e6-a721-4010c6ee0f81/volumes" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.512312 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="535446c9-3136-49fe-9b67-354c2ae9be55" path="/var/lib/kubelet/pods/535446c9-3136-49fe-9b67-354c2ae9be55/volumes" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.513460 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817" path="/var/lib/kubelet/pods/ebcb0ed0-29b3-4ff9-8bbf-c2cd5d963817/volumes" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.515336 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe806b26-bc8c-4e16-ba35-ffa489befc9a" path="/var/lib/kubelet/pods/fe806b26-bc8c-4e16-ba35-ffa489befc9a/volumes" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.520987 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-b6vj7" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.532336 4810 scope.go:117] "RemoveContainer" containerID="1b63d6fe729ca9dcc82b75dcf4c9bb4d47b7cd2d50638513fa29253a470549d4" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.606852 4810 scope.go:117] "RemoveContainer" containerID="8cb6af2c29480513ff7981f7f9f023b7a4bc21f738b0305082e69934bfea96c3" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.621509 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/077fd1b8-6283-4a1b-b159-437897afa556-config\") pod \"077fd1b8-6283-4a1b-b159-437897afa556\" (UID: \"077fd1b8-6283-4a1b-b159-437897afa556\") " Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.621588 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/077fd1b8-6283-4a1b-b159-437897afa556-combined-ca-bundle\") pod \"077fd1b8-6283-4a1b-b159-437897afa556\" (UID: \"077fd1b8-6283-4a1b-b159-437897afa556\") " Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.621653 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ln4z\" (UniqueName: \"kubernetes.io/projected/077fd1b8-6283-4a1b-b159-437897afa556-kube-api-access-7ln4z\") pod \"077fd1b8-6283-4a1b-b159-437897afa556\" (UID: \"077fd1b8-6283-4a1b-b159-437897afa556\") " Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.627634 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/077fd1b8-6283-4a1b-b159-437897afa556-kube-api-access-7ln4z" (OuterVolumeSpecName: "kube-api-access-7ln4z") pod "077fd1b8-6283-4a1b-b159-437897afa556" (UID: "077fd1b8-6283-4a1b-b159-437897afa556"). InnerVolumeSpecName "kube-api-access-7ln4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.635107 4810 scope.go:117] "RemoveContainer" containerID="d0da55680ea404f4e37e5d9ef3fc938bbec4754bf125f24758a733873b008d4e" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.653285 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/077fd1b8-6283-4a1b-b159-437897afa556-config" (OuterVolumeSpecName: "config") pod "077fd1b8-6283-4a1b-b159-437897afa556" (UID: "077fd1b8-6283-4a1b-b159-437897afa556"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.675014 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/077fd1b8-6283-4a1b-b159-437897afa556-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "077fd1b8-6283-4a1b-b159-437897afa556" (UID: "077fd1b8-6283-4a1b-b159-437897afa556"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.691072 4810 scope.go:117] "RemoveContainer" containerID="47157a1db99702e1e8ffb9402d3bdb1c9952d455b59c92707ba64c98952bb76c" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.724578 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/077fd1b8-6283-4a1b-b159-437897afa556-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.724609 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/077fd1b8-6283-4a1b-b159-437897afa556-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.724619 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ln4z\" (UniqueName: \"kubernetes.io/projected/077fd1b8-6283-4a1b-b159-437897afa556-kube-api-access-7ln4z\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.942299 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-v9d27"] Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.957748 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-l4gdq" event={"ID":"51938ba3-de39-414b-9777-fdc90cc14a97","Type":"ContainerStarted","Data":"ca5bbd70522206ebd37cdd1720b8e0f4acb738863472b35a2d0739bbc6d28b23"} Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.966513 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zwn5q" event={"ID":"ef385349-2803-45cf-9307-8657305dc6cf","Type":"ContainerStarted","Data":"877b17a1615addffb19f186b3b5d4ca546eeda45758f240cdd33f5e0fe8a09dc"} Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.970887 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-b6vj7" event={"ID":"077fd1b8-6283-4a1b-b159-437897afa556","Type":"ContainerDied","Data":"b12b30254bfa5200bd9b31b78fe1e9a77d98d3731e8d0cebc9eb97ba2600041f"} Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.970948 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b12b30254bfa5200bd9b31b78fe1e9a77d98d3731e8d0cebc9eb97ba2600041f" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.970913 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-b6vj7" Dec 01 14:55:42 crc kubenswrapper[4810]: E1201 14:55:42.971812 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b5266c9a26766fce2b92f95dff52d362a760f7baf1474cdcb33bd68570e096c0\\\"\"" pod="openstack/cinder-db-sync-vlw7c" podUID="09861728-4533-4ddf-a32d-3cb1b70fc325" Dec 01 14:55:42 crc kubenswrapper[4810]: I1201 14:55:42.975358 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-l4gdq" podStartSLOduration=4.115464051 podStartE2EDuration="23.975336324s" podCreationTimestamp="2025-12-01 14:55:19 +0000 UTC" firstStartedPulling="2025-12-01 14:55:20.534393601 +0000 UTC m=+1286.297903194" lastFinishedPulling="2025-12-01 14:55:40.394265864 +0000 UTC m=+1306.157775467" observedRunningTime="2025-12-01 14:55:42.97375949 +0000 UTC m=+1308.737269093" watchObservedRunningTime="2025-12-01 14:55:42.975336324 +0000 UTC m=+1308.738845927" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.023244 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.044433 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-zwn5q" podStartSLOduration=2.20805726 podStartE2EDuration="24.04441345s" podCreationTimestamp="2025-12-01 14:55:19 +0000 UTC" firstStartedPulling="2025-12-01 14:55:20.525831859 +0000 UTC m=+1286.289341462" lastFinishedPulling="2025-12-01 14:55:42.362188049 +0000 UTC m=+1308.125697652" observedRunningTime="2025-12-01 14:55:43.022955717 +0000 UTC m=+1308.786465340" watchObservedRunningTime="2025-12-01 14:55:43.04441345 +0000 UTC m=+1308.807923053" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.201555 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f677dd449-kgbj2"] Dec 01 14:55:43 crc kubenswrapper[4810]: E1201 14:55:43.202031 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="077fd1b8-6283-4a1b-b159-437897afa556" containerName="neutron-db-sync" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.202058 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="077fd1b8-6283-4a1b-b159-437897afa556" containerName="neutron-db-sync" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.202309 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="077fd1b8-6283-4a1b-b159-437897afa556" containerName="neutron-db-sync" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.203450 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.228445 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f677dd449-kgbj2"] Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.340349 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9k84\" (UniqueName: \"kubernetes.io/projected/35fd3838-aed7-46cf-aefa-df19199dc98e-kube-api-access-k9k84\") pod \"dnsmasq-dns-7f677dd449-kgbj2\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.340431 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-config\") pod \"dnsmasq-dns-7f677dd449-kgbj2\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.340513 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-dns-swift-storage-0\") pod \"dnsmasq-dns-7f677dd449-kgbj2\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.340542 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-dns-svc\") pod \"dnsmasq-dns-7f677dd449-kgbj2\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.340658 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-ovsdbserver-nb\") pod \"dnsmasq-dns-7f677dd449-kgbj2\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.340875 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-ovsdbserver-sb\") pod \"dnsmasq-dns-7f677dd449-kgbj2\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.365380 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5df84c984d-frb8v"] Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.367053 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.376084 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.376369 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-22xgp" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.376495 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.379035 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.398875 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5df84c984d-frb8v"] Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.412530 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f6d79597f-wvf92" podUID="1aec0ffb-75a0-43e6-a721-4010c6ee0f81" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: i/o timeout" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.442072 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9k84\" (UniqueName: \"kubernetes.io/projected/35fd3838-aed7-46cf-aefa-df19199dc98e-kube-api-access-k9k84\") pod \"dnsmasq-dns-7f677dd449-kgbj2\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.442136 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-combined-ca-bundle\") pod \"neutron-5df84c984d-frb8v\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.442186 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-ovndb-tls-certs\") pod \"neutron-5df84c984d-frb8v\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.442220 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-config\") pod \"dnsmasq-dns-7f677dd449-kgbj2\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.442277 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-dns-swift-storage-0\") pod \"dnsmasq-dns-7f677dd449-kgbj2\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.442304 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-dns-svc\") pod \"dnsmasq-dns-7f677dd449-kgbj2\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.442372 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-ovsdbserver-nb\") pod \"dnsmasq-dns-7f677dd449-kgbj2\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.442403 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-ovsdbserver-sb\") pod \"dnsmasq-dns-7f677dd449-kgbj2\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.442435 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-config\") pod \"neutron-5df84c984d-frb8v\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.442507 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tffmc\" (UniqueName: \"kubernetes.io/projected/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-kube-api-access-tffmc\") pod \"neutron-5df84c984d-frb8v\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.442571 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-httpd-config\") pod \"neutron-5df84c984d-frb8v\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.443937 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-ovsdbserver-nb\") pod \"dnsmasq-dns-7f677dd449-kgbj2\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.444539 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-config\") pod \"dnsmasq-dns-7f677dd449-kgbj2\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.444710 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-dns-swift-storage-0\") pod \"dnsmasq-dns-7f677dd449-kgbj2\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.446504 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-ovsdbserver-sb\") pod \"dnsmasq-dns-7f677dd449-kgbj2\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.447994 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-dns-svc\") pod \"dnsmasq-dns-7f677dd449-kgbj2\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.459643 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9k84\" (UniqueName: \"kubernetes.io/projected/35fd3838-aed7-46cf-aefa-df19199dc98e-kube-api-access-k9k84\") pod \"dnsmasq-dns-7f677dd449-kgbj2\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.520812 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.547000 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-config\") pod \"neutron-5df84c984d-frb8v\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.547962 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tffmc\" (UniqueName: \"kubernetes.io/projected/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-kube-api-access-tffmc\") pod \"neutron-5df84c984d-frb8v\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.548316 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-httpd-config\") pod \"neutron-5df84c984d-frb8v\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.548887 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-combined-ca-bundle\") pod \"neutron-5df84c984d-frb8v\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.549065 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-ovndb-tls-certs\") pod \"neutron-5df84c984d-frb8v\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.554366 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-combined-ca-bundle\") pod \"neutron-5df84c984d-frb8v\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.554685 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-config\") pod \"neutron-5df84c984d-frb8v\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.556091 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-httpd-config\") pod \"neutron-5df84c984d-frb8v\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.558320 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-ovndb-tls-certs\") pod \"neutron-5df84c984d-frb8v\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.569280 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tffmc\" (UniqueName: \"kubernetes.io/projected/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-kube-api-access-tffmc\") pod \"neutron-5df84c984d-frb8v\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.690159 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.994182 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3039d59e-4377-411d-98e7-9fd84a2d93b6","Type":"ContainerStarted","Data":"8ad63be1f8544e127ea84a20e8394c29ddf23c0d67240904b40820218c663e4d"} Dec 01 14:55:43 crc kubenswrapper[4810]: I1201 14:55:43.996571 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"073cab97-c80f-4905-92df-82b2537ed262","Type":"ContainerStarted","Data":"1d39fda5deb853dc5f22a662445df2f289042a4470edfdd81faeef5d2732a791"} Dec 01 14:55:44 crc kubenswrapper[4810]: I1201 14:55:44.000691 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v9d27" event={"ID":"5b341ba0-f0a3-40f7-a99b-7eb045ba4739","Type":"ContainerStarted","Data":"375d1e18c5849fe9602a5edfd221a72767570f1b703cd433970ef2374a9b748b"} Dec 01 14:55:44 crc kubenswrapper[4810]: I1201 14:55:44.000720 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v9d27" event={"ID":"5b341ba0-f0a3-40f7-a99b-7eb045ba4739","Type":"ContainerStarted","Data":"1aefe40b089d6cd4e6a399f6d61e8b50902c24622cbb25efe1aec1863ec62bb6"} Dec 01 14:55:44 crc kubenswrapper[4810]: I1201 14:55:44.035364 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-v9d27" podStartSLOduration=3.035344136 podStartE2EDuration="3.035344136s" podCreationTimestamp="2025-12-01 14:55:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:55:44.029274561 +0000 UTC m=+1309.792784164" watchObservedRunningTime="2025-12-01 14:55:44.035344136 +0000 UTC m=+1309.798853739" Dec 01 14:55:44 crc kubenswrapper[4810]: I1201 14:55:44.089911 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f677dd449-kgbj2"] Dec 01 14:55:44 crc kubenswrapper[4810]: W1201 14:55:44.101950 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35fd3838_aed7_46cf_aefa_df19199dc98e.slice/crio-f4c4ebea0e0cc62f70f59ebcde98ac571b044ff1a0ff48ed2b3fb5bed5264727 WatchSource:0}: Error finding container f4c4ebea0e0cc62f70f59ebcde98ac571b044ff1a0ff48ed2b3fb5bed5264727: Status 404 returned error can't find the container with id f4c4ebea0e0cc62f70f59ebcde98ac571b044ff1a0ff48ed2b3fb5bed5264727 Dec 01 14:55:44 crc kubenswrapper[4810]: I1201 14:55:44.180908 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 14:55:44 crc kubenswrapper[4810]: W1201 14:55:44.202426 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd18c05bc_3e20_4305_b8bf_d45e6ff46308.slice/crio-91fb5799b749f63d23a20589068152c4f2153471a0828e73df67eaee15e98a21 WatchSource:0}: Error finding container 91fb5799b749f63d23a20589068152c4f2153471a0828e73df67eaee15e98a21: Status 404 returned error can't find the container with id 91fb5799b749f63d23a20589068152c4f2153471a0828e73df67eaee15e98a21 Dec 01 14:55:44 crc kubenswrapper[4810]: I1201 14:55:44.438094 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5df84c984d-frb8v"] Dec 01 14:55:44 crc kubenswrapper[4810]: W1201 14:55:44.467743 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccf7f439_d5a8_4bcc_a71d_dc6ceae59e42.slice/crio-ab9c35c84443abd5384392ddbe947e86dbbc99b0630611d10f21f1953f5394aa WatchSource:0}: Error finding container ab9c35c84443abd5384392ddbe947e86dbbc99b0630611d10f21f1953f5394aa: Status 404 returned error can't find the container with id ab9c35c84443abd5384392ddbe947e86dbbc99b0630611d10f21f1953f5394aa Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.014663 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5df84c984d-frb8v" event={"ID":"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42","Type":"ContainerStarted","Data":"a81f5574834d150e9892017d8ae052bc6da9601608100fab0fac319ce5cd3cf8"} Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.014952 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5df84c984d-frb8v" event={"ID":"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42","Type":"ContainerStarted","Data":"ab9c35c84443abd5384392ddbe947e86dbbc99b0630611d10f21f1953f5394aa"} Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.018296 4810 generic.go:334] "Generic (PLEG): container finished" podID="35fd3838-aed7-46cf-aefa-df19199dc98e" containerID="12c228f4af297ebce306363df7b12d715ab3025c024395e1166eba451956709f" exitCode=0 Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.018370 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" event={"ID":"35fd3838-aed7-46cf-aefa-df19199dc98e","Type":"ContainerDied","Data":"12c228f4af297ebce306363df7b12d715ab3025c024395e1166eba451956709f"} Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.018437 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" event={"ID":"35fd3838-aed7-46cf-aefa-df19199dc98e","Type":"ContainerStarted","Data":"f4c4ebea0e0cc62f70f59ebcde98ac571b044ff1a0ff48ed2b3fb5bed5264727"} Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.030039 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3039d59e-4377-411d-98e7-9fd84a2d93b6","Type":"ContainerStarted","Data":"ad42623d524fd2d5246b4533a6cc14b8730ace1654df85830c52e33fdeecce05"} Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.030965 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d18c05bc-3e20-4305-b8bf-d45e6ff46308","Type":"ContainerStarted","Data":"91fb5799b749f63d23a20589068152c4f2153471a0828e73df67eaee15e98a21"} Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.774494 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5477779d95-wg2vp"] Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.776300 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.780148 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.784121 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.792447 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5477779d95-wg2vp"] Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.921135 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-public-tls-certs\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.921184 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-internal-tls-certs\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.921237 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mrmd\" (UniqueName: \"kubernetes.io/projected/ba5f8a69-22cf-4196-bc89-ffa60eda9303-kube-api-access-6mrmd\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.921272 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-combined-ca-bundle\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.921393 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-config\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.921420 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-ovndb-tls-certs\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:45 crc kubenswrapper[4810]: I1201 14:55:45.921501 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-httpd-config\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.024635 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-config\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.024685 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-ovndb-tls-certs\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.024731 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-httpd-config\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.024756 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-public-tls-certs\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.024784 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-internal-tls-certs\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.024818 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mrmd\" (UniqueName: \"kubernetes.io/projected/ba5f8a69-22cf-4196-bc89-ffa60eda9303-kube-api-access-6mrmd\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.024867 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-combined-ca-bundle\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.032961 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-internal-tls-certs\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.039695 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-combined-ca-bundle\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.040263 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-public-tls-certs\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.041032 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-ovndb-tls-certs\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.041282 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-config\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.044203 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-httpd-config\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.048170 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mrmd\" (UniqueName: \"kubernetes.io/projected/ba5f8a69-22cf-4196-bc89-ffa60eda9303-kube-api-access-6mrmd\") pod \"neutron-5477779d95-wg2vp\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.063737 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3039d59e-4377-411d-98e7-9fd84a2d93b6","Type":"ContainerStarted","Data":"0ce22d98360b8f8557cdeab1c36e735725358d0cfe0253144934e0ceb6ba1349"} Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.069614 4810 generic.go:334] "Generic (PLEG): container finished" podID="51938ba3-de39-414b-9777-fdc90cc14a97" containerID="ca5bbd70522206ebd37cdd1720b8e0f4acb738863472b35a2d0739bbc6d28b23" exitCode=0 Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.069723 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-l4gdq" event={"ID":"51938ba3-de39-414b-9777-fdc90cc14a97","Type":"ContainerDied","Data":"ca5bbd70522206ebd37cdd1720b8e0f4acb738863472b35a2d0739bbc6d28b23"} Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.072217 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d18c05bc-3e20-4305-b8bf-d45e6ff46308","Type":"ContainerStarted","Data":"11cb52869bfafc31bb1d1d214f6a4b58a7e39a377578ec37c708f280ae68b49f"} Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.072250 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d18c05bc-3e20-4305-b8bf-d45e6ff46308","Type":"ContainerStarted","Data":"373d8ee5e28ab65af88de3532b2fd0604c91e4cbc4b1d3f1e4e86653bfc62288"} Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.074332 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5df84c984d-frb8v" event={"ID":"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42","Type":"ContainerStarted","Data":"ce152a2dcedab6dfa2222a388bf9973c1bcce7517a6f7e4bf7a4901cbcf8f86a"} Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.075170 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.078452 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" event={"ID":"35fd3838-aed7-46cf-aefa-df19199dc98e","Type":"ContainerStarted","Data":"68a5d8b615cff419baff4fa7f934f7a2242a8bc7b3b97c2aab2e9ba25573f120"} Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.079751 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.131618 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.136665 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.136649063 podStartE2EDuration="5.136649063s" podCreationTimestamp="2025-12-01 14:55:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:55:46.135282616 +0000 UTC m=+1311.898792219" watchObservedRunningTime="2025-12-01 14:55:46.136649063 +0000 UTC m=+1311.900158666" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.139857 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.13983986 podStartE2EDuration="6.13983986s" podCreationTimestamp="2025-12-01 14:55:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:55:46.107008558 +0000 UTC m=+1311.870518161" watchObservedRunningTime="2025-12-01 14:55:46.13983986 +0000 UTC m=+1311.903349453" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.171180 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5df84c984d-frb8v" podStartSLOduration=3.17114145 podStartE2EDuration="3.17114145s" podCreationTimestamp="2025-12-01 14:55:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:55:46.154649052 +0000 UTC m=+1311.918158665" watchObservedRunningTime="2025-12-01 14:55:46.17114145 +0000 UTC m=+1311.934651053" Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.237604 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" podStartSLOduration=3.237573925 podStartE2EDuration="3.237573925s" podCreationTimestamp="2025-12-01 14:55:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:55:46.197545627 +0000 UTC m=+1311.961055230" watchObservedRunningTime="2025-12-01 14:55:46.237573925 +0000 UTC m=+1312.001083528" Dec 01 14:55:46 crc kubenswrapper[4810]: W1201 14:55:46.853373 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba5f8a69_22cf_4196_bc89_ffa60eda9303.slice/crio-8c05d4229dd218e5bf9e4223256a65e9728e52b9a83b569c8efa023476792fe0 WatchSource:0}: Error finding container 8c05d4229dd218e5bf9e4223256a65e9728e52b9a83b569c8efa023476792fe0: Status 404 returned error can't find the container with id 8c05d4229dd218e5bf9e4223256a65e9728e52b9a83b569c8efa023476792fe0 Dec 01 14:55:46 crc kubenswrapper[4810]: I1201 14:55:46.854130 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5477779d95-wg2vp"] Dec 01 14:55:47 crc kubenswrapper[4810]: I1201 14:55:47.095937 4810 generic.go:334] "Generic (PLEG): container finished" podID="ef385349-2803-45cf-9307-8657305dc6cf" containerID="877b17a1615addffb19f186b3b5d4ca546eeda45758f240cdd33f5e0fe8a09dc" exitCode=0 Dec 01 14:55:47 crc kubenswrapper[4810]: I1201 14:55:47.096314 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zwn5q" event={"ID":"ef385349-2803-45cf-9307-8657305dc6cf","Type":"ContainerDied","Data":"877b17a1615addffb19f186b3b5d4ca546eeda45758f240cdd33f5e0fe8a09dc"} Dec 01 14:55:47 crc kubenswrapper[4810]: I1201 14:55:47.101090 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5477779d95-wg2vp" event={"ID":"ba5f8a69-22cf-4196-bc89-ffa60eda9303","Type":"ContainerStarted","Data":"8c05d4229dd218e5bf9e4223256a65e9728e52b9a83b569c8efa023476792fe0"} Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.110446 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5477779d95-wg2vp" event={"ID":"ba5f8a69-22cf-4196-bc89-ffa60eda9303","Type":"ContainerStarted","Data":"7b6de236d9a353033c45404d0d0dce626c48768c8ec55f0ea9bf63e5d411c381"} Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.112060 4810 generic.go:334] "Generic (PLEG): container finished" podID="5b341ba0-f0a3-40f7-a99b-7eb045ba4739" containerID="375d1e18c5849fe9602a5edfd221a72767570f1b703cd433970ef2374a9b748b" exitCode=0 Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.112128 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v9d27" event={"ID":"5b341ba0-f0a3-40f7-a99b-7eb045ba4739","Type":"ContainerDied","Data":"375d1e18c5849fe9602a5edfd221a72767570f1b703cd433970ef2374a9b748b"} Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.640659 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.648752 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zwn5q" Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.789872 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcv9v\" (UniqueName: \"kubernetes.io/projected/ef385349-2803-45cf-9307-8657305dc6cf-kube-api-access-rcv9v\") pod \"ef385349-2803-45cf-9307-8657305dc6cf\" (UID: \"ef385349-2803-45cf-9307-8657305dc6cf\") " Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.789979 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51938ba3-de39-414b-9777-fdc90cc14a97-config-data\") pod \"51938ba3-de39-414b-9777-fdc90cc14a97\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.790062 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51938ba3-de39-414b-9777-fdc90cc14a97-combined-ca-bundle\") pod \"51938ba3-de39-414b-9777-fdc90cc14a97\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.790090 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnzbj\" (UniqueName: \"kubernetes.io/projected/51938ba3-de39-414b-9777-fdc90cc14a97-kube-api-access-bnzbj\") pod \"51938ba3-de39-414b-9777-fdc90cc14a97\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.790540 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51938ba3-de39-414b-9777-fdc90cc14a97-logs\") pod \"51938ba3-de39-414b-9777-fdc90cc14a97\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.790607 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51938ba3-de39-414b-9777-fdc90cc14a97-scripts\") pod \"51938ba3-de39-414b-9777-fdc90cc14a97\" (UID: \"51938ba3-de39-414b-9777-fdc90cc14a97\") " Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.790694 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef385349-2803-45cf-9307-8657305dc6cf-combined-ca-bundle\") pod \"ef385349-2803-45cf-9307-8657305dc6cf\" (UID: \"ef385349-2803-45cf-9307-8657305dc6cf\") " Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.790740 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef385349-2803-45cf-9307-8657305dc6cf-db-sync-config-data\") pod \"ef385349-2803-45cf-9307-8657305dc6cf\" (UID: \"ef385349-2803-45cf-9307-8657305dc6cf\") " Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.792985 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51938ba3-de39-414b-9777-fdc90cc14a97-logs" (OuterVolumeSpecName: "logs") pod "51938ba3-de39-414b-9777-fdc90cc14a97" (UID: "51938ba3-de39-414b-9777-fdc90cc14a97"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.796930 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51938ba3-de39-414b-9777-fdc90cc14a97-kube-api-access-bnzbj" (OuterVolumeSpecName: "kube-api-access-bnzbj") pod "51938ba3-de39-414b-9777-fdc90cc14a97" (UID: "51938ba3-de39-414b-9777-fdc90cc14a97"). InnerVolumeSpecName "kube-api-access-bnzbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.796988 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef385349-2803-45cf-9307-8657305dc6cf-kube-api-access-rcv9v" (OuterVolumeSpecName: "kube-api-access-rcv9v") pod "ef385349-2803-45cf-9307-8657305dc6cf" (UID: "ef385349-2803-45cf-9307-8657305dc6cf"). InnerVolumeSpecName "kube-api-access-rcv9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.796989 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51938ba3-de39-414b-9777-fdc90cc14a97-scripts" (OuterVolumeSpecName: "scripts") pod "51938ba3-de39-414b-9777-fdc90cc14a97" (UID: "51938ba3-de39-414b-9777-fdc90cc14a97"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.797150 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef385349-2803-45cf-9307-8657305dc6cf-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ef385349-2803-45cf-9307-8657305dc6cf" (UID: "ef385349-2803-45cf-9307-8657305dc6cf"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.816030 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51938ba3-de39-414b-9777-fdc90cc14a97-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51938ba3-de39-414b-9777-fdc90cc14a97" (UID: "51938ba3-de39-414b-9777-fdc90cc14a97"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.816442 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef385349-2803-45cf-9307-8657305dc6cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef385349-2803-45cf-9307-8657305dc6cf" (UID: "ef385349-2803-45cf-9307-8657305dc6cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.817724 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51938ba3-de39-414b-9777-fdc90cc14a97-config-data" (OuterVolumeSpecName: "config-data") pod "51938ba3-de39-414b-9777-fdc90cc14a97" (UID: "51938ba3-de39-414b-9777-fdc90cc14a97"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.893510 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef385349-2803-45cf-9307-8657305dc6cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.893551 4810 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef385349-2803-45cf-9307-8657305dc6cf-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.893566 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcv9v\" (UniqueName: \"kubernetes.io/projected/ef385349-2803-45cf-9307-8657305dc6cf-kube-api-access-rcv9v\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.893591 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51938ba3-de39-414b-9777-fdc90cc14a97-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.893604 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51938ba3-de39-414b-9777-fdc90cc14a97-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.893615 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnzbj\" (UniqueName: \"kubernetes.io/projected/51938ba3-de39-414b-9777-fdc90cc14a97-kube-api-access-bnzbj\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.893626 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51938ba3-de39-414b-9777-fdc90cc14a97-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:48 crc kubenswrapper[4810]: I1201 14:55:48.893638 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51938ba3-de39-414b-9777-fdc90cc14a97-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.127111 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zwn5q" event={"ID":"ef385349-2803-45cf-9307-8657305dc6cf","Type":"ContainerDied","Data":"f0ad77614f91f3187631da4101355ee3cc53c89228e5f70fc1166fad86a7f416"} Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.127153 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0ad77614f91f3187631da4101355ee3cc53c89228e5f70fc1166fad86a7f416" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.127938 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zwn5q" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.130916 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-l4gdq" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.131515 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-l4gdq" event={"ID":"51938ba3-de39-414b-9777-fdc90cc14a97","Type":"ContainerDied","Data":"b490863d702616d194a2068a92afead078ac76cd4000ddeaa65b07572b8e0ccc"} Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.131540 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b490863d702616d194a2068a92afead078ac76cd4000ddeaa65b07572b8e0ccc" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.460853 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6d8fb448c4-2hg9g"] Dec 01 14:55:49 crc kubenswrapper[4810]: E1201 14:55:49.461768 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51938ba3-de39-414b-9777-fdc90cc14a97" containerName="placement-db-sync" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.461786 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="51938ba3-de39-414b-9777-fdc90cc14a97" containerName="placement-db-sync" Dec 01 14:55:49 crc kubenswrapper[4810]: E1201 14:55:49.461809 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef385349-2803-45cf-9307-8657305dc6cf" containerName="barbican-db-sync" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.461817 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef385349-2803-45cf-9307-8657305dc6cf" containerName="barbican-db-sync" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.462012 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef385349-2803-45cf-9307-8657305dc6cf" containerName="barbican-db-sync" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.462032 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="51938ba3-de39-414b-9777-fdc90cc14a97" containerName="placement-db-sync" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.464316 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.480686 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.480820 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-ppms8" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.481086 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.497501 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-84864d5bd5-wmq86"] Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.499139 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.511864 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-combined-ca-bundle\") pod \"barbican-keystone-listener-6d8fb448c4-2hg9g\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.512251 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-config-data\") pod \"barbican-keystone-listener-6d8fb448c4-2hg9g\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.512291 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9xp8\" (UniqueName: \"kubernetes.io/projected/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-kube-api-access-d9xp8\") pod \"barbican-keystone-listener-6d8fb448c4-2hg9g\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.512365 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-config-data-custom\") pod \"barbican-keystone-listener-6d8fb448c4-2hg9g\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.512420 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-logs\") pod \"barbican-keystone-listener-6d8fb448c4-2hg9g\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.541255 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.548973 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-84864d5bd5-wmq86"] Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.575518 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6d8fb448c4-2hg9g"] Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.596320 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f677dd449-kgbj2"] Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.596563 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" podUID="35fd3838-aed7-46cf-aefa-df19199dc98e" containerName="dnsmasq-dns" containerID="cri-o://68a5d8b615cff419baff4fa7f934f7a2242a8bc7b3b97c2aab2e9ba25573f120" gracePeriod=10 Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.613389 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-config-data\") pod \"barbican-keystone-listener-6d8fb448c4-2hg9g\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.613436 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9xp8\" (UniqueName: \"kubernetes.io/projected/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-kube-api-access-d9xp8\") pod \"barbican-keystone-listener-6d8fb448c4-2hg9g\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.613475 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tls46\" (UniqueName: \"kubernetes.io/projected/129ae854-1430-465f-a177-0c459bd8b4c8-kube-api-access-tls46\") pod \"barbican-worker-84864d5bd5-wmq86\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.613507 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-config-data-custom\") pod \"barbican-keystone-listener-6d8fb448c4-2hg9g\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.613546 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-logs\") pod \"barbican-keystone-listener-6d8fb448c4-2hg9g\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.613570 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/129ae854-1430-465f-a177-0c459bd8b4c8-config-data\") pod \"barbican-worker-84864d5bd5-wmq86\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.613588 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/129ae854-1430-465f-a177-0c459bd8b4c8-logs\") pod \"barbican-worker-84864d5bd5-wmq86\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.613618 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/129ae854-1430-465f-a177-0c459bd8b4c8-config-data-custom\") pod \"barbican-worker-84864d5bd5-wmq86\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.613641 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-combined-ca-bundle\") pod \"barbican-keystone-listener-6d8fb448c4-2hg9g\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.613675 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/129ae854-1430-465f-a177-0c459bd8b4c8-combined-ca-bundle\") pod \"barbican-worker-84864d5bd5-wmq86\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.614642 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-logs\") pod \"barbican-keystone-listener-6d8fb448c4-2hg9g\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.633570 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-config-data\") pod \"barbican-keystone-listener-6d8fb448c4-2hg9g\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.635428 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-config-data-custom\") pod \"barbican-keystone-listener-6d8fb448c4-2hg9g\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.644757 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9xp8\" (UniqueName: \"kubernetes.io/projected/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-kube-api-access-d9xp8\") pod \"barbican-keystone-listener-6d8fb448c4-2hg9g\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.651105 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64dfd64c45-nssxw"] Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.652595 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.655356 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-combined-ca-bundle\") pod \"barbican-keystone-listener-6d8fb448c4-2hg9g\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.678540 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-dbfcbd88b-69qn5"] Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.681076 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.685268 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.709354 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64dfd64c45-nssxw"] Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.715502 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tls46\" (UniqueName: \"kubernetes.io/projected/129ae854-1430-465f-a177-0c459bd8b4c8-kube-api-access-tls46\") pod \"barbican-worker-84864d5bd5-wmq86\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.715585 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/129ae854-1430-465f-a177-0c459bd8b4c8-config-data\") pod \"barbican-worker-84864d5bd5-wmq86\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.715607 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/129ae854-1430-465f-a177-0c459bd8b4c8-logs\") pod \"barbican-worker-84864d5bd5-wmq86\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.715642 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/129ae854-1430-465f-a177-0c459bd8b4c8-config-data-custom\") pod \"barbican-worker-84864d5bd5-wmq86\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.715683 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/129ae854-1430-465f-a177-0c459bd8b4c8-combined-ca-bundle\") pod \"barbican-worker-84864d5bd5-wmq86\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.721247 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-dbfcbd88b-69qn5"] Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.722265 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/129ae854-1430-465f-a177-0c459bd8b4c8-logs\") pod \"barbican-worker-84864d5bd5-wmq86\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.724928 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/129ae854-1430-465f-a177-0c459bd8b4c8-config-data\") pod \"barbican-worker-84864d5bd5-wmq86\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.726693 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/129ae854-1430-465f-a177-0c459bd8b4c8-combined-ca-bundle\") pod \"barbican-worker-84864d5bd5-wmq86\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.734806 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/129ae854-1430-465f-a177-0c459bd8b4c8-config-data-custom\") pod \"barbican-worker-84864d5bd5-wmq86\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.748490 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tls46\" (UniqueName: \"kubernetes.io/projected/129ae854-1430-465f-a177-0c459bd8b4c8-kube-api-access-tls46\") pod \"barbican-worker-84864d5bd5-wmq86\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.799568 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6c5f549c7d-52fkz"] Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.816679 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66f9e303-31b7-4355-9acd-2720e03a9bb5-config-data-custom\") pod \"barbican-api-dbfcbd88b-69qn5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.816725 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-ovsdbserver-nb\") pod \"dnsmasq-dns-64dfd64c45-nssxw\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.816743 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdbjs\" (UniqueName: \"kubernetes.io/projected/66f9e303-31b7-4355-9acd-2720e03a9bb5-kube-api-access-gdbjs\") pod \"barbican-api-dbfcbd88b-69qn5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.816824 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66f9e303-31b7-4355-9acd-2720e03a9bb5-logs\") pod \"barbican-api-dbfcbd88b-69qn5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.816847 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-ovsdbserver-sb\") pod \"dnsmasq-dns-64dfd64c45-nssxw\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.816882 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-config\") pod \"dnsmasq-dns-64dfd64c45-nssxw\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.816931 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k2tq\" (UniqueName: \"kubernetes.io/projected/21371b53-8ea0-40e1-a68e-c140a606b610-kube-api-access-2k2tq\") pod \"dnsmasq-dns-64dfd64c45-nssxw\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.816974 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66f9e303-31b7-4355-9acd-2720e03a9bb5-config-data\") pod \"barbican-api-dbfcbd88b-69qn5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.816999 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66f9e303-31b7-4355-9acd-2720e03a9bb5-combined-ca-bundle\") pod \"barbican-api-dbfcbd88b-69qn5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.817024 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-dns-swift-storage-0\") pod \"dnsmasq-dns-64dfd64c45-nssxw\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.817042 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-dns-svc\") pod \"dnsmasq-dns-64dfd64c45-nssxw\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.818698 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.821946 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.822344 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.823288 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.823405 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-hgvgs" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.825933 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.836891 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6c5f549c7d-52fkz"] Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.847961 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.878777 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.918668 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66f9e303-31b7-4355-9acd-2720e03a9bb5-config-data\") pod \"barbican-api-dbfcbd88b-69qn5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.918741 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66f9e303-31b7-4355-9acd-2720e03a9bb5-combined-ca-bundle\") pod \"barbican-api-dbfcbd88b-69qn5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.918780 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-dns-swift-storage-0\") pod \"dnsmasq-dns-64dfd64c45-nssxw\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.918807 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-dns-svc\") pod \"dnsmasq-dns-64dfd64c45-nssxw\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.918842 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-internal-tls-certs\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.918864 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-public-tls-certs\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.918907 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-combined-ca-bundle\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.918930 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66f9e303-31b7-4355-9acd-2720e03a9bb5-config-data-custom\") pod \"barbican-api-dbfcbd88b-69qn5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.918952 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8hpz\" (UniqueName: \"kubernetes.io/projected/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-kube-api-access-q8hpz\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.918970 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-ovsdbserver-nb\") pod \"dnsmasq-dns-64dfd64c45-nssxw\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.918987 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdbjs\" (UniqueName: \"kubernetes.io/projected/66f9e303-31b7-4355-9acd-2720e03a9bb5-kube-api-access-gdbjs\") pod \"barbican-api-dbfcbd88b-69qn5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.919006 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-logs\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.919035 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-config-data\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.919059 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66f9e303-31b7-4355-9acd-2720e03a9bb5-logs\") pod \"barbican-api-dbfcbd88b-69qn5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.919079 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-ovsdbserver-sb\") pod \"dnsmasq-dns-64dfd64c45-nssxw\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.919114 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-config\") pod \"dnsmasq-dns-64dfd64c45-nssxw\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.919147 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-scripts\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.919163 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k2tq\" (UniqueName: \"kubernetes.io/projected/21371b53-8ea0-40e1-a68e-c140a606b610-kube-api-access-2k2tq\") pod \"dnsmasq-dns-64dfd64c45-nssxw\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.920583 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-ovsdbserver-nb\") pod \"dnsmasq-dns-64dfd64c45-nssxw\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.921130 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-dns-swift-storage-0\") pod \"dnsmasq-dns-64dfd64c45-nssxw\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.921156 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66f9e303-31b7-4355-9acd-2720e03a9bb5-logs\") pod \"barbican-api-dbfcbd88b-69qn5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.922266 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-config\") pod \"dnsmasq-dns-64dfd64c45-nssxw\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.922829 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-dns-svc\") pod \"dnsmasq-dns-64dfd64c45-nssxw\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.923766 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66f9e303-31b7-4355-9acd-2720e03a9bb5-config-data-custom\") pod \"barbican-api-dbfcbd88b-69qn5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.924272 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66f9e303-31b7-4355-9acd-2720e03a9bb5-combined-ca-bundle\") pod \"barbican-api-dbfcbd88b-69qn5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.935862 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-ovsdbserver-sb\") pod \"dnsmasq-dns-64dfd64c45-nssxw\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.936109 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66f9e303-31b7-4355-9acd-2720e03a9bb5-config-data\") pod \"barbican-api-dbfcbd88b-69qn5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.939808 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k2tq\" (UniqueName: \"kubernetes.io/projected/21371b53-8ea0-40e1-a68e-c140a606b610-kube-api-access-2k2tq\") pod \"dnsmasq-dns-64dfd64c45-nssxw\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:49 crc kubenswrapper[4810]: I1201 14:55:49.940888 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdbjs\" (UniqueName: \"kubernetes.io/projected/66f9e303-31b7-4355-9acd-2720e03a9bb5-kube-api-access-gdbjs\") pod \"barbican-api-dbfcbd88b-69qn5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.021383 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-config-data\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.021535 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-scripts\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.021646 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-internal-tls-certs\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.021667 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-public-tls-certs\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.021708 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-combined-ca-bundle\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.021743 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8hpz\" (UniqueName: \"kubernetes.io/projected/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-kube-api-access-q8hpz\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.021781 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-logs\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.022620 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-logs\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.027423 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-public-tls-certs\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.028334 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-config-data\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.028888 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-combined-ca-bundle\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.030175 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-scripts\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.035773 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-internal-tls-certs\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.039168 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8hpz\" (UniqueName: \"kubernetes.io/projected/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-kube-api-access-q8hpz\") pod \"placement-6c5f549c7d-52fkz\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.129539 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.154669 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.155448 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.162889 4810 generic.go:334] "Generic (PLEG): container finished" podID="35fd3838-aed7-46cf-aefa-df19199dc98e" containerID="68a5d8b615cff419baff4fa7f934f7a2242a8bc7b3b97c2aab2e9ba25573f120" exitCode=0 Dec 01 14:55:50 crc kubenswrapper[4810]: I1201 14:55:50.162949 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" event={"ID":"35fd3838-aed7-46cf-aefa-df19199dc98e","Type":"ContainerDied","Data":"68a5d8b615cff419baff4fa7f934f7a2242a8bc7b3b97c2aab2e9ba25573f120"} Dec 01 14:55:51 crc kubenswrapper[4810]: I1201 14:55:51.464726 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 14:55:51 crc kubenswrapper[4810]: I1201 14:55:51.465011 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 14:55:51 crc kubenswrapper[4810]: I1201 14:55:51.498179 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 14:55:51 crc kubenswrapper[4810]: I1201 14:55:51.510643 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 14:55:51 crc kubenswrapper[4810]: I1201 14:55:51.737565 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 14:55:51 crc kubenswrapper[4810]: I1201 14:55:51.737614 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 14:55:51 crc kubenswrapper[4810]: I1201 14:55:51.763298 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 14:55:51 crc kubenswrapper[4810]: I1201 14:55:51.775577 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.115853 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5b4ff7d6cb-m2g6c"] Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.117817 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.124094 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.124244 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.135252 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5b4ff7d6cb-m2g6c"] Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.180044 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.180175 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.180194 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.180207 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.268165 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-config-data\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.268205 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mrbz\" (UniqueName: \"kubernetes.io/projected/bdd90cbb-4242-4227-9de4-09f99e60a516-kube-api-access-9mrbz\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.268238 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-config-data-custom\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.268564 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-combined-ca-bundle\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.268626 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-public-tls-certs\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.268643 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-internal-tls-certs\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.268679 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdd90cbb-4242-4227-9de4-09f99e60a516-logs\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.370283 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-public-tls-certs\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.370455 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-internal-tls-certs\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.370652 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdd90cbb-4242-4227-9de4-09f99e60a516-logs\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.370899 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-config-data\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.371003 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mrbz\" (UniqueName: \"kubernetes.io/projected/bdd90cbb-4242-4227-9de4-09f99e60a516-kube-api-access-9mrbz\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.370951 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdd90cbb-4242-4227-9de4-09f99e60a516-logs\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.371085 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-config-data-custom\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.371201 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-combined-ca-bundle\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.377333 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-config-data-custom\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.381595 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-combined-ca-bundle\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.381649 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-public-tls-certs\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.382854 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-config-data\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.385056 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-internal-tls-certs\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.398041 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mrbz\" (UniqueName: \"kubernetes.io/projected/bdd90cbb-4242-4227-9de4-09f99e60a516-kube-api-access-9mrbz\") pod \"barbican-api-5b4ff7d6cb-m2g6c\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.445764 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.518571 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.574047 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-credential-keys\") pod \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.574427 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-fernet-keys\") pod \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.574480 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74n65\" (UniqueName: \"kubernetes.io/projected/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-kube-api-access-74n65\") pod \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.574567 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-scripts\") pod \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.574600 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-config-data\") pod \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.574656 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-combined-ca-bundle\") pod \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\" (UID: \"5b341ba0-f0a3-40f7-a99b-7eb045ba4739\") " Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.579748 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-kube-api-access-74n65" (OuterVolumeSpecName: "kube-api-access-74n65") pod "5b341ba0-f0a3-40f7-a99b-7eb045ba4739" (UID: "5b341ba0-f0a3-40f7-a99b-7eb045ba4739"). InnerVolumeSpecName "kube-api-access-74n65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.585204 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5b341ba0-f0a3-40f7-a99b-7eb045ba4739" (UID: "5b341ba0-f0a3-40f7-a99b-7eb045ba4739"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.585651 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5b341ba0-f0a3-40f7-a99b-7eb045ba4739" (UID: "5b341ba0-f0a3-40f7-a99b-7eb045ba4739"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.616524 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-scripts" (OuterVolumeSpecName: "scripts") pod "5b341ba0-f0a3-40f7-a99b-7eb045ba4739" (UID: "5b341ba0-f0a3-40f7-a99b-7eb045ba4739"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.622621 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b341ba0-f0a3-40f7-a99b-7eb045ba4739" (UID: "5b341ba0-f0a3-40f7-a99b-7eb045ba4739"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.679844 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74n65\" (UniqueName: \"kubernetes.io/projected/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-kube-api-access-74n65\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.680228 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.680247 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.680257 4810 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.680267 4810 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.697519 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-config-data" (OuterVolumeSpecName: "config-data") pod "5b341ba0-f0a3-40f7-a99b-7eb045ba4739" (UID: "5b341ba0-f0a3-40f7-a99b-7eb045ba4739"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.784070 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b341ba0-f0a3-40f7-a99b-7eb045ba4739-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.849050 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.986308 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-config\") pod \"35fd3838-aed7-46cf-aefa-df19199dc98e\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.986734 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-ovsdbserver-sb\") pod \"35fd3838-aed7-46cf-aefa-df19199dc98e\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.986772 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9k84\" (UniqueName: \"kubernetes.io/projected/35fd3838-aed7-46cf-aefa-df19199dc98e-kube-api-access-k9k84\") pod \"35fd3838-aed7-46cf-aefa-df19199dc98e\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.986832 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-dns-svc\") pod \"35fd3838-aed7-46cf-aefa-df19199dc98e\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.986863 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-dns-swift-storage-0\") pod \"35fd3838-aed7-46cf-aefa-df19199dc98e\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " Dec 01 14:55:52 crc kubenswrapper[4810]: I1201 14:55:52.986899 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-ovsdbserver-nb\") pod \"35fd3838-aed7-46cf-aefa-df19199dc98e\" (UID: \"35fd3838-aed7-46cf-aefa-df19199dc98e\") " Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.012069 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35fd3838-aed7-46cf-aefa-df19199dc98e-kube-api-access-k9k84" (OuterVolumeSpecName: "kube-api-access-k9k84") pod "35fd3838-aed7-46cf-aefa-df19199dc98e" (UID: "35fd3838-aed7-46cf-aefa-df19199dc98e"). InnerVolumeSpecName "kube-api-access-k9k84". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.089204 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9k84\" (UniqueName: \"kubernetes.io/projected/35fd3838-aed7-46cf-aefa-df19199dc98e-kube-api-access-k9k84\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.174828 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-config" (OuterVolumeSpecName: "config") pod "35fd3838-aed7-46cf-aefa-df19199dc98e" (UID: "35fd3838-aed7-46cf-aefa-df19199dc98e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.183770 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "35fd3838-aed7-46cf-aefa-df19199dc98e" (UID: "35fd3838-aed7-46cf-aefa-df19199dc98e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.196215 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.196240 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.197719 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "35fd3838-aed7-46cf-aefa-df19199dc98e" (UID: "35fd3838-aed7-46cf-aefa-df19199dc98e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.216461 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-dbfcbd88b-69qn5"] Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.218417 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "35fd3838-aed7-46cf-aefa-df19199dc98e" (UID: "35fd3838-aed7-46cf-aefa-df19199dc98e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.223898 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" event={"ID":"35fd3838-aed7-46cf-aefa-df19199dc98e","Type":"ContainerDied","Data":"f4c4ebea0e0cc62f70f59ebcde98ac571b044ff1a0ff48ed2b3fb5bed5264727"} Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.223954 4810 scope.go:117] "RemoveContainer" containerID="68a5d8b615cff419baff4fa7f934f7a2242a8bc7b3b97c2aab2e9ba25573f120" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.224088 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f677dd449-kgbj2" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.232208 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "35fd3838-aed7-46cf-aefa-df19199dc98e" (UID: "35fd3838-aed7-46cf-aefa-df19199dc98e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.234710 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"073cab97-c80f-4905-92df-82b2537ed262","Type":"ContainerStarted","Data":"5e1ffc3bf5219302f8daa00b25be0c838237ca853ad9bafeaa3d4c4829b55773"} Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.248064 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5477779d95-wg2vp" event={"ID":"ba5f8a69-22cf-4196-bc89-ffa60eda9303","Type":"ContainerStarted","Data":"8439e2334c700fa63f3562d0a3892b76d49cb65b3961834e76812a6b483afe47"} Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.248442 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.268287 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v9d27" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.268840 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v9d27" event={"ID":"5b341ba0-f0a3-40f7-a99b-7eb045ba4739","Type":"ContainerDied","Data":"1aefe40b089d6cd4e6a399f6d61e8b50902c24622cbb25efe1aec1863ec62bb6"} Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.268873 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1aefe40b089d6cd4e6a399f6d61e8b50902c24622cbb25efe1aec1863ec62bb6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.269473 4810 scope.go:117] "RemoveContainer" containerID="12c228f4af297ebce306363df7b12d715ab3025c024395e1166eba451956709f" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.298880 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5477779d95-wg2vp" podStartSLOduration=8.298806867 podStartE2EDuration="8.298806867s" podCreationTimestamp="2025-12-01 14:55:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:55:53.27605237 +0000 UTC m=+1319.039561973" watchObservedRunningTime="2025-12-01 14:55:53.298806867 +0000 UTC m=+1319.062316470" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.299709 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.299744 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.299753 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35fd3838-aed7-46cf-aefa-df19199dc98e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.378527 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64dfd64c45-nssxw"] Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.395422 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6c5f549c7d-52fkz"] Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.511868 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-84864d5bd5-wmq86"] Dec 01 14:55:53 crc kubenswrapper[4810]: W1201 14:55:53.512227 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20964a12_27b5_46f7_a9b9_e68b8c7dfce2.slice/crio-7f009842a256ec9e51a6a68edbc02dff2e8fa77481ed53d9a21a09772491ce71 WatchSource:0}: Error finding container 7f009842a256ec9e51a6a68edbc02dff2e8fa77481ed53d9a21a09772491ce71: Status 404 returned error can't find the container with id 7f009842a256ec9e51a6a68edbc02dff2e8fa77481ed53d9a21a09772491ce71 Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.520664 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6d8fb448c4-2hg9g"] Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.538785 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5b4ff7d6cb-m2g6c"] Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.665676 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f677dd449-kgbj2"] Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.682103 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f677dd449-kgbj2"] Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.716626 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-65bbfc7449-rtqm6"] Dec 01 14:55:53 crc kubenswrapper[4810]: E1201 14:55:53.717136 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35fd3838-aed7-46cf-aefa-df19199dc98e" containerName="dnsmasq-dns" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.717165 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="35fd3838-aed7-46cf-aefa-df19199dc98e" containerName="dnsmasq-dns" Dec 01 14:55:53 crc kubenswrapper[4810]: E1201 14:55:53.717186 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35fd3838-aed7-46cf-aefa-df19199dc98e" containerName="init" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.717195 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="35fd3838-aed7-46cf-aefa-df19199dc98e" containerName="init" Dec 01 14:55:53 crc kubenswrapper[4810]: E1201 14:55:53.717223 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b341ba0-f0a3-40f7-a99b-7eb045ba4739" containerName="keystone-bootstrap" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.717231 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b341ba0-f0a3-40f7-a99b-7eb045ba4739" containerName="keystone-bootstrap" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.717444 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="35fd3838-aed7-46cf-aefa-df19199dc98e" containerName="dnsmasq-dns" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.717500 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b341ba0-f0a3-40f7-a99b-7eb045ba4739" containerName="keystone-bootstrap" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.718300 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.722097 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.722307 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.722618 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zlbmq" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.722741 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.722997 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.723096 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.737834 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-65bbfc7449-rtqm6"] Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.808275 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-credential-keys\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.808330 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-internal-tls-certs\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.808359 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-public-tls-certs\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.808391 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-fernet-keys\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.808431 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-combined-ca-bundle\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.808460 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-scripts\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.808515 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7kzl\" (UniqueName: \"kubernetes.io/projected/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-kube-api-access-p7kzl\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.808538 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-config-data\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.913426 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-combined-ca-bundle\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.913502 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-scripts\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.913546 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7kzl\" (UniqueName: \"kubernetes.io/projected/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-kube-api-access-p7kzl\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.913562 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-config-data\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.913608 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-credential-keys\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.913639 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-internal-tls-certs\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.913661 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-public-tls-certs\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.913691 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-fernet-keys\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.921320 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-internal-tls-certs\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.922006 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-public-tls-certs\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.922624 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-combined-ca-bundle\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.928935 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-fernet-keys\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.929321 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7kzl\" (UniqueName: \"kubernetes.io/projected/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-kube-api-access-p7kzl\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.983450 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-scripts\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.986111 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-config-data\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:53 crc kubenswrapper[4810]: I1201 14:55:53.986565 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-credential-keys\") pod \"keystone-65bbfc7449-rtqm6\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.043930 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.281750 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84864d5bd5-wmq86" event={"ID":"129ae854-1430-465f-a177-0c459bd8b4c8","Type":"ContainerStarted","Data":"46e8408cf96c668ecb1af9153b47b16715be23c99fb60a271b3559ab78086e7e"} Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.285983 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" event={"ID":"bdd90cbb-4242-4227-9de4-09f99e60a516","Type":"ContainerStarted","Data":"4a8d673779dd5bbbb9060129ba8fe9ee012663d8194ce9496b1684874e3903dc"} Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.286022 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" event={"ID":"bdd90cbb-4242-4227-9de4-09f99e60a516","Type":"ContainerStarted","Data":"0af6bd36655b389f1090223a090ab3a8d47cace5e570943004f63de4c038edb3"} Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.286035 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" event={"ID":"bdd90cbb-4242-4227-9de4-09f99e60a516","Type":"ContainerStarted","Data":"ea546ae2e75fe338dbf8885d66eb7865f2845a29a179ed939efd251c8116fbf0"} Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.287504 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.287541 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.306434 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6c5f549c7d-52fkz" event={"ID":"31b83b20-83c4-4e06-8eb0-1e1d83d4b160","Type":"ContainerStarted","Data":"c81abf769f22e0110d5df574ceaf9e47a00a5723bf172e5cebaebd7bde8ecee2"} Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.306504 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6c5f549c7d-52fkz" event={"ID":"31b83b20-83c4-4e06-8eb0-1e1d83d4b160","Type":"ContainerStarted","Data":"ba0d20517bfab8cdb0b867e26e4238d3696e54b13598a77e2b77184f47eb47d0"} Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.306519 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6c5f549c7d-52fkz" event={"ID":"31b83b20-83c4-4e06-8eb0-1e1d83d4b160","Type":"ContainerStarted","Data":"0fc0f37ab3b0b1ca3c876a18bef1e6d8eeb42170c39b40db211606adcf278956"} Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.307557 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.307593 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.314818 4810 generic.go:334] "Generic (PLEG): container finished" podID="21371b53-8ea0-40e1-a68e-c140a606b610" containerID="be0f50bda2529e30fd38080d5e23545e84b94561feb08c8bbedbb42f9d764c46" exitCode=0 Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.314940 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" event={"ID":"21371b53-8ea0-40e1-a68e-c140a606b610","Type":"ContainerDied","Data":"be0f50bda2529e30fd38080d5e23545e84b94561feb08c8bbedbb42f9d764c46"} Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.314973 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" event={"ID":"21371b53-8ea0-40e1-a68e-c140a606b610","Type":"ContainerStarted","Data":"30f6aab787fe48b569b9c9989bf6362f5ab52c8c5714076566ad91e6646cfde2"} Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.320827 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" podStartSLOduration=2.320809878 podStartE2EDuration="2.320809878s" podCreationTimestamp="2025-12-01 14:55:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:55:54.306301593 +0000 UTC m=+1320.069811196" watchObservedRunningTime="2025-12-01 14:55:54.320809878 +0000 UTC m=+1320.084319481" Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.323608 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dbfcbd88b-69qn5" event={"ID":"66f9e303-31b7-4355-9acd-2720e03a9bb5","Type":"ContainerStarted","Data":"20703cc0174712162d24e0139f50da7302d7a2f19242ad7070a8af3f01b4eee9"} Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.323678 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dbfcbd88b-69qn5" event={"ID":"66f9e303-31b7-4355-9acd-2720e03a9bb5","Type":"ContainerStarted","Data":"8b9b17c744a8a4fd26bfc8645930c1dcc44173d2d0ba4b4bd66fc6ac3e2a264c"} Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.323692 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dbfcbd88b-69qn5" event={"ID":"66f9e303-31b7-4355-9acd-2720e03a9bb5","Type":"ContainerStarted","Data":"f1d52a4c1bc52513c64a791d489093557207eacf26bd67cf8bac38b3ddbe4b49"} Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.326726 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" event={"ID":"20964a12-27b5-46f7-a9b9-e68b8c7dfce2","Type":"ContainerStarted","Data":"7f009842a256ec9e51a6a68edbc02dff2e8fa77481ed53d9a21a09772491ce71"} Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.332588 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.348859 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.349179 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.353884 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6c5f549c7d-52fkz" podStartSLOduration=5.353866665 podStartE2EDuration="5.353866665s" podCreationTimestamp="2025-12-01 14:55:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:55:54.329299129 +0000 UTC m=+1320.092808732" watchObservedRunningTime="2025-12-01 14:55:54.353866665 +0000 UTC m=+1320.117376268" Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.382938 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-dbfcbd88b-69qn5" podStartSLOduration=5.382917335 podStartE2EDuration="5.382917335s" podCreationTimestamp="2025-12-01 14:55:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:55:54.378757922 +0000 UTC m=+1320.142267525" watchObservedRunningTime="2025-12-01 14:55:54.382917335 +0000 UTC m=+1320.146426938" Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.614850 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35fd3838-aed7-46cf-aefa-df19199dc98e" path="/var/lib/kubelet/pods/35fd3838-aed7-46cf-aefa-df19199dc98e/volumes" Dec 01 14:55:54 crc kubenswrapper[4810]: I1201 14:55:54.709613 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-65bbfc7449-rtqm6"] Dec 01 14:55:55 crc kubenswrapper[4810]: I1201 14:55:55.022694 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 14:55:55 crc kubenswrapper[4810]: I1201 14:55:55.155111 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 14:55:55 crc kubenswrapper[4810]: I1201 14:55:55.155228 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 14:55:55 crc kubenswrapper[4810]: I1201 14:55:55.155851 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:55:55 crc kubenswrapper[4810]: I1201 14:55:55.157385 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 14:55:55 crc kubenswrapper[4810]: I1201 14:55:55.360039 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-65bbfc7449-rtqm6" event={"ID":"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0","Type":"ContainerStarted","Data":"0f78ed302304f1a4ae276cf04f45b339eb2e16c4d7fe8c01f8a3c129526d347d"} Dec 01 14:55:55 crc kubenswrapper[4810]: I1201 14:55:55.360318 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 14:55:55 crc kubenswrapper[4810]: I1201 14:55:55.394542 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 14:55:56 crc kubenswrapper[4810]: I1201 14:55:56.407025 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-65bbfc7449-rtqm6" event={"ID":"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0","Type":"ContainerStarted","Data":"e064be1964d87f2e16398d0a402ce3f57b2caf18c89d6b179f3ef2a94e1374ac"} Dec 01 14:55:56 crc kubenswrapper[4810]: I1201 14:55:56.407324 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:55:56 crc kubenswrapper[4810]: I1201 14:55:56.411991 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" event={"ID":"21371b53-8ea0-40e1-a68e-c140a606b610","Type":"ContainerStarted","Data":"78d365d0d6eb8ecf1b2484dc5acf8196acada7b43e9907e6fc8e57c44892e66e"} Dec 01 14:55:56 crc kubenswrapper[4810]: I1201 14:55:56.412060 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:55:56 crc kubenswrapper[4810]: I1201 14:55:56.445751 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-65bbfc7449-rtqm6" podStartSLOduration=3.445734166 podStartE2EDuration="3.445734166s" podCreationTimestamp="2025-12-01 14:55:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:55:56.429591078 +0000 UTC m=+1322.193100691" watchObservedRunningTime="2025-12-01 14:55:56.445734166 +0000 UTC m=+1322.209243769" Dec 01 14:55:56 crc kubenswrapper[4810]: I1201 14:55:56.466456 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" podStartSLOduration=7.466436219 podStartE2EDuration="7.466436219s" podCreationTimestamp="2025-12-01 14:55:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:55:56.458816272 +0000 UTC m=+1322.222325895" watchObservedRunningTime="2025-12-01 14:55:56.466436219 +0000 UTC m=+1322.229945822" Dec 01 14:55:57 crc kubenswrapper[4810]: I1201 14:55:57.420818 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84864d5bd5-wmq86" event={"ID":"129ae854-1430-465f-a177-0c459bd8b4c8","Type":"ContainerStarted","Data":"8c00c76d0791b2c2bc8a97328993ef7e1e8aa6997c54b9aa8cb7c3385b227dd7"} Dec 01 14:55:57 crc kubenswrapper[4810]: I1201 14:55:57.422414 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84864d5bd5-wmq86" event={"ID":"129ae854-1430-465f-a177-0c459bd8b4c8","Type":"ContainerStarted","Data":"b8a19753910ff67350d9da738f60efda1e12ea5411279e28d972ad0cb28c8f79"} Dec 01 14:55:57 crc kubenswrapper[4810]: I1201 14:55:57.425615 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vlw7c" event={"ID":"09861728-4533-4ddf-a32d-3cb1b70fc325","Type":"ContainerStarted","Data":"e06e409cdd63d5c7f2fe85cf645dd9d82ceaa580d3ea7ea520c851fddaa5d10d"} Dec 01 14:55:57 crc kubenswrapper[4810]: I1201 14:55:57.465440 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-84864d5bd5-wmq86" podStartSLOduration=5.81492747 podStartE2EDuration="8.465423905s" podCreationTimestamp="2025-12-01 14:55:49 +0000 UTC" firstStartedPulling="2025-12-01 14:55:53.501100142 +0000 UTC m=+1319.264609745" lastFinishedPulling="2025-12-01 14:55:56.151596577 +0000 UTC m=+1321.915106180" observedRunningTime="2025-12-01 14:55:57.435379848 +0000 UTC m=+1323.198889451" watchObservedRunningTime="2025-12-01 14:55:57.465423905 +0000 UTC m=+1323.228933508" Dec 01 14:55:57 crc kubenswrapper[4810]: I1201 14:55:57.481908 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-vlw7c" podStartSLOduration=3.8742743060000002 podStartE2EDuration="39.481887231s" podCreationTimestamp="2025-12-01 14:55:18 +0000 UTC" firstStartedPulling="2025-12-01 14:55:20.507723327 +0000 UTC m=+1286.271232930" lastFinishedPulling="2025-12-01 14:55:56.115336252 +0000 UTC m=+1321.878845855" observedRunningTime="2025-12-01 14:55:57.455095943 +0000 UTC m=+1323.218605546" watchObservedRunningTime="2025-12-01 14:55:57.481887231 +0000 UTC m=+1323.245396834" Dec 01 14:55:58 crc kubenswrapper[4810]: I1201 14:55:58.439103 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" event={"ID":"20964a12-27b5-46f7-a9b9-e68b8c7dfce2","Type":"ContainerStarted","Data":"a40c19049624e61426f700a570719d59ec7a3b73107e4fbc2d3c42482e3793b0"} Dec 01 14:55:58 crc kubenswrapper[4810]: I1201 14:55:58.439631 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" event={"ID":"20964a12-27b5-46f7-a9b9-e68b8c7dfce2","Type":"ContainerStarted","Data":"5a5b7334cdb5668aa597f83c43e1f06755c33b97ba720234d356c46686f469f6"} Dec 01 14:55:58 crc kubenswrapper[4810]: I1201 14:55:58.462730 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" podStartSLOduration=5.790415563 podStartE2EDuration="9.462710163s" podCreationTimestamp="2025-12-01 14:55:49 +0000 UTC" firstStartedPulling="2025-12-01 14:55:53.5625012 +0000 UTC m=+1319.326010803" lastFinishedPulling="2025-12-01 14:55:57.2347958 +0000 UTC m=+1322.998305403" observedRunningTime="2025-12-01 14:55:58.453076522 +0000 UTC m=+1324.216586145" watchObservedRunningTime="2025-12-01 14:55:58.462710163 +0000 UTC m=+1324.226219766" Dec 01 14:56:00 crc kubenswrapper[4810]: I1201 14:56:00.132249 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:56:00 crc kubenswrapper[4810]: I1201 14:56:00.211023 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f645789c-lshtm"] Dec 01 14:56:00 crc kubenswrapper[4810]: I1201 14:56:00.211519 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f645789c-lshtm" podUID="59e1bba9-e680-453b-bb4d-9f7f202551eb" containerName="dnsmasq-dns" containerID="cri-o://e0deec04a2b1f5902cc569561667e0c9f848a4cc54bd1aa77303e35ec109d9b5" gracePeriod=10 Dec 01 14:56:01 crc kubenswrapper[4810]: I1201 14:56:01.470816 4810 generic.go:334] "Generic (PLEG): container finished" podID="59e1bba9-e680-453b-bb4d-9f7f202551eb" containerID="e0deec04a2b1f5902cc569561667e0c9f848a4cc54bd1aa77303e35ec109d9b5" exitCode=0 Dec 01 14:56:01 crc kubenswrapper[4810]: I1201 14:56:01.471101 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f645789c-lshtm" event={"ID":"59e1bba9-e680-453b-bb4d-9f7f202551eb","Type":"ContainerDied","Data":"e0deec04a2b1f5902cc569561667e0c9f848a4cc54bd1aa77303e35ec109d9b5"} Dec 01 14:56:01 crc kubenswrapper[4810]: I1201 14:56:01.632186 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:56:01 crc kubenswrapper[4810]: I1201 14:56:01.984767 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:56:03 crc kubenswrapper[4810]: I1201 14:56:03.503207 4810 generic.go:334] "Generic (PLEG): container finished" podID="09861728-4533-4ddf-a32d-3cb1b70fc325" containerID="e06e409cdd63d5c7f2fe85cf645dd9d82ceaa580d3ea7ea520c851fddaa5d10d" exitCode=0 Dec 01 14:56:03 crc kubenswrapper[4810]: I1201 14:56:03.503436 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vlw7c" event={"ID":"09861728-4533-4ddf-a32d-3cb1b70fc325","Type":"ContainerDied","Data":"e06e409cdd63d5c7f2fe85cf645dd9d82ceaa580d3ea7ea520c851fddaa5d10d"} Dec 01 14:56:03 crc kubenswrapper[4810]: I1201 14:56:03.883485 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:56:03 crc kubenswrapper[4810]: I1201 14:56:03.961373 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:56:03 crc kubenswrapper[4810]: E1201 14:56:03.986864 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="073cab97-c80f-4905-92df-82b2537ed262" Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.029470 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-dbfcbd88b-69qn5"] Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.029791 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-dbfcbd88b-69qn5" podUID="66f9e303-31b7-4355-9acd-2720e03a9bb5" containerName="barbican-api-log" containerID="cri-o://8b9b17c744a8a4fd26bfc8645930c1dcc44173d2d0ba4b4bd66fc6ac3e2a264c" gracePeriod=30 Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.030161 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-dbfcbd88b-69qn5" podUID="66f9e303-31b7-4355-9acd-2720e03a9bb5" containerName="barbican-api" containerID="cri-o://20703cc0174712162d24e0139f50da7302d7a2f19242ad7070a8af3f01b4eee9" gracePeriod=30 Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.086376 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.237323 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-dns-svc\") pod \"59e1bba9-e680-453b-bb4d-9f7f202551eb\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.237451 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwl68\" (UniqueName: \"kubernetes.io/projected/59e1bba9-e680-453b-bb4d-9f7f202551eb-kube-api-access-hwl68\") pod \"59e1bba9-e680-453b-bb4d-9f7f202551eb\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.237519 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-ovsdbserver-sb\") pod \"59e1bba9-e680-453b-bb4d-9f7f202551eb\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.237548 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-dns-swift-storage-0\") pod \"59e1bba9-e680-453b-bb4d-9f7f202551eb\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.237576 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-ovsdbserver-nb\") pod \"59e1bba9-e680-453b-bb4d-9f7f202551eb\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.237628 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-config\") pod \"59e1bba9-e680-453b-bb4d-9f7f202551eb\" (UID: \"59e1bba9-e680-453b-bb4d-9f7f202551eb\") " Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.244190 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59e1bba9-e680-453b-bb4d-9f7f202551eb-kube-api-access-hwl68" (OuterVolumeSpecName: "kube-api-access-hwl68") pod "59e1bba9-e680-453b-bb4d-9f7f202551eb" (UID: "59e1bba9-e680-453b-bb4d-9f7f202551eb"). InnerVolumeSpecName "kube-api-access-hwl68". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.288393 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "59e1bba9-e680-453b-bb4d-9f7f202551eb" (UID: "59e1bba9-e680-453b-bb4d-9f7f202551eb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.291018 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "59e1bba9-e680-453b-bb4d-9f7f202551eb" (UID: "59e1bba9-e680-453b-bb4d-9f7f202551eb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.292944 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "59e1bba9-e680-453b-bb4d-9f7f202551eb" (UID: "59e1bba9-e680-453b-bb4d-9f7f202551eb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.303267 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-config" (OuterVolumeSpecName: "config") pod "59e1bba9-e680-453b-bb4d-9f7f202551eb" (UID: "59e1bba9-e680-453b-bb4d-9f7f202551eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.311015 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "59e1bba9-e680-453b-bb4d-9f7f202551eb" (UID: "59e1bba9-e680-453b-bb4d-9f7f202551eb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.340245 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.340287 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.340301 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwl68\" (UniqueName: \"kubernetes.io/projected/59e1bba9-e680-453b-bb4d-9f7f202551eb-kube-api-access-hwl68\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.340318 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.340331 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.340368 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59e1bba9-e680-453b-bb4d-9f7f202551eb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.519605 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f645789c-lshtm" event={"ID":"59e1bba9-e680-453b-bb4d-9f7f202551eb","Type":"ContainerDied","Data":"8ea87e13c69593db6ceeac2aa9dbf2944c96f815b6b0ad3fbafd9cc86bbed27a"} Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.519703 4810 scope.go:117] "RemoveContainer" containerID="e0deec04a2b1f5902cc569561667e0c9f848a4cc54bd1aa77303e35ec109d9b5" Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.519633 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f645789c-lshtm" Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.538850 4810 scope.go:117] "RemoveContainer" containerID="655da5f31bd45ec5aae36c4aec6d16ce77eb5e9fdbbd4c24b84f240e54627620" Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.539097 4810 generic.go:334] "Generic (PLEG): container finished" podID="66f9e303-31b7-4355-9acd-2720e03a9bb5" containerID="8b9b17c744a8a4fd26bfc8645930c1dcc44173d2d0ba4b4bd66fc6ac3e2a264c" exitCode=143 Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.539168 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dbfcbd88b-69qn5" event={"ID":"66f9e303-31b7-4355-9acd-2720e03a9bb5","Type":"ContainerDied","Data":"8b9b17c744a8a4fd26bfc8645930c1dcc44173d2d0ba4b4bd66fc6ac3e2a264c"} Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.565876 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"073cab97-c80f-4905-92df-82b2537ed262","Type":"ContainerStarted","Data":"66dab48997486cb3696b8a9bf3f8feab00e21edf9ee7e1b08337e550fde3b47c"} Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.566081 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.566091 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="073cab97-c80f-4905-92df-82b2537ed262" containerName="ceilometer-notification-agent" containerID="cri-o://1d39fda5deb853dc5f22a662445df2f289042a4470edfdd81faeef5d2732a791" gracePeriod=30 Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.566229 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="073cab97-c80f-4905-92df-82b2537ed262" containerName="proxy-httpd" containerID="cri-o://66dab48997486cb3696b8a9bf3f8feab00e21edf9ee7e1b08337e550fde3b47c" gracePeriod=30 Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.566289 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="073cab97-c80f-4905-92df-82b2537ed262" containerName="sg-core" containerID="cri-o://5e1ffc3bf5219302f8daa00b25be0c838237ca853ad9bafeaa3d4c4829b55773" gracePeriod=30 Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.576073 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f645789c-lshtm"] Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.587189 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f645789c-lshtm"] Dec 01 14:56:04 crc kubenswrapper[4810]: I1201 14:56:04.925954 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.052455 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-config-data\") pod \"09861728-4533-4ddf-a32d-3cb1b70fc325\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.052564 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-combined-ca-bundle\") pod \"09861728-4533-4ddf-a32d-3cb1b70fc325\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.052633 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/09861728-4533-4ddf-a32d-3cb1b70fc325-etc-machine-id\") pod \"09861728-4533-4ddf-a32d-3cb1b70fc325\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.052670 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-scripts\") pod \"09861728-4533-4ddf-a32d-3cb1b70fc325\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.052705 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngs2b\" (UniqueName: \"kubernetes.io/projected/09861728-4533-4ddf-a32d-3cb1b70fc325-kube-api-access-ngs2b\") pod \"09861728-4533-4ddf-a32d-3cb1b70fc325\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.052813 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-db-sync-config-data\") pod \"09861728-4533-4ddf-a32d-3cb1b70fc325\" (UID: \"09861728-4533-4ddf-a32d-3cb1b70fc325\") " Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.053016 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09861728-4533-4ddf-a32d-3cb1b70fc325-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "09861728-4533-4ddf-a32d-3cb1b70fc325" (UID: "09861728-4533-4ddf-a32d-3cb1b70fc325"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.053270 4810 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/09861728-4533-4ddf-a32d-3cb1b70fc325-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.058749 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "09861728-4533-4ddf-a32d-3cb1b70fc325" (UID: "09861728-4533-4ddf-a32d-3cb1b70fc325"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.059423 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-scripts" (OuterVolumeSpecName: "scripts") pod "09861728-4533-4ddf-a32d-3cb1b70fc325" (UID: "09861728-4533-4ddf-a32d-3cb1b70fc325"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.078666 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09861728-4533-4ddf-a32d-3cb1b70fc325-kube-api-access-ngs2b" (OuterVolumeSpecName: "kube-api-access-ngs2b") pod "09861728-4533-4ddf-a32d-3cb1b70fc325" (UID: "09861728-4533-4ddf-a32d-3cb1b70fc325"). InnerVolumeSpecName "kube-api-access-ngs2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.102606 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "09861728-4533-4ddf-a32d-3cb1b70fc325" (UID: "09861728-4533-4ddf-a32d-3cb1b70fc325"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.117437 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-config-data" (OuterVolumeSpecName: "config-data") pod "09861728-4533-4ddf-a32d-3cb1b70fc325" (UID: "09861728-4533-4ddf-a32d-3cb1b70fc325"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.154758 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.154794 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.154959 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngs2b\" (UniqueName: \"kubernetes.io/projected/09861728-4533-4ddf-a32d-3cb1b70fc325-kube-api-access-ngs2b\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.154994 4810 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.155013 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09861728-4533-4ddf-a32d-3cb1b70fc325-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.587953 4810 generic.go:334] "Generic (PLEG): container finished" podID="073cab97-c80f-4905-92df-82b2537ed262" containerID="66dab48997486cb3696b8a9bf3f8feab00e21edf9ee7e1b08337e550fde3b47c" exitCode=0 Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.587987 4810 generic.go:334] "Generic (PLEG): container finished" podID="073cab97-c80f-4905-92df-82b2537ed262" containerID="5e1ffc3bf5219302f8daa00b25be0c838237ca853ad9bafeaa3d4c4829b55773" exitCode=2 Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.587993 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"073cab97-c80f-4905-92df-82b2537ed262","Type":"ContainerDied","Data":"66dab48997486cb3696b8a9bf3f8feab00e21edf9ee7e1b08337e550fde3b47c"} Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.588033 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"073cab97-c80f-4905-92df-82b2537ed262","Type":"ContainerDied","Data":"5e1ffc3bf5219302f8daa00b25be0c838237ca853ad9bafeaa3d4c4829b55773"} Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.592196 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vlw7c" event={"ID":"09861728-4533-4ddf-a32d-3cb1b70fc325","Type":"ContainerDied","Data":"58a06685d2101791f5bea371d664d3b80d561952217772fa304642ba18a8665d"} Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.592222 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58a06685d2101791f5bea371d664d3b80d561952217772fa304642ba18a8665d" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.592288 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vlw7c" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.815799 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8ccb5c7cf-r4dnc"] Dec 01 14:56:05 crc kubenswrapper[4810]: E1201 14:56:05.816167 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59e1bba9-e680-453b-bb4d-9f7f202551eb" containerName="init" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.816183 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="59e1bba9-e680-453b-bb4d-9f7f202551eb" containerName="init" Dec 01 14:56:05 crc kubenswrapper[4810]: E1201 14:56:05.816195 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59e1bba9-e680-453b-bb4d-9f7f202551eb" containerName="dnsmasq-dns" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.816201 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="59e1bba9-e680-453b-bb4d-9f7f202551eb" containerName="dnsmasq-dns" Dec 01 14:56:05 crc kubenswrapper[4810]: E1201 14:56:05.816224 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09861728-4533-4ddf-a32d-3cb1b70fc325" containerName="cinder-db-sync" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.816232 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="09861728-4533-4ddf-a32d-3cb1b70fc325" containerName="cinder-db-sync" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.816401 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="09861728-4533-4ddf-a32d-3cb1b70fc325" containerName="cinder-db-sync" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.816420 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="59e1bba9-e680-453b-bb4d-9f7f202551eb" containerName="dnsmasq-dns" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.817321 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.844612 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8ccb5c7cf-r4dnc"] Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.860408 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.861944 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.867256 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-8q2gm" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.867629 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.867763 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.867884 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.894611 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.980419 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-ovsdbserver-sb\") pod \"dnsmasq-dns-8ccb5c7cf-r4dnc\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.980634 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-scripts\") pod \"cinder-scheduler-0\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.980711 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-dns-svc\") pod \"dnsmasq-dns-8ccb5c7cf-r4dnc\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.980778 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-ovsdbserver-nb\") pod \"dnsmasq-dns-8ccb5c7cf-r4dnc\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.980900 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.980973 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.981063 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv6d5\" (UniqueName: \"kubernetes.io/projected/42ccc7b5-8af0-418b-8ca7-5a05095280c6-kube-api-access-nv6d5\") pod \"cinder-scheduler-0\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.981169 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-dns-swift-storage-0\") pod \"dnsmasq-dns-8ccb5c7cf-r4dnc\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.981246 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh8rk\" (UniqueName: \"kubernetes.io/projected/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-kube-api-access-zh8rk\") pod \"dnsmasq-dns-8ccb5c7cf-r4dnc\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.981312 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-config\") pod \"dnsmasq-dns-8ccb5c7cf-r4dnc\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.981370 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-config-data\") pod \"cinder-scheduler-0\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:05 crc kubenswrapper[4810]: I1201 14:56:05.981439 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/42ccc7b5-8af0-418b-8ca7-5a05095280c6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.043089 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.045417 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.047673 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.057560 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.084023 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv6d5\" (UniqueName: \"kubernetes.io/projected/42ccc7b5-8af0-418b-8ca7-5a05095280c6-kube-api-access-nv6d5\") pod \"cinder-scheduler-0\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.084290 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-config-data\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.084459 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-scripts\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.084631 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-dns-swift-storage-0\") pod \"dnsmasq-dns-8ccb5c7cf-r4dnc\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.084731 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh8rk\" (UniqueName: \"kubernetes.io/projected/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-kube-api-access-zh8rk\") pod \"dnsmasq-dns-8ccb5c7cf-r4dnc\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.084801 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.084875 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-config-data\") pod \"cinder-scheduler-0\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.084949 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-config\") pod \"dnsmasq-dns-8ccb5c7cf-r4dnc\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.085040 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/42ccc7b5-8af0-418b-8ca7-5a05095280c6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.085165 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-ovsdbserver-sb\") pod \"dnsmasq-dns-8ccb5c7cf-r4dnc\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.085299 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa46114f-d15e-416f-9085-7ada02bfbfdf-logs\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.085409 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-scripts\") pod \"cinder-scheduler-0\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.085530 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzgtd\" (UniqueName: \"kubernetes.io/projected/fa46114f-d15e-416f-9085-7ada02bfbfdf-kube-api-access-kzgtd\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.085631 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-dns-svc\") pod \"dnsmasq-dns-8ccb5c7cf-r4dnc\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.085704 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-config-data-custom\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.085779 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-ovsdbserver-nb\") pod \"dnsmasq-dns-8ccb5c7cf-r4dnc\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.085905 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fa46114f-d15e-416f-9085-7ada02bfbfdf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.086029 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.086145 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.086905 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-ovsdbserver-sb\") pod \"dnsmasq-dns-8ccb5c7cf-r4dnc\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.087068 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-dns-svc\") pod \"dnsmasq-dns-8ccb5c7cf-r4dnc\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.087430 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-dns-swift-storage-0\") pod \"dnsmasq-dns-8ccb5c7cf-r4dnc\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.087808 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-ovsdbserver-nb\") pod \"dnsmasq-dns-8ccb5c7cf-r4dnc\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.087865 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/42ccc7b5-8af0-418b-8ca7-5a05095280c6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.088033 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-config\") pod \"dnsmasq-dns-8ccb5c7cf-r4dnc\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.094683 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.102954 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.106916 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv6d5\" (UniqueName: \"kubernetes.io/projected/42ccc7b5-8af0-418b-8ca7-5a05095280c6-kube-api-access-nv6d5\") pod \"cinder-scheduler-0\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.110219 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh8rk\" (UniqueName: \"kubernetes.io/projected/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-kube-api-access-zh8rk\") pod \"dnsmasq-dns-8ccb5c7cf-r4dnc\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.111035 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-scripts\") pod \"cinder-scheduler-0\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.119612 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-config-data\") pod \"cinder-scheduler-0\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.145742 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.187316 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-config-data\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.187354 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-scripts\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.187396 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.187447 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa46114f-d15e-416f-9085-7ada02bfbfdf-logs\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.187463 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzgtd\" (UniqueName: \"kubernetes.io/projected/fa46114f-d15e-416f-9085-7ada02bfbfdf-kube-api-access-kzgtd\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.187549 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-config-data-custom\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.187581 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fa46114f-d15e-416f-9085-7ada02bfbfdf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.187666 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fa46114f-d15e-416f-9085-7ada02bfbfdf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.188189 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa46114f-d15e-416f-9085-7ada02bfbfdf-logs\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.189716 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.190533 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-scripts\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.191021 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-config-data\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.191432 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.193484 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-config-data-custom\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.207239 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzgtd\" (UniqueName: \"kubernetes.io/projected/fa46114f-d15e-416f-9085-7ada02bfbfdf-kube-api-access-kzgtd\") pod \"cinder-api-0\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.479407 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.512643 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59e1bba9-e680-453b-bb4d-9f7f202551eb" path="/var/lib/kubelet/pods/59e1bba9-e680-453b-bb4d-9f7f202551eb/volumes" Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.672364 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 14:56:06 crc kubenswrapper[4810]: W1201 14:56:06.688145 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42ccc7b5_8af0_418b_8ca7_5a05095280c6.slice/crio-45f1c50adfd43e0b65d03a64d89ec867ca10758a043fbb694ccbcdb701225af9 WatchSource:0}: Error finding container 45f1c50adfd43e0b65d03a64d89ec867ca10758a043fbb694ccbcdb701225af9: Status 404 returned error can't find the container with id 45f1c50adfd43e0b65d03a64d89ec867ca10758a043fbb694ccbcdb701225af9 Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.774367 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8ccb5c7cf-r4dnc"] Dec 01 14:56:06 crc kubenswrapper[4810]: W1201 14:56:06.779172 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ef69bb1_54d2_47b5_bcbd_5d0f3b60d2ed.slice/crio-317bfd153a31d9dcdad645a673394f82bb72293881221f6bce7bc109f6db70fc WatchSource:0}: Error finding container 317bfd153a31d9dcdad645a673394f82bb72293881221f6bce7bc109f6db70fc: Status 404 returned error can't find the container with id 317bfd153a31d9dcdad645a673394f82bb72293881221f6bce7bc109f6db70fc Dec 01 14:56:06 crc kubenswrapper[4810]: I1201 14:56:06.968541 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 14:56:07 crc kubenswrapper[4810]: W1201 14:56:07.017610 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa46114f_d15e_416f_9085_7ada02bfbfdf.slice/crio-08dcdb366ccc5bc992a72e19d7b98460b530f9f7b073f457b60f7aa7d6f081f9 WatchSource:0}: Error finding container 08dcdb366ccc5bc992a72e19d7b98460b530f9f7b073f457b60f7aa7d6f081f9: Status 404 returned error can't find the container with id 08dcdb366ccc5bc992a72e19d7b98460b530f9f7b073f457b60f7aa7d6f081f9 Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.189683 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-dbfcbd88b-69qn5" podUID="66f9e303-31b7-4355-9acd-2720e03a9bb5" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": read tcp 10.217.0.2:50246->10.217.0.154:9311: read: connection reset by peer" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.189731 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-dbfcbd88b-69qn5" podUID="66f9e303-31b7-4355-9acd-2720e03a9bb5" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": read tcp 10.217.0.2:50248->10.217.0.154:9311: read: connection reset by peer" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.279150 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.320619 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/073cab97-c80f-4905-92df-82b2537ed262-log-httpd\") pod \"073cab97-c80f-4905-92df-82b2537ed262\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.320690 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-sg-core-conf-yaml\") pod \"073cab97-c80f-4905-92df-82b2537ed262\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.320746 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-config-data\") pod \"073cab97-c80f-4905-92df-82b2537ed262\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.320792 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/073cab97-c80f-4905-92df-82b2537ed262-run-httpd\") pod \"073cab97-c80f-4905-92df-82b2537ed262\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.320820 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfp2z\" (UniqueName: \"kubernetes.io/projected/073cab97-c80f-4905-92df-82b2537ed262-kube-api-access-vfp2z\") pod \"073cab97-c80f-4905-92df-82b2537ed262\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.320849 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-scripts\") pod \"073cab97-c80f-4905-92df-82b2537ed262\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.320892 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-combined-ca-bundle\") pod \"073cab97-c80f-4905-92df-82b2537ed262\" (UID: \"073cab97-c80f-4905-92df-82b2537ed262\") " Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.321704 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/073cab97-c80f-4905-92df-82b2537ed262-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "073cab97-c80f-4905-92df-82b2537ed262" (UID: "073cab97-c80f-4905-92df-82b2537ed262"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.323907 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/073cab97-c80f-4905-92df-82b2537ed262-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "073cab97-c80f-4905-92df-82b2537ed262" (UID: "073cab97-c80f-4905-92df-82b2537ed262"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.327360 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-scripts" (OuterVolumeSpecName: "scripts") pod "073cab97-c80f-4905-92df-82b2537ed262" (UID: "073cab97-c80f-4905-92df-82b2537ed262"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.351294 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/073cab97-c80f-4905-92df-82b2537ed262-kube-api-access-vfp2z" (OuterVolumeSpecName: "kube-api-access-vfp2z") pod "073cab97-c80f-4905-92df-82b2537ed262" (UID: "073cab97-c80f-4905-92df-82b2537ed262"). InnerVolumeSpecName "kube-api-access-vfp2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.377434 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "073cab97-c80f-4905-92df-82b2537ed262" (UID: "073cab97-c80f-4905-92df-82b2537ed262"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.399248 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "073cab97-c80f-4905-92df-82b2537ed262" (UID: "073cab97-c80f-4905-92df-82b2537ed262"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.421685 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-config-data" (OuterVolumeSpecName: "config-data") pod "073cab97-c80f-4905-92df-82b2537ed262" (UID: "073cab97-c80f-4905-92df-82b2537ed262"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.423134 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/073cab97-c80f-4905-92df-82b2537ed262-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.423160 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.423172 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.423184 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/073cab97-c80f-4905-92df-82b2537ed262-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.423196 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfp2z\" (UniqueName: \"kubernetes.io/projected/073cab97-c80f-4905-92df-82b2537ed262-kube-api-access-vfp2z\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.423208 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.423219 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/073cab97-c80f-4905-92df-82b2537ed262-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.532679 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.616239 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fa46114f-d15e-416f-9085-7ada02bfbfdf","Type":"ContainerStarted","Data":"08dcdb366ccc5bc992a72e19d7b98460b530f9f7b073f457b60f7aa7d6f081f9"} Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.617243 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"42ccc7b5-8af0-418b-8ca7-5a05095280c6","Type":"ContainerStarted","Data":"45f1c50adfd43e0b65d03a64d89ec867ca10758a043fbb694ccbcdb701225af9"} Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.618973 4810 generic.go:334] "Generic (PLEG): container finished" podID="66f9e303-31b7-4355-9acd-2720e03a9bb5" containerID="20703cc0174712162d24e0139f50da7302d7a2f19242ad7070a8af3f01b4eee9" exitCode=0 Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.619028 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dbfcbd88b-69qn5" event={"ID":"66f9e303-31b7-4355-9acd-2720e03a9bb5","Type":"ContainerDied","Data":"20703cc0174712162d24e0139f50da7302d7a2f19242ad7070a8af3f01b4eee9"} Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.619029 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-dbfcbd88b-69qn5" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.619047 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dbfcbd88b-69qn5" event={"ID":"66f9e303-31b7-4355-9acd-2720e03a9bb5","Type":"ContainerDied","Data":"f1d52a4c1bc52513c64a791d489093557207eacf26bd67cf8bac38b3ddbe4b49"} Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.619065 4810 scope.go:117] "RemoveContainer" containerID="20703cc0174712162d24e0139f50da7302d7a2f19242ad7070a8af3f01b4eee9" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.625644 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66f9e303-31b7-4355-9acd-2720e03a9bb5-logs\") pod \"66f9e303-31b7-4355-9acd-2720e03a9bb5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.625688 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66f9e303-31b7-4355-9acd-2720e03a9bb5-combined-ca-bundle\") pod \"66f9e303-31b7-4355-9acd-2720e03a9bb5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.625820 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66f9e303-31b7-4355-9acd-2720e03a9bb5-config-data\") pod \"66f9e303-31b7-4355-9acd-2720e03a9bb5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.625901 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdbjs\" (UniqueName: \"kubernetes.io/projected/66f9e303-31b7-4355-9acd-2720e03a9bb5-kube-api-access-gdbjs\") pod \"66f9e303-31b7-4355-9acd-2720e03a9bb5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.625947 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66f9e303-31b7-4355-9acd-2720e03a9bb5-config-data-custom\") pod \"66f9e303-31b7-4355-9acd-2720e03a9bb5\" (UID: \"66f9e303-31b7-4355-9acd-2720e03a9bb5\") " Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.627001 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66f9e303-31b7-4355-9acd-2720e03a9bb5-logs" (OuterVolumeSpecName: "logs") pod "66f9e303-31b7-4355-9acd-2720e03a9bb5" (UID: "66f9e303-31b7-4355-9acd-2720e03a9bb5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.628422 4810 generic.go:334] "Generic (PLEG): container finished" podID="073cab97-c80f-4905-92df-82b2537ed262" containerID="1d39fda5deb853dc5f22a662445df2f289042a4470edfdd81faeef5d2732a791" exitCode=0 Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.628511 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"073cab97-c80f-4905-92df-82b2537ed262","Type":"ContainerDied","Data":"1d39fda5deb853dc5f22a662445df2f289042a4470edfdd81faeef5d2732a791"} Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.628537 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"073cab97-c80f-4905-92df-82b2537ed262","Type":"ContainerDied","Data":"3630a694a7b3b317b18e38d6907ec6c0512366643411e5d75feb5115e7d5f499"} Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.628627 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.631102 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66f9e303-31b7-4355-9acd-2720e03a9bb5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "66f9e303-31b7-4355-9acd-2720e03a9bb5" (UID: "66f9e303-31b7-4355-9acd-2720e03a9bb5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.631175 4810 generic.go:334] "Generic (PLEG): container finished" podID="4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed" containerID="0ba04bb2acced6ad016122958c11c701353a9196ca7ece5be1e1a74b3b65fc40" exitCode=0 Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.631216 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" event={"ID":"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed","Type":"ContainerDied","Data":"0ba04bb2acced6ad016122958c11c701353a9196ca7ece5be1e1a74b3b65fc40"} Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.631248 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" event={"ID":"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed","Type":"ContainerStarted","Data":"317bfd153a31d9dcdad645a673394f82bb72293881221f6bce7bc109f6db70fc"} Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.634712 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66f9e303-31b7-4355-9acd-2720e03a9bb5-kube-api-access-gdbjs" (OuterVolumeSpecName: "kube-api-access-gdbjs") pod "66f9e303-31b7-4355-9acd-2720e03a9bb5" (UID: "66f9e303-31b7-4355-9acd-2720e03a9bb5"). InnerVolumeSpecName "kube-api-access-gdbjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.665097 4810 scope.go:117] "RemoveContainer" containerID="8b9b17c744a8a4fd26bfc8645930c1dcc44173d2d0ba4b4bd66fc6ac3e2a264c" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.692884 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66f9e303-31b7-4355-9acd-2720e03a9bb5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66f9e303-31b7-4355-9acd-2720e03a9bb5" (UID: "66f9e303-31b7-4355-9acd-2720e03a9bb5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.693441 4810 scope.go:117] "RemoveContainer" containerID="20703cc0174712162d24e0139f50da7302d7a2f19242ad7070a8af3f01b4eee9" Dec 01 14:56:07 crc kubenswrapper[4810]: E1201 14:56:07.696355 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20703cc0174712162d24e0139f50da7302d7a2f19242ad7070a8af3f01b4eee9\": container with ID starting with 20703cc0174712162d24e0139f50da7302d7a2f19242ad7070a8af3f01b4eee9 not found: ID does not exist" containerID="20703cc0174712162d24e0139f50da7302d7a2f19242ad7070a8af3f01b4eee9" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.696402 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20703cc0174712162d24e0139f50da7302d7a2f19242ad7070a8af3f01b4eee9"} err="failed to get container status \"20703cc0174712162d24e0139f50da7302d7a2f19242ad7070a8af3f01b4eee9\": rpc error: code = NotFound desc = could not find container \"20703cc0174712162d24e0139f50da7302d7a2f19242ad7070a8af3f01b4eee9\": container with ID starting with 20703cc0174712162d24e0139f50da7302d7a2f19242ad7070a8af3f01b4eee9 not found: ID does not exist" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.696447 4810 scope.go:117] "RemoveContainer" containerID="8b9b17c744a8a4fd26bfc8645930c1dcc44173d2d0ba4b4bd66fc6ac3e2a264c" Dec 01 14:56:07 crc kubenswrapper[4810]: E1201 14:56:07.703008 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b9b17c744a8a4fd26bfc8645930c1dcc44173d2d0ba4b4bd66fc6ac3e2a264c\": container with ID starting with 8b9b17c744a8a4fd26bfc8645930c1dcc44173d2d0ba4b4bd66fc6ac3e2a264c not found: ID does not exist" containerID="8b9b17c744a8a4fd26bfc8645930c1dcc44173d2d0ba4b4bd66fc6ac3e2a264c" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.703044 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b9b17c744a8a4fd26bfc8645930c1dcc44173d2d0ba4b4bd66fc6ac3e2a264c"} err="failed to get container status \"8b9b17c744a8a4fd26bfc8645930c1dcc44173d2d0ba4b4bd66fc6ac3e2a264c\": rpc error: code = NotFound desc = could not find container \"8b9b17c744a8a4fd26bfc8645930c1dcc44173d2d0ba4b4bd66fc6ac3e2a264c\": container with ID starting with 8b9b17c744a8a4fd26bfc8645930c1dcc44173d2d0ba4b4bd66fc6ac3e2a264c not found: ID does not exist" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.703064 4810 scope.go:117] "RemoveContainer" containerID="66dab48997486cb3696b8a9bf3f8feab00e21edf9ee7e1b08337e550fde3b47c" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.733501 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66f9e303-31b7-4355-9acd-2720e03a9bb5-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.733529 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66f9e303-31b7-4355-9acd-2720e03a9bb5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.733538 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdbjs\" (UniqueName: \"kubernetes.io/projected/66f9e303-31b7-4355-9acd-2720e03a9bb5-kube-api-access-gdbjs\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.733548 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/66f9e303-31b7-4355-9acd-2720e03a9bb5-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.734610 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.758969 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66f9e303-31b7-4355-9acd-2720e03a9bb5-config-data" (OuterVolumeSpecName: "config-data") pod "66f9e303-31b7-4355-9acd-2720e03a9bb5" (UID: "66f9e303-31b7-4355-9acd-2720e03a9bb5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.759047 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.782650 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:56:07 crc kubenswrapper[4810]: E1201 14:56:07.783509 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66f9e303-31b7-4355-9acd-2720e03a9bb5" containerName="barbican-api" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.783527 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="66f9e303-31b7-4355-9acd-2720e03a9bb5" containerName="barbican-api" Dec 01 14:56:07 crc kubenswrapper[4810]: E1201 14:56:07.783544 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="073cab97-c80f-4905-92df-82b2537ed262" containerName="sg-core" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.783551 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="073cab97-c80f-4905-92df-82b2537ed262" containerName="sg-core" Dec 01 14:56:07 crc kubenswrapper[4810]: E1201 14:56:07.783573 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="073cab97-c80f-4905-92df-82b2537ed262" containerName="proxy-httpd" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.783581 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="073cab97-c80f-4905-92df-82b2537ed262" containerName="proxy-httpd" Dec 01 14:56:07 crc kubenswrapper[4810]: E1201 14:56:07.783598 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66f9e303-31b7-4355-9acd-2720e03a9bb5" containerName="barbican-api-log" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.783607 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="66f9e303-31b7-4355-9acd-2720e03a9bb5" containerName="barbican-api-log" Dec 01 14:56:07 crc kubenswrapper[4810]: E1201 14:56:07.783713 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="073cab97-c80f-4905-92df-82b2537ed262" containerName="ceilometer-notification-agent" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.783735 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="073cab97-c80f-4905-92df-82b2537ed262" containerName="ceilometer-notification-agent" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.783948 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="073cab97-c80f-4905-92df-82b2537ed262" containerName="proxy-httpd" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.783965 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="66f9e303-31b7-4355-9acd-2720e03a9bb5" containerName="barbican-api-log" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.783978 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="66f9e303-31b7-4355-9acd-2720e03a9bb5" containerName="barbican-api" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.783990 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="073cab97-c80f-4905-92df-82b2537ed262" containerName="sg-core" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.784005 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="073cab97-c80f-4905-92df-82b2537ed262" containerName="ceilometer-notification-agent" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.786035 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.789062 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.790338 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.797679 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.835650 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-run-httpd\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.835886 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-config-data\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.836039 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.836130 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-scripts\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.836266 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9twtd\" (UniqueName: \"kubernetes.io/projected/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-kube-api-access-9twtd\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.836348 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.836441 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-log-httpd\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.836613 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66f9e303-31b7-4355-9acd-2720e03a9bb5-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.938284 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9twtd\" (UniqueName: \"kubernetes.io/projected/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-kube-api-access-9twtd\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.938320 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.938367 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-log-httpd\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.938434 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-run-httpd\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.938452 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-config-data\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.938482 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.938499 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-scripts\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.940992 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-log-httpd\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.941396 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-run-httpd\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.942596 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-scripts\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.944058 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.946836 4810 scope.go:117] "RemoveContainer" containerID="5e1ffc3bf5219302f8daa00b25be0c838237ca853ad9bafeaa3d4c4829b55773" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.947294 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.953092 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-config-data\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:07 crc kubenswrapper[4810]: I1201 14:56:07.958982 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9twtd\" (UniqueName: \"kubernetes.io/projected/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-kube-api-access-9twtd\") pod \"ceilometer-0\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " pod="openstack/ceilometer-0" Dec 01 14:56:08 crc kubenswrapper[4810]: I1201 14:56:08.130957 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-dbfcbd88b-69qn5"] Dec 01 14:56:08 crc kubenswrapper[4810]: I1201 14:56:08.140086 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-dbfcbd88b-69qn5"] Dec 01 14:56:08 crc kubenswrapper[4810]: I1201 14:56:08.142530 4810 scope.go:117] "RemoveContainer" containerID="1d39fda5deb853dc5f22a662445df2f289042a4470edfdd81faeef5d2732a791" Dec 01 14:56:08 crc kubenswrapper[4810]: I1201 14:56:08.161588 4810 scope.go:117] "RemoveContainer" containerID="66dab48997486cb3696b8a9bf3f8feab00e21edf9ee7e1b08337e550fde3b47c" Dec 01 14:56:08 crc kubenswrapper[4810]: E1201 14:56:08.161977 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66dab48997486cb3696b8a9bf3f8feab00e21edf9ee7e1b08337e550fde3b47c\": container with ID starting with 66dab48997486cb3696b8a9bf3f8feab00e21edf9ee7e1b08337e550fde3b47c not found: ID does not exist" containerID="66dab48997486cb3696b8a9bf3f8feab00e21edf9ee7e1b08337e550fde3b47c" Dec 01 14:56:08 crc kubenswrapper[4810]: I1201 14:56:08.162006 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66dab48997486cb3696b8a9bf3f8feab00e21edf9ee7e1b08337e550fde3b47c"} err="failed to get container status \"66dab48997486cb3696b8a9bf3f8feab00e21edf9ee7e1b08337e550fde3b47c\": rpc error: code = NotFound desc = could not find container \"66dab48997486cb3696b8a9bf3f8feab00e21edf9ee7e1b08337e550fde3b47c\": container with ID starting with 66dab48997486cb3696b8a9bf3f8feab00e21edf9ee7e1b08337e550fde3b47c not found: ID does not exist" Dec 01 14:56:08 crc kubenswrapper[4810]: I1201 14:56:08.162047 4810 scope.go:117] "RemoveContainer" containerID="5e1ffc3bf5219302f8daa00b25be0c838237ca853ad9bafeaa3d4c4829b55773" Dec 01 14:56:08 crc kubenswrapper[4810]: E1201 14:56:08.162765 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e1ffc3bf5219302f8daa00b25be0c838237ca853ad9bafeaa3d4c4829b55773\": container with ID starting with 5e1ffc3bf5219302f8daa00b25be0c838237ca853ad9bafeaa3d4c4829b55773 not found: ID does not exist" containerID="5e1ffc3bf5219302f8daa00b25be0c838237ca853ad9bafeaa3d4c4829b55773" Dec 01 14:56:08 crc kubenswrapper[4810]: I1201 14:56:08.162789 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e1ffc3bf5219302f8daa00b25be0c838237ca853ad9bafeaa3d4c4829b55773"} err="failed to get container status \"5e1ffc3bf5219302f8daa00b25be0c838237ca853ad9bafeaa3d4c4829b55773\": rpc error: code = NotFound desc = could not find container \"5e1ffc3bf5219302f8daa00b25be0c838237ca853ad9bafeaa3d4c4829b55773\": container with ID starting with 5e1ffc3bf5219302f8daa00b25be0c838237ca853ad9bafeaa3d4c4829b55773 not found: ID does not exist" Dec 01 14:56:08 crc kubenswrapper[4810]: I1201 14:56:08.162803 4810 scope.go:117] "RemoveContainer" containerID="1d39fda5deb853dc5f22a662445df2f289042a4470edfdd81faeef5d2732a791" Dec 01 14:56:08 crc kubenswrapper[4810]: E1201 14:56:08.168685 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d39fda5deb853dc5f22a662445df2f289042a4470edfdd81faeef5d2732a791\": container with ID starting with 1d39fda5deb853dc5f22a662445df2f289042a4470edfdd81faeef5d2732a791 not found: ID does not exist" containerID="1d39fda5deb853dc5f22a662445df2f289042a4470edfdd81faeef5d2732a791" Dec 01 14:56:08 crc kubenswrapper[4810]: I1201 14:56:08.168795 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d39fda5deb853dc5f22a662445df2f289042a4470edfdd81faeef5d2732a791"} err="failed to get container status \"1d39fda5deb853dc5f22a662445df2f289042a4470edfdd81faeef5d2732a791\": rpc error: code = NotFound desc = could not find container \"1d39fda5deb853dc5f22a662445df2f289042a4470edfdd81faeef5d2732a791\": container with ID starting with 1d39fda5deb853dc5f22a662445df2f289042a4470edfdd81faeef5d2732a791 not found: ID does not exist" Dec 01 14:56:08 crc kubenswrapper[4810]: I1201 14:56:08.191099 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 14:56:08 crc kubenswrapper[4810]: I1201 14:56:08.221052 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:56:08 crc kubenswrapper[4810]: I1201 14:56:08.501680 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="073cab97-c80f-4905-92df-82b2537ed262" path="/var/lib/kubelet/pods/073cab97-c80f-4905-92df-82b2537ed262/volumes" Dec 01 14:56:08 crc kubenswrapper[4810]: I1201 14:56:08.511763 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66f9e303-31b7-4355-9acd-2720e03a9bb5" path="/var/lib/kubelet/pods/66f9e303-31b7-4355-9acd-2720e03a9bb5/volumes" Dec 01 14:56:08 crc kubenswrapper[4810]: I1201 14:56:08.706938 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fa46114f-d15e-416f-9085-7ada02bfbfdf","Type":"ContainerStarted","Data":"1c0a01407912518707090d666465d2ab681d6e1ad54b6d3ecf9772dad6bbe3c7"} Dec 01 14:56:08 crc kubenswrapper[4810]: I1201 14:56:08.815462 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:56:08 crc kubenswrapper[4810]: I1201 14:56:08.815730 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" event={"ID":"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed","Type":"ContainerStarted","Data":"fe264e2791d0c18cc3286e99e91fe1842646747df8709e9962de684167859864"} Dec 01 14:56:08 crc kubenswrapper[4810]: I1201 14:56:08.815912 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:08 crc kubenswrapper[4810]: W1201 14:56:08.815935 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a454ac3_a4dc_41c9_b548_3b57cad82bc5.slice/crio-29804ee1f5096e755708f4ef3aa4bd594ed3ebe547d076ebcc7bd7a8474f45c9 WatchSource:0}: Error finding container 29804ee1f5096e755708f4ef3aa4bd594ed3ebe547d076ebcc7bd7a8474f45c9: Status 404 returned error can't find the container with id 29804ee1f5096e755708f4ef3aa4bd594ed3ebe547d076ebcc7bd7a8474f45c9 Dec 01 14:56:08 crc kubenswrapper[4810]: I1201 14:56:08.853736 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" podStartSLOduration=3.853713962 podStartE2EDuration="3.853713962s" podCreationTimestamp="2025-12-01 14:56:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:56:08.837599144 +0000 UTC m=+1334.601108747" watchObservedRunningTime="2025-12-01 14:56:08.853713962 +0000 UTC m=+1334.617223565" Dec 01 14:56:09 crc kubenswrapper[4810]: I1201 14:56:09.826764 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"42ccc7b5-8af0-418b-8ca7-5a05095280c6","Type":"ContainerStarted","Data":"51175a934514541664d84258dbcce9bbef04606fd747040a29dfeb5cf200c6b5"} Dec 01 14:56:09 crc kubenswrapper[4810]: I1201 14:56:09.827438 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"42ccc7b5-8af0-418b-8ca7-5a05095280c6","Type":"ContainerStarted","Data":"86785eb4efabc3e128734d53b6b7cc16e62db1ee18d6a497fd91efce191ea40b"} Dec 01 14:56:09 crc kubenswrapper[4810]: I1201 14:56:09.828233 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a454ac3-a4dc-41c9-b548-3b57cad82bc5","Type":"ContainerStarted","Data":"040cad2764ca8aa7273ce97e369d59aed9af01a3cc2aef0ac5e4e35d012e003a"} Dec 01 14:56:09 crc kubenswrapper[4810]: I1201 14:56:09.828279 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a454ac3-a4dc-41c9-b548-3b57cad82bc5","Type":"ContainerStarted","Data":"29804ee1f5096e755708f4ef3aa4bd594ed3ebe547d076ebcc7bd7a8474f45c9"} Dec 01 14:56:09 crc kubenswrapper[4810]: I1201 14:56:09.830776 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fa46114f-d15e-416f-9085-7ada02bfbfdf","Type":"ContainerStarted","Data":"9cbbf02601c53c34cda7cec12732439176f2cfe36d36a61508b4e599ee9d0926"} Dec 01 14:56:09 crc kubenswrapper[4810]: I1201 14:56:09.830863 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="fa46114f-d15e-416f-9085-7ada02bfbfdf" containerName="cinder-api-log" containerID="cri-o://1c0a01407912518707090d666465d2ab681d6e1ad54b6d3ecf9772dad6bbe3c7" gracePeriod=30 Dec 01 14:56:09 crc kubenswrapper[4810]: I1201 14:56:09.830908 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 01 14:56:09 crc kubenswrapper[4810]: I1201 14:56:09.830918 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="fa46114f-d15e-416f-9085-7ada02bfbfdf" containerName="cinder-api" containerID="cri-o://9cbbf02601c53c34cda7cec12732439176f2cfe36d36a61508b4e599ee9d0926" gracePeriod=30 Dec 01 14:56:09 crc kubenswrapper[4810]: I1201 14:56:09.856082 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.56615606 podStartE2EDuration="4.856060038s" podCreationTimestamp="2025-12-01 14:56:05 +0000 UTC" firstStartedPulling="2025-12-01 14:56:06.693607787 +0000 UTC m=+1332.457117390" lastFinishedPulling="2025-12-01 14:56:07.983511755 +0000 UTC m=+1333.747021368" observedRunningTime="2025-12-01 14:56:09.850189979 +0000 UTC m=+1335.613699622" watchObservedRunningTime="2025-12-01 14:56:09.856060038 +0000 UTC m=+1335.619569641" Dec 01 14:56:09 crc kubenswrapper[4810]: I1201 14:56:09.870836 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.87081475 podStartE2EDuration="3.87081475s" podCreationTimestamp="2025-12-01 14:56:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:56:09.8660648 +0000 UTC m=+1335.629574433" watchObservedRunningTime="2025-12-01 14:56:09.87081475 +0000 UTC m=+1335.634324353" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.403171 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.528295 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-combined-ca-bundle\") pod \"fa46114f-d15e-416f-9085-7ada02bfbfdf\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.528414 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa46114f-d15e-416f-9085-7ada02bfbfdf-logs\") pod \"fa46114f-d15e-416f-9085-7ada02bfbfdf\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.528433 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fa46114f-d15e-416f-9085-7ada02bfbfdf-etc-machine-id\") pod \"fa46114f-d15e-416f-9085-7ada02bfbfdf\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.528452 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-config-data-custom\") pod \"fa46114f-d15e-416f-9085-7ada02bfbfdf\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.528574 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-scripts\") pod \"fa46114f-d15e-416f-9085-7ada02bfbfdf\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.528621 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzgtd\" (UniqueName: \"kubernetes.io/projected/fa46114f-d15e-416f-9085-7ada02bfbfdf-kube-api-access-kzgtd\") pod \"fa46114f-d15e-416f-9085-7ada02bfbfdf\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.528671 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-config-data\") pod \"fa46114f-d15e-416f-9085-7ada02bfbfdf\" (UID: \"fa46114f-d15e-416f-9085-7ada02bfbfdf\") " Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.529221 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa46114f-d15e-416f-9085-7ada02bfbfdf-logs" (OuterVolumeSpecName: "logs") pod "fa46114f-d15e-416f-9085-7ada02bfbfdf" (UID: "fa46114f-d15e-416f-9085-7ada02bfbfdf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.529302 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa46114f-d15e-416f-9085-7ada02bfbfdf-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "fa46114f-d15e-416f-9085-7ada02bfbfdf" (UID: "fa46114f-d15e-416f-9085-7ada02bfbfdf"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.532501 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fa46114f-d15e-416f-9085-7ada02bfbfdf" (UID: "fa46114f-d15e-416f-9085-7ada02bfbfdf"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.532590 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-scripts" (OuterVolumeSpecName: "scripts") pod "fa46114f-d15e-416f-9085-7ada02bfbfdf" (UID: "fa46114f-d15e-416f-9085-7ada02bfbfdf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.533225 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa46114f-d15e-416f-9085-7ada02bfbfdf-kube-api-access-kzgtd" (OuterVolumeSpecName: "kube-api-access-kzgtd") pod "fa46114f-d15e-416f-9085-7ada02bfbfdf" (UID: "fa46114f-d15e-416f-9085-7ada02bfbfdf"). InnerVolumeSpecName "kube-api-access-kzgtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.555758 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fa46114f-d15e-416f-9085-7ada02bfbfdf" (UID: "fa46114f-d15e-416f-9085-7ada02bfbfdf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.605532 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-config-data" (OuterVolumeSpecName: "config-data") pod "fa46114f-d15e-416f-9085-7ada02bfbfdf" (UID: "fa46114f-d15e-416f-9085-7ada02bfbfdf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.633945 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.634075 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa46114f-d15e-416f-9085-7ada02bfbfdf-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.634145 4810 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fa46114f-d15e-416f-9085-7ada02bfbfdf-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.634210 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.634354 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.634421 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzgtd\" (UniqueName: \"kubernetes.io/projected/fa46114f-d15e-416f-9085-7ada02bfbfdf-kube-api-access-kzgtd\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.634513 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa46114f-d15e-416f-9085-7ada02bfbfdf-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.844957 4810 generic.go:334] "Generic (PLEG): container finished" podID="fa46114f-d15e-416f-9085-7ada02bfbfdf" containerID="9cbbf02601c53c34cda7cec12732439176f2cfe36d36a61508b4e599ee9d0926" exitCode=0 Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.845292 4810 generic.go:334] "Generic (PLEG): container finished" podID="fa46114f-d15e-416f-9085-7ada02bfbfdf" containerID="1c0a01407912518707090d666465d2ab681d6e1ad54b6d3ecf9772dad6bbe3c7" exitCode=143 Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.845035 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.845036 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fa46114f-d15e-416f-9085-7ada02bfbfdf","Type":"ContainerDied","Data":"9cbbf02601c53c34cda7cec12732439176f2cfe36d36a61508b4e599ee9d0926"} Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.845464 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fa46114f-d15e-416f-9085-7ada02bfbfdf","Type":"ContainerDied","Data":"1c0a01407912518707090d666465d2ab681d6e1ad54b6d3ecf9772dad6bbe3c7"} Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.845528 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"fa46114f-d15e-416f-9085-7ada02bfbfdf","Type":"ContainerDied","Data":"08dcdb366ccc5bc992a72e19d7b98460b530f9f7b073f457b60f7aa7d6f081f9"} Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.845545 4810 scope.go:117] "RemoveContainer" containerID="9cbbf02601c53c34cda7cec12732439176f2cfe36d36a61508b4e599ee9d0926" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.849428 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a454ac3-a4dc-41c9-b548-3b57cad82bc5","Type":"ContainerStarted","Data":"6f2250ffa5a8fa4cbd93aff65636f6a04fd9b92fdbf1ffbc4e97507590ff20a3"} Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.887716 4810 scope.go:117] "RemoveContainer" containerID="1c0a01407912518707090d666465d2ab681d6e1ad54b6d3ecf9772dad6bbe3c7" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.895390 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.907636 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.921608 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 01 14:56:10 crc kubenswrapper[4810]: E1201 14:56:10.922118 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa46114f-d15e-416f-9085-7ada02bfbfdf" containerName="cinder-api-log" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.922134 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa46114f-d15e-416f-9085-7ada02bfbfdf" containerName="cinder-api-log" Dec 01 14:56:10 crc kubenswrapper[4810]: E1201 14:56:10.922154 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa46114f-d15e-416f-9085-7ada02bfbfdf" containerName="cinder-api" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.922161 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa46114f-d15e-416f-9085-7ada02bfbfdf" containerName="cinder-api" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.922400 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa46114f-d15e-416f-9085-7ada02bfbfdf" containerName="cinder-api-log" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.922430 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa46114f-d15e-416f-9085-7ada02bfbfdf" containerName="cinder-api" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.923595 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.926933 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.927080 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.927100 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.933997 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.969193 4810 scope.go:117] "RemoveContainer" containerID="9cbbf02601c53c34cda7cec12732439176f2cfe36d36a61508b4e599ee9d0926" Dec 01 14:56:10 crc kubenswrapper[4810]: E1201 14:56:10.971283 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cbbf02601c53c34cda7cec12732439176f2cfe36d36a61508b4e599ee9d0926\": container with ID starting with 9cbbf02601c53c34cda7cec12732439176f2cfe36d36a61508b4e599ee9d0926 not found: ID does not exist" containerID="9cbbf02601c53c34cda7cec12732439176f2cfe36d36a61508b4e599ee9d0926" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.971333 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cbbf02601c53c34cda7cec12732439176f2cfe36d36a61508b4e599ee9d0926"} err="failed to get container status \"9cbbf02601c53c34cda7cec12732439176f2cfe36d36a61508b4e599ee9d0926\": rpc error: code = NotFound desc = could not find container \"9cbbf02601c53c34cda7cec12732439176f2cfe36d36a61508b4e599ee9d0926\": container with ID starting with 9cbbf02601c53c34cda7cec12732439176f2cfe36d36a61508b4e599ee9d0926 not found: ID does not exist" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.971356 4810 scope.go:117] "RemoveContainer" containerID="1c0a01407912518707090d666465d2ab681d6e1ad54b6d3ecf9772dad6bbe3c7" Dec 01 14:56:10 crc kubenswrapper[4810]: E1201 14:56:10.975818 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c0a01407912518707090d666465d2ab681d6e1ad54b6d3ecf9772dad6bbe3c7\": container with ID starting with 1c0a01407912518707090d666465d2ab681d6e1ad54b6d3ecf9772dad6bbe3c7 not found: ID does not exist" containerID="1c0a01407912518707090d666465d2ab681d6e1ad54b6d3ecf9772dad6bbe3c7" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.975876 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c0a01407912518707090d666465d2ab681d6e1ad54b6d3ecf9772dad6bbe3c7"} err="failed to get container status \"1c0a01407912518707090d666465d2ab681d6e1ad54b6d3ecf9772dad6bbe3c7\": rpc error: code = NotFound desc = could not find container \"1c0a01407912518707090d666465d2ab681d6e1ad54b6d3ecf9772dad6bbe3c7\": container with ID starting with 1c0a01407912518707090d666465d2ab681d6e1ad54b6d3ecf9772dad6bbe3c7 not found: ID does not exist" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.975910 4810 scope.go:117] "RemoveContainer" containerID="9cbbf02601c53c34cda7cec12732439176f2cfe36d36a61508b4e599ee9d0926" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.976374 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cbbf02601c53c34cda7cec12732439176f2cfe36d36a61508b4e599ee9d0926"} err="failed to get container status \"9cbbf02601c53c34cda7cec12732439176f2cfe36d36a61508b4e599ee9d0926\": rpc error: code = NotFound desc = could not find container \"9cbbf02601c53c34cda7cec12732439176f2cfe36d36a61508b4e599ee9d0926\": container with ID starting with 9cbbf02601c53c34cda7cec12732439176f2cfe36d36a61508b4e599ee9d0926 not found: ID does not exist" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.976405 4810 scope.go:117] "RemoveContainer" containerID="1c0a01407912518707090d666465d2ab681d6e1ad54b6d3ecf9772dad6bbe3c7" Dec 01 14:56:10 crc kubenswrapper[4810]: I1201 14:56:10.976748 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c0a01407912518707090d666465d2ab681d6e1ad54b6d3ecf9772dad6bbe3c7"} err="failed to get container status \"1c0a01407912518707090d666465d2ab681d6e1ad54b6d3ecf9772dad6bbe3c7\": rpc error: code = NotFound desc = could not find container \"1c0a01407912518707090d666465d2ab681d6e1ad54b6d3ecf9772dad6bbe3c7\": container with ID starting with 1c0a01407912518707090d666465d2ab681d6e1ad54b6d3ecf9772dad6bbe3c7 not found: ID does not exist" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.044832 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.044914 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.045092 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-scripts\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.045277 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.045413 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-logs\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.045495 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-config-data-custom\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.045537 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-config-data\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.045570 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.045593 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlz9r\" (UniqueName: \"kubernetes.io/projected/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-kube-api-access-nlz9r\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.147099 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.147164 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-logs\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.147190 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-config-data-custom\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.147213 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-config-data\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.147242 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.147262 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlz9r\" (UniqueName: \"kubernetes.io/projected/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-kube-api-access-nlz9r\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.147327 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.147363 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.147396 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-scripts\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.147437 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.147657 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-logs\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.154394 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-scripts\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.157164 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-config-data-custom\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.159579 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.163742 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.164368 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-config-data\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.171945 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.173391 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlz9r\" (UniqueName: \"kubernetes.io/projected/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-kube-api-access-nlz9r\") pod \"cinder-api-0\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.190370 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.250006 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.711043 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 14:56:11 crc kubenswrapper[4810]: W1201 14:56:11.715244 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode79c5698_ab2d_4292_bda9_fab8dbbb2abb.slice/crio-d3af5f0b7ed3a270d5b91e569b451e6bd182c3e93d804653e0703cad50ee10d6 WatchSource:0}: Error finding container d3af5f0b7ed3a270d5b91e569b451e6bd182c3e93d804653e0703cad50ee10d6: Status 404 returned error can't find the container with id d3af5f0b7ed3a270d5b91e569b451e6bd182c3e93d804653e0703cad50ee10d6 Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.889862 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e79c5698-ab2d-4292-bda9-fab8dbbb2abb","Type":"ContainerStarted","Data":"d3af5f0b7ed3a270d5b91e569b451e6bd182c3e93d804653e0703cad50ee10d6"} Dec 01 14:56:11 crc kubenswrapper[4810]: I1201 14:56:11.897850 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a454ac3-a4dc-41c9-b548-3b57cad82bc5","Type":"ContainerStarted","Data":"5c9a3f70bc821a5731836c8aed647e817a723b5616565d4bf6d518bff2fdb117"} Dec 01 14:56:12 crc kubenswrapper[4810]: I1201 14:56:12.503232 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa46114f-d15e-416f-9085-7ada02bfbfdf" path="/var/lib/kubelet/pods/fa46114f-d15e-416f-9085-7ada02bfbfdf/volumes" Dec 01 14:56:12 crc kubenswrapper[4810]: I1201 14:56:12.920421 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e79c5698-ab2d-4292-bda9-fab8dbbb2abb","Type":"ContainerStarted","Data":"ebb0282754daba83e3b7cf8ef6784b85c2feea8bbea0bae1fde73a5239ce6c9c"} Dec 01 14:56:13 crc kubenswrapper[4810]: I1201 14:56:13.702892 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:56:13 crc kubenswrapper[4810]: I1201 14:56:13.932494 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e79c5698-ab2d-4292-bda9-fab8dbbb2abb","Type":"ContainerStarted","Data":"085bfd508911509acb2f401ff11dc8f197ed30234eac469830bf853fef1a201a"} Dec 01 14:56:13 crc kubenswrapper[4810]: I1201 14:56:13.932568 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 01 14:56:13 crc kubenswrapper[4810]: I1201 14:56:13.938153 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a454ac3-a4dc-41c9-b548-3b57cad82bc5","Type":"ContainerStarted","Data":"0720428d31bf19f31e912919435903191f3bcb90677211a516c8c3dbd918a037"} Dec 01 14:56:13 crc kubenswrapper[4810]: I1201 14:56:13.938255 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 14:56:13 crc kubenswrapper[4810]: I1201 14:56:13.954856 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.954813713 podStartE2EDuration="3.954813713s" podCreationTimestamp="2025-12-01 14:56:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:56:13.952345885 +0000 UTC m=+1339.715855498" watchObservedRunningTime="2025-12-01 14:56:13.954813713 +0000 UTC m=+1339.718323306" Dec 01 14:56:13 crc kubenswrapper[4810]: I1201 14:56:13.977081 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.897430251 podStartE2EDuration="6.977064336s" podCreationTimestamp="2025-12-01 14:56:07 +0000 UTC" firstStartedPulling="2025-12-01 14:56:08.818230728 +0000 UTC m=+1334.581740331" lastFinishedPulling="2025-12-01 14:56:12.897864813 +0000 UTC m=+1338.661374416" observedRunningTime="2025-12-01 14:56:13.975297638 +0000 UTC m=+1339.738807241" watchObservedRunningTime="2025-12-01 14:56:13.977064336 +0000 UTC m=+1339.740573939" Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.142287 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5477779d95-wg2vp" Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.147455 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.232394 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5df84c984d-frb8v"] Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.232679 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5df84c984d-frb8v" podUID="ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42" containerName="neutron-api" containerID="cri-o://a81f5574834d150e9892017d8ae052bc6da9601608100fab0fac319ce5cd3cf8" gracePeriod=30 Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.233161 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5df84c984d-frb8v" podUID="ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42" containerName="neutron-httpd" containerID="cri-o://ce152a2dcedab6dfa2222a388bf9973c1bcce7517a6f7e4bf7a4901cbcf8f86a" gracePeriod=30 Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.244887 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64dfd64c45-nssxw"] Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.245147 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" podUID="21371b53-8ea0-40e1-a68e-c140a606b610" containerName="dnsmasq-dns" containerID="cri-o://78d365d0d6eb8ecf1b2484dc5acf8196acada7b43e9907e6fc8e57c44892e66e" gracePeriod=10 Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.481142 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.522240 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.780431 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.877368 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-dns-swift-storage-0\") pod \"21371b53-8ea0-40e1-a68e-c140a606b610\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.877664 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-ovsdbserver-nb\") pod \"21371b53-8ea0-40e1-a68e-c140a606b610\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.877738 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-config\") pod \"21371b53-8ea0-40e1-a68e-c140a606b610\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.877817 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-dns-svc\") pod \"21371b53-8ea0-40e1-a68e-c140a606b610\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.877893 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-ovsdbserver-sb\") pod \"21371b53-8ea0-40e1-a68e-c140a606b610\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.877965 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k2tq\" (UniqueName: \"kubernetes.io/projected/21371b53-8ea0-40e1-a68e-c140a606b610-kube-api-access-2k2tq\") pod \"21371b53-8ea0-40e1-a68e-c140a606b610\" (UID: \"21371b53-8ea0-40e1-a68e-c140a606b610\") " Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.904203 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21371b53-8ea0-40e1-a68e-c140a606b610-kube-api-access-2k2tq" (OuterVolumeSpecName: "kube-api-access-2k2tq") pod "21371b53-8ea0-40e1-a68e-c140a606b610" (UID: "21371b53-8ea0-40e1-a68e-c140a606b610"). InnerVolumeSpecName "kube-api-access-2k2tq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.933304 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "21371b53-8ea0-40e1-a68e-c140a606b610" (UID: "21371b53-8ea0-40e1-a68e-c140a606b610"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.935032 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "21371b53-8ea0-40e1-a68e-c140a606b610" (UID: "21371b53-8ea0-40e1-a68e-c140a606b610"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.945358 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "21371b53-8ea0-40e1-a68e-c140a606b610" (UID: "21371b53-8ea0-40e1-a68e-c140a606b610"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.946131 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "21371b53-8ea0-40e1-a68e-c140a606b610" (UID: "21371b53-8ea0-40e1-a68e-c140a606b610"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.964392 4810 generic.go:334] "Generic (PLEG): container finished" podID="21371b53-8ea0-40e1-a68e-c140a606b610" containerID="78d365d0d6eb8ecf1b2484dc5acf8196acada7b43e9907e6fc8e57c44892e66e" exitCode=0 Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.964434 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" event={"ID":"21371b53-8ea0-40e1-a68e-c140a606b610","Type":"ContainerDied","Data":"78d365d0d6eb8ecf1b2484dc5acf8196acada7b43e9907e6fc8e57c44892e66e"} Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.964495 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" event={"ID":"21371b53-8ea0-40e1-a68e-c140a606b610","Type":"ContainerDied","Data":"30f6aab787fe48b569b9c9989bf6362f5ab52c8c5714076566ad91e6646cfde2"} Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.964512 4810 scope.go:117] "RemoveContainer" containerID="78d365d0d6eb8ecf1b2484dc5acf8196acada7b43e9907e6fc8e57c44892e66e" Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.964505 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dfd64c45-nssxw" Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.967347 4810 generic.go:334] "Generic (PLEG): container finished" podID="ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42" containerID="ce152a2dcedab6dfa2222a388bf9973c1bcce7517a6f7e4bf7a4901cbcf8f86a" exitCode=0 Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.968098 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="42ccc7b5-8af0-418b-8ca7-5a05095280c6" containerName="cinder-scheduler" containerID="cri-o://86785eb4efabc3e128734d53b6b7cc16e62db1ee18d6a497fd91efce191ea40b" gracePeriod=30 Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.968444 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5df84c984d-frb8v" event={"ID":"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42","Type":"ContainerDied","Data":"ce152a2dcedab6dfa2222a388bf9973c1bcce7517a6f7e4bf7a4901cbcf8f86a"} Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.970308 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="42ccc7b5-8af0-418b-8ca7-5a05095280c6" containerName="probe" containerID="cri-o://51175a934514541664d84258dbcce9bbef04606fd747040a29dfeb5cf200c6b5" gracePeriod=30 Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.982947 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k2tq\" (UniqueName: \"kubernetes.io/projected/21371b53-8ea0-40e1-a68e-c140a606b610-kube-api-access-2k2tq\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.982984 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.982993 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.983002 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.983011 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.983321 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-config" (OuterVolumeSpecName: "config") pod "21371b53-8ea0-40e1-a68e-c140a606b610" (UID: "21371b53-8ea0-40e1-a68e-c140a606b610"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:56:16 crc kubenswrapper[4810]: I1201 14:56:16.992083 4810 scope.go:117] "RemoveContainer" containerID="be0f50bda2529e30fd38080d5e23545e84b94561feb08c8bbedbb42f9d764c46" Dec 01 14:56:17 crc kubenswrapper[4810]: I1201 14:56:17.010527 4810 scope.go:117] "RemoveContainer" containerID="78d365d0d6eb8ecf1b2484dc5acf8196acada7b43e9907e6fc8e57c44892e66e" Dec 01 14:56:17 crc kubenswrapper[4810]: E1201 14:56:17.010907 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78d365d0d6eb8ecf1b2484dc5acf8196acada7b43e9907e6fc8e57c44892e66e\": container with ID starting with 78d365d0d6eb8ecf1b2484dc5acf8196acada7b43e9907e6fc8e57c44892e66e not found: ID does not exist" containerID="78d365d0d6eb8ecf1b2484dc5acf8196acada7b43e9907e6fc8e57c44892e66e" Dec 01 14:56:17 crc kubenswrapper[4810]: I1201 14:56:17.010939 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78d365d0d6eb8ecf1b2484dc5acf8196acada7b43e9907e6fc8e57c44892e66e"} err="failed to get container status \"78d365d0d6eb8ecf1b2484dc5acf8196acada7b43e9907e6fc8e57c44892e66e\": rpc error: code = NotFound desc = could not find container \"78d365d0d6eb8ecf1b2484dc5acf8196acada7b43e9907e6fc8e57c44892e66e\": container with ID starting with 78d365d0d6eb8ecf1b2484dc5acf8196acada7b43e9907e6fc8e57c44892e66e not found: ID does not exist" Dec 01 14:56:17 crc kubenswrapper[4810]: I1201 14:56:17.010960 4810 scope.go:117] "RemoveContainer" containerID="be0f50bda2529e30fd38080d5e23545e84b94561feb08c8bbedbb42f9d764c46" Dec 01 14:56:17 crc kubenswrapper[4810]: E1201 14:56:17.011676 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be0f50bda2529e30fd38080d5e23545e84b94561feb08c8bbedbb42f9d764c46\": container with ID starting with be0f50bda2529e30fd38080d5e23545e84b94561feb08c8bbedbb42f9d764c46 not found: ID does not exist" containerID="be0f50bda2529e30fd38080d5e23545e84b94561feb08c8bbedbb42f9d764c46" Dec 01 14:56:17 crc kubenswrapper[4810]: I1201 14:56:17.011704 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be0f50bda2529e30fd38080d5e23545e84b94561feb08c8bbedbb42f9d764c46"} err="failed to get container status \"be0f50bda2529e30fd38080d5e23545e84b94561feb08c8bbedbb42f9d764c46\": rpc error: code = NotFound desc = could not find container \"be0f50bda2529e30fd38080d5e23545e84b94561feb08c8bbedbb42f9d764c46\": container with ID starting with be0f50bda2529e30fd38080d5e23545e84b94561feb08c8bbedbb42f9d764c46 not found: ID does not exist" Dec 01 14:56:17 crc kubenswrapper[4810]: I1201 14:56:17.085088 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21371b53-8ea0-40e1-a68e-c140a606b610-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:17 crc kubenswrapper[4810]: I1201 14:56:17.294546 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64dfd64c45-nssxw"] Dec 01 14:56:17 crc kubenswrapper[4810]: I1201 14:56:17.302442 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64dfd64c45-nssxw"] Dec 01 14:56:17 crc kubenswrapper[4810]: I1201 14:56:17.980389 4810 generic.go:334] "Generic (PLEG): container finished" podID="42ccc7b5-8af0-418b-8ca7-5a05095280c6" containerID="51175a934514541664d84258dbcce9bbef04606fd747040a29dfeb5cf200c6b5" exitCode=0 Dec 01 14:56:17 crc kubenswrapper[4810]: I1201 14:56:17.980446 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"42ccc7b5-8af0-418b-8ca7-5a05095280c6","Type":"ContainerDied","Data":"51175a934514541664d84258dbcce9bbef04606fd747040a29dfeb5cf200c6b5"} Dec 01 14:56:18 crc kubenswrapper[4810]: I1201 14:56:18.503127 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21371b53-8ea0-40e1-a68e-c140a606b610" path="/var/lib/kubelet/pods/21371b53-8ea0-40e1-a68e-c140a606b610/volumes" Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.580676 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.642682 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-scripts\") pod \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.642752 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-config-data\") pod \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.642778 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nv6d5\" (UniqueName: \"kubernetes.io/projected/42ccc7b5-8af0-418b-8ca7-5a05095280c6-kube-api-access-nv6d5\") pod \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.642846 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-config-data-custom\") pod \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.643012 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-combined-ca-bundle\") pod \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.643062 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/42ccc7b5-8af0-418b-8ca7-5a05095280c6-etc-machine-id\") pod \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\" (UID: \"42ccc7b5-8af0-418b-8ca7-5a05095280c6\") " Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.643514 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/42ccc7b5-8af0-418b-8ca7-5a05095280c6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "42ccc7b5-8af0-418b-8ca7-5a05095280c6" (UID: "42ccc7b5-8af0-418b-8ca7-5a05095280c6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.651243 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "42ccc7b5-8af0-418b-8ca7-5a05095280c6" (UID: "42ccc7b5-8af0-418b-8ca7-5a05095280c6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.655017 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-scripts" (OuterVolumeSpecName: "scripts") pod "42ccc7b5-8af0-418b-8ca7-5a05095280c6" (UID: "42ccc7b5-8af0-418b-8ca7-5a05095280c6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.664608 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42ccc7b5-8af0-418b-8ca7-5a05095280c6-kube-api-access-nv6d5" (OuterVolumeSpecName: "kube-api-access-nv6d5") pod "42ccc7b5-8af0-418b-8ca7-5a05095280c6" (UID: "42ccc7b5-8af0-418b-8ca7-5a05095280c6"). InnerVolumeSpecName "kube-api-access-nv6d5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.701105 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42ccc7b5-8af0-418b-8ca7-5a05095280c6" (UID: "42ccc7b5-8af0-418b-8ca7-5a05095280c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.745123 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.745157 4810 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/42ccc7b5-8af0-418b-8ca7-5a05095280c6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.745166 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.745174 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nv6d5\" (UniqueName: \"kubernetes.io/projected/42ccc7b5-8af0-418b-8ca7-5a05095280c6-kube-api-access-nv6d5\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.745183 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.799612 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-config-data" (OuterVolumeSpecName: "config-data") pod "42ccc7b5-8af0-418b-8ca7-5a05095280c6" (UID: "42ccc7b5-8af0-418b-8ca7-5a05095280c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:20 crc kubenswrapper[4810]: I1201 14:56:20.847143 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42ccc7b5-8af0-418b-8ca7-5a05095280c6-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.008963 4810 generic.go:334] "Generic (PLEG): container finished" podID="42ccc7b5-8af0-418b-8ca7-5a05095280c6" containerID="86785eb4efabc3e128734d53b6b7cc16e62db1ee18d6a497fd91efce191ea40b" exitCode=0 Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.009004 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"42ccc7b5-8af0-418b-8ca7-5a05095280c6","Type":"ContainerDied","Data":"86785eb4efabc3e128734d53b6b7cc16e62db1ee18d6a497fd91efce191ea40b"} Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.009036 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"42ccc7b5-8af0-418b-8ca7-5a05095280c6","Type":"ContainerDied","Data":"45f1c50adfd43e0b65d03a64d89ec867ca10758a043fbb694ccbcdb701225af9"} Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.009055 4810 scope.go:117] "RemoveContainer" containerID="51175a934514541664d84258dbcce9bbef04606fd747040a29dfeb5cf200c6b5" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.009351 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.039567 4810 scope.go:117] "RemoveContainer" containerID="86785eb4efabc3e128734d53b6b7cc16e62db1ee18d6a497fd91efce191ea40b" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.046640 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.056136 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.074965 4810 scope.go:117] "RemoveContainer" containerID="51175a934514541664d84258dbcce9bbef04606fd747040a29dfeb5cf200c6b5" Dec 01 14:56:21 crc kubenswrapper[4810]: E1201 14:56:21.075567 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51175a934514541664d84258dbcce9bbef04606fd747040a29dfeb5cf200c6b5\": container with ID starting with 51175a934514541664d84258dbcce9bbef04606fd747040a29dfeb5cf200c6b5 not found: ID does not exist" containerID="51175a934514541664d84258dbcce9bbef04606fd747040a29dfeb5cf200c6b5" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.075622 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51175a934514541664d84258dbcce9bbef04606fd747040a29dfeb5cf200c6b5"} err="failed to get container status \"51175a934514541664d84258dbcce9bbef04606fd747040a29dfeb5cf200c6b5\": rpc error: code = NotFound desc = could not find container \"51175a934514541664d84258dbcce9bbef04606fd747040a29dfeb5cf200c6b5\": container with ID starting with 51175a934514541664d84258dbcce9bbef04606fd747040a29dfeb5cf200c6b5 not found: ID does not exist" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.075661 4810 scope.go:117] "RemoveContainer" containerID="86785eb4efabc3e128734d53b6b7cc16e62db1ee18d6a497fd91efce191ea40b" Dec 01 14:56:21 crc kubenswrapper[4810]: E1201 14:56:21.075997 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86785eb4efabc3e128734d53b6b7cc16e62db1ee18d6a497fd91efce191ea40b\": container with ID starting with 86785eb4efabc3e128734d53b6b7cc16e62db1ee18d6a497fd91efce191ea40b not found: ID does not exist" containerID="86785eb4efabc3e128734d53b6b7cc16e62db1ee18d6a497fd91efce191ea40b" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.076080 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86785eb4efabc3e128734d53b6b7cc16e62db1ee18d6a497fd91efce191ea40b"} err="failed to get container status \"86785eb4efabc3e128734d53b6b7cc16e62db1ee18d6a497fd91efce191ea40b\": rpc error: code = NotFound desc = could not find container \"86785eb4efabc3e128734d53b6b7cc16e62db1ee18d6a497fd91efce191ea40b\": container with ID starting with 86785eb4efabc3e128734d53b6b7cc16e62db1ee18d6a497fd91efce191ea40b not found: ID does not exist" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.081323 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 14:56:21 crc kubenswrapper[4810]: E1201 14:56:21.081886 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21371b53-8ea0-40e1-a68e-c140a606b610" containerName="init" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.081979 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="21371b53-8ea0-40e1-a68e-c140a606b610" containerName="init" Dec 01 14:56:21 crc kubenswrapper[4810]: E1201 14:56:21.082055 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42ccc7b5-8af0-418b-8ca7-5a05095280c6" containerName="probe" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.082122 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="42ccc7b5-8af0-418b-8ca7-5a05095280c6" containerName="probe" Dec 01 14:56:21 crc kubenswrapper[4810]: E1201 14:56:21.082226 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21371b53-8ea0-40e1-a68e-c140a606b610" containerName="dnsmasq-dns" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.082302 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="21371b53-8ea0-40e1-a68e-c140a606b610" containerName="dnsmasq-dns" Dec 01 14:56:21 crc kubenswrapper[4810]: E1201 14:56:21.082369 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42ccc7b5-8af0-418b-8ca7-5a05095280c6" containerName="cinder-scheduler" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.082417 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="42ccc7b5-8af0-418b-8ca7-5a05095280c6" containerName="cinder-scheduler" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.082711 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="21371b53-8ea0-40e1-a68e-c140a606b610" containerName="dnsmasq-dns" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.082787 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="42ccc7b5-8af0-418b-8ca7-5a05095280c6" containerName="cinder-scheduler" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.082849 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="42ccc7b5-8af0-418b-8ca7-5a05095280c6" containerName="probe" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.083878 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.088350 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.097708 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.151520 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-config-data\") pod \"cinder-scheduler-0\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.151611 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.151630 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bef8b0c7-66cd-4930-a78b-bf73804a73d2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.151699 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.151716 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-scripts\") pod \"cinder-scheduler-0\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.151745 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v485w\" (UniqueName: \"kubernetes.io/projected/bef8b0c7-66cd-4930-a78b-bf73804a73d2-kube-api-access-v485w\") pod \"cinder-scheduler-0\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.253056 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.253096 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-scripts\") pod \"cinder-scheduler-0\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.253130 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v485w\" (UniqueName: \"kubernetes.io/projected/bef8b0c7-66cd-4930-a78b-bf73804a73d2-kube-api-access-v485w\") pod \"cinder-scheduler-0\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.253178 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-config-data\") pod \"cinder-scheduler-0\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.253239 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bef8b0c7-66cd-4930-a78b-bf73804a73d2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.253253 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.253323 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bef8b0c7-66cd-4930-a78b-bf73804a73d2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.258685 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-config-data\") pod \"cinder-scheduler-0\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.260046 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-scripts\") pod \"cinder-scheduler-0\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.265184 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.265826 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.274077 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v485w\" (UniqueName: \"kubernetes.io/projected/bef8b0c7-66cd-4930-a78b-bf73804a73d2-kube-api-access-v485w\") pod \"cinder-scheduler-0\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.357323 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.424952 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 14:56:21 crc kubenswrapper[4810]: I1201 14:56:21.933332 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 14:56:22 crc kubenswrapper[4810]: I1201 14:56:22.021146 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bef8b0c7-66cd-4930-a78b-bf73804a73d2","Type":"ContainerStarted","Data":"f7dff30ba20cc0a4c5812b5cf1ac8ac65e4d45ec8049c46527f8c1a87513a936"} Dec 01 14:56:22 crc kubenswrapper[4810]: I1201 14:56:22.368805 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:56:22 crc kubenswrapper[4810]: I1201 14:56:22.504031 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42ccc7b5-8af0-418b-8ca7-5a05095280c6" path="/var/lib/kubelet/pods/42ccc7b5-8af0-418b-8ca7-5a05095280c6/volumes" Dec 01 14:56:23 crc kubenswrapper[4810]: I1201 14:56:23.032172 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bef8b0c7-66cd-4930-a78b-bf73804a73d2","Type":"ContainerStarted","Data":"d01cd11ccf42943023a23e361f67460aa3a8387758c28bbf2c4f4a6dba26d397"} Dec 01 14:56:23 crc kubenswrapper[4810]: I1201 14:56:23.461601 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.044772 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5df84c984d-frb8v" event={"ID":"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42","Type":"ContainerDied","Data":"a81f5574834d150e9892017d8ae052bc6da9601608100fab0fac319ce5cd3cf8"} Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.044713 4810 generic.go:334] "Generic (PLEG): container finished" podID="ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42" containerID="a81f5574834d150e9892017d8ae052bc6da9601608100fab0fac319ce5cd3cf8" exitCode=0 Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.053294 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bef8b0c7-66cd-4930-a78b-bf73804a73d2","Type":"ContainerStarted","Data":"c3e8f637d25ffa49a1106b343b7c3861c122e5ccdeb4f52b034a012b1bf12322"} Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.079578 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.079545018 podStartE2EDuration="3.079545018s" podCreationTimestamp="2025-12-01 14:56:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:56:24.075460257 +0000 UTC m=+1349.838969860" watchObservedRunningTime="2025-12-01 14:56:24.079545018 +0000 UTC m=+1349.843054621" Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.394955 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.428979 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tffmc\" (UniqueName: \"kubernetes.io/projected/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-kube-api-access-tffmc\") pod \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.429055 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-httpd-config\") pod \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.429084 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-ovndb-tls-certs\") pod \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.429106 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-config\") pod \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.429174 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-combined-ca-bundle\") pod \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\" (UID: \"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42\") " Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.443701 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42" (UID: "ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.445718 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-kube-api-access-tffmc" (OuterVolumeSpecName: "kube-api-access-tffmc") pod "ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42" (UID: "ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42"). InnerVolumeSpecName "kube-api-access-tffmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.534556 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.534587 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tffmc\" (UniqueName: \"kubernetes.io/projected/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-kube-api-access-tffmc\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.564645 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-config" (OuterVolumeSpecName: "config") pod "ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42" (UID: "ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.573658 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42" (UID: "ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.639564 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.639603 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.653821 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42" (UID: "ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:24 crc kubenswrapper[4810]: I1201 14:56:24.741524 4810 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:25 crc kubenswrapper[4810]: I1201 14:56:25.065746 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5df84c984d-frb8v" Dec 01 14:56:25 crc kubenswrapper[4810]: I1201 14:56:25.070675 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5df84c984d-frb8v" event={"ID":"ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42","Type":"ContainerDied","Data":"ab9c35c84443abd5384392ddbe947e86dbbc99b0630611d10f21f1953f5394aa"} Dec 01 14:56:25 crc kubenswrapper[4810]: I1201 14:56:25.070872 4810 scope.go:117] "RemoveContainer" containerID="ce152a2dcedab6dfa2222a388bf9973c1bcce7517a6f7e4bf7a4901cbcf8f86a" Dec 01 14:56:25 crc kubenswrapper[4810]: I1201 14:56:25.090706 4810 scope.go:117] "RemoveContainer" containerID="a81f5574834d150e9892017d8ae052bc6da9601608100fab0fac319ce5cd3cf8" Dec 01 14:56:25 crc kubenswrapper[4810]: I1201 14:56:25.102980 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5df84c984d-frb8v"] Dec 01 14:56:25 crc kubenswrapper[4810]: I1201 14:56:25.120662 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5df84c984d-frb8v"] Dec 01 14:56:26 crc kubenswrapper[4810]: I1201 14:56:26.040933 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:56:26 crc kubenswrapper[4810]: I1201 14:56:26.425312 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 01 14:56:26 crc kubenswrapper[4810]: I1201 14:56:26.500589 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42" path="/var/lib/kubelet/pods/ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42/volumes" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.461030 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 14:56:27 crc kubenswrapper[4810]: E1201 14:56:27.461953 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42" containerName="neutron-api" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.461970 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42" containerName="neutron-api" Dec 01 14:56:27 crc kubenswrapper[4810]: E1201 14:56:27.461985 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42" containerName="neutron-httpd" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.461993 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42" containerName="neutron-httpd" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.462223 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42" containerName="neutron-httpd" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.462257 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccf7f439-d5a8-4bcc-a71d-dc6ceae59e42" containerName="neutron-api" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.463030 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.465707 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.465829 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.466027 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-qqn85" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.471591 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.498997 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c35fc22d-59ab-4396-87c5-fadfc17ecc53-openstack-config\") pod \"openstackclient\" (UID: \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\") " pod="openstack/openstackclient" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.499050 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kckmw\" (UniqueName: \"kubernetes.io/projected/c35fc22d-59ab-4396-87c5-fadfc17ecc53-kube-api-access-kckmw\") pod \"openstackclient\" (UID: \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\") " pod="openstack/openstackclient" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.499089 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c35fc22d-59ab-4396-87c5-fadfc17ecc53-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\") " pod="openstack/openstackclient" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.499132 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c35fc22d-59ab-4396-87c5-fadfc17ecc53-openstack-config-secret\") pod \"openstackclient\" (UID: \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\") " pod="openstack/openstackclient" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.600114 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c35fc22d-59ab-4396-87c5-fadfc17ecc53-openstack-config-secret\") pod \"openstackclient\" (UID: \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\") " pod="openstack/openstackclient" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.600252 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c35fc22d-59ab-4396-87c5-fadfc17ecc53-openstack-config\") pod \"openstackclient\" (UID: \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\") " pod="openstack/openstackclient" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.600301 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kckmw\" (UniqueName: \"kubernetes.io/projected/c35fc22d-59ab-4396-87c5-fadfc17ecc53-kube-api-access-kckmw\") pod \"openstackclient\" (UID: \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\") " pod="openstack/openstackclient" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.600349 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c35fc22d-59ab-4396-87c5-fadfc17ecc53-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\") " pod="openstack/openstackclient" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.602072 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c35fc22d-59ab-4396-87c5-fadfc17ecc53-openstack-config\") pod \"openstackclient\" (UID: \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\") " pod="openstack/openstackclient" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.606087 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c35fc22d-59ab-4396-87c5-fadfc17ecc53-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\") " pod="openstack/openstackclient" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.620353 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c35fc22d-59ab-4396-87c5-fadfc17ecc53-openstack-config-secret\") pod \"openstackclient\" (UID: \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\") " pod="openstack/openstackclient" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.620974 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kckmw\" (UniqueName: \"kubernetes.io/projected/c35fc22d-59ab-4396-87c5-fadfc17ecc53-kube-api-access-kckmw\") pod \"openstackclient\" (UID: \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\") " pod="openstack/openstackclient" Dec 01 14:56:27 crc kubenswrapper[4810]: I1201 14:56:27.785522 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 14:56:28 crc kubenswrapper[4810]: I1201 14:56:28.276639 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 14:56:29 crc kubenswrapper[4810]: I1201 14:56:29.108861 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"c35fc22d-59ab-4396-87c5-fadfc17ecc53","Type":"ContainerStarted","Data":"20d0aa4e26b165bc3f4355044df83264b401d34a1b27cc603c5dce4b4a255062"} Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.142799 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6c8f8b5cf7-65pxh"] Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.145111 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.147705 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.147859 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.147950 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.174866 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6c8f8b5cf7-65pxh"] Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.282815 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-config-data\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.282862 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrx4f\" (UniqueName: \"kubernetes.io/projected/7478f42b-4be2-430b-9854-5930dccaee43-kube-api-access-mrx4f\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.282895 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-internal-tls-certs\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.283133 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7478f42b-4be2-430b-9854-5930dccaee43-log-httpd\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.283335 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7478f42b-4be2-430b-9854-5930dccaee43-run-httpd\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.283409 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-combined-ca-bundle\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.283493 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-public-tls-certs\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.283578 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7478f42b-4be2-430b-9854-5930dccaee43-etc-swift\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.385753 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-internal-tls-certs\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.385822 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7478f42b-4be2-430b-9854-5930dccaee43-log-httpd\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.385871 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7478f42b-4be2-430b-9854-5930dccaee43-run-httpd\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.385898 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-combined-ca-bundle\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.385921 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-public-tls-certs\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.385945 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7478f42b-4be2-430b-9854-5930dccaee43-etc-swift\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.386018 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-config-data\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.386042 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrx4f\" (UniqueName: \"kubernetes.io/projected/7478f42b-4be2-430b-9854-5930dccaee43-kube-api-access-mrx4f\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.386662 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7478f42b-4be2-430b-9854-5930dccaee43-run-httpd\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.386967 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7478f42b-4be2-430b-9854-5930dccaee43-log-httpd\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.391725 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-internal-tls-certs\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.392209 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-config-data\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.392455 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-combined-ca-bundle\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.392981 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-public-tls-certs\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.396305 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7478f42b-4be2-430b-9854-5930dccaee43-etc-swift\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.405755 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrx4f\" (UniqueName: \"kubernetes.io/projected/7478f42b-4be2-430b-9854-5930dccaee43-kube-api-access-mrx4f\") pod \"swift-proxy-6c8f8b5cf7-65pxh\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.465082 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:31 crc kubenswrapper[4810]: I1201 14:56:31.832173 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 01 14:56:32 crc kubenswrapper[4810]: I1201 14:56:32.090890 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6c8f8b5cf7-65pxh"] Dec 01 14:56:32 crc kubenswrapper[4810]: I1201 14:56:32.153233 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" event={"ID":"7478f42b-4be2-430b-9854-5930dccaee43","Type":"ContainerStarted","Data":"4cfb038fda022537c9c1494bc93f9700a37720afcac62a202c221904f4195331"} Dec 01 14:56:32 crc kubenswrapper[4810]: I1201 14:56:32.205314 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:56:32 crc kubenswrapper[4810]: I1201 14:56:32.205729 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerName="ceilometer-central-agent" containerID="cri-o://040cad2764ca8aa7273ce97e369d59aed9af01a3cc2aef0ac5e4e35d012e003a" gracePeriod=30 Dec 01 14:56:32 crc kubenswrapper[4810]: I1201 14:56:32.205868 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerName="proxy-httpd" containerID="cri-o://0720428d31bf19f31e912919435903191f3bcb90677211a516c8c3dbd918a037" gracePeriod=30 Dec 01 14:56:32 crc kubenswrapper[4810]: I1201 14:56:32.205928 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerName="sg-core" containerID="cri-o://5c9a3f70bc821a5731836c8aed647e817a723b5616565d4bf6d518bff2fdb117" gracePeriod=30 Dec 01 14:56:32 crc kubenswrapper[4810]: I1201 14:56:32.205977 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerName="ceilometer-notification-agent" containerID="cri-o://6f2250ffa5a8fa4cbd93aff65636f6a04fd9b92fdbf1ffbc4e97507590ff20a3" gracePeriod=30 Dec 01 14:56:32 crc kubenswrapper[4810]: I1201 14:56:32.211551 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 14:56:33 crc kubenswrapper[4810]: I1201 14:56:33.064280 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 14:56:33 crc kubenswrapper[4810]: I1201 14:56:33.064853 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3039d59e-4377-411d-98e7-9fd84a2d93b6" containerName="glance-log" containerID="cri-o://ad42623d524fd2d5246b4533a6cc14b8730ace1654df85830c52e33fdeecce05" gracePeriod=30 Dec 01 14:56:33 crc kubenswrapper[4810]: I1201 14:56:33.064930 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3039d59e-4377-411d-98e7-9fd84a2d93b6" containerName="glance-httpd" containerID="cri-o://0ce22d98360b8f8557cdeab1c36e735725358d0cfe0253144934e0ceb6ba1349" gracePeriod=30 Dec 01 14:56:33 crc kubenswrapper[4810]: I1201 14:56:33.190108 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" event={"ID":"7478f42b-4be2-430b-9854-5930dccaee43","Type":"ContainerStarted","Data":"1cc6810ebfaa94211940478b494df0cc163de04d06f714491ab4d71d2b9f032d"} Dec 01 14:56:33 crc kubenswrapper[4810]: I1201 14:56:33.190150 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" event={"ID":"7478f42b-4be2-430b-9854-5930dccaee43","Type":"ContainerStarted","Data":"65c24e7fbab051767fd3a4645fd6dfed9e0a34a7bc48306bde5ba5dbb76b4aad"} Dec 01 14:56:33 crc kubenswrapper[4810]: I1201 14:56:33.190256 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:33 crc kubenswrapper[4810]: I1201 14:56:33.190277 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:33 crc kubenswrapper[4810]: I1201 14:56:33.218664 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" podStartSLOduration=2.21864882 podStartE2EDuration="2.21864882s" podCreationTimestamp="2025-12-01 14:56:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:56:33.214497578 +0000 UTC m=+1358.978007171" watchObservedRunningTime="2025-12-01 14:56:33.21864882 +0000 UTC m=+1358.982158423" Dec 01 14:56:33 crc kubenswrapper[4810]: I1201 14:56:33.222934 4810 generic.go:334] "Generic (PLEG): container finished" podID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerID="0720428d31bf19f31e912919435903191f3bcb90677211a516c8c3dbd918a037" exitCode=0 Dec 01 14:56:33 crc kubenswrapper[4810]: I1201 14:56:33.222986 4810 generic.go:334] "Generic (PLEG): container finished" podID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerID="5c9a3f70bc821a5731836c8aed647e817a723b5616565d4bf6d518bff2fdb117" exitCode=2 Dec 01 14:56:33 crc kubenswrapper[4810]: I1201 14:56:33.222998 4810 generic.go:334] "Generic (PLEG): container finished" podID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerID="040cad2764ca8aa7273ce97e369d59aed9af01a3cc2aef0ac5e4e35d012e003a" exitCode=0 Dec 01 14:56:33 crc kubenswrapper[4810]: I1201 14:56:33.223036 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a454ac3-a4dc-41c9-b548-3b57cad82bc5","Type":"ContainerDied","Data":"0720428d31bf19f31e912919435903191f3bcb90677211a516c8c3dbd918a037"} Dec 01 14:56:33 crc kubenswrapper[4810]: I1201 14:56:33.223086 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a454ac3-a4dc-41c9-b548-3b57cad82bc5","Type":"ContainerDied","Data":"5c9a3f70bc821a5731836c8aed647e817a723b5616565d4bf6d518bff2fdb117"} Dec 01 14:56:33 crc kubenswrapper[4810]: I1201 14:56:33.223104 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a454ac3-a4dc-41c9-b548-3b57cad82bc5","Type":"ContainerDied","Data":"040cad2764ca8aa7273ce97e369d59aed9af01a3cc2aef0ac5e4e35d012e003a"} Dec 01 14:56:33 crc kubenswrapper[4810]: E1201 14:56:33.284410 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3039d59e_4377_411d_98e7_9fd84a2d93b6.slice/crio-conmon-ad42623d524fd2d5246b4533a6cc14b8730ace1654df85830c52e33fdeecce05.scope\": RecentStats: unable to find data in memory cache]" Dec 01 14:56:34 crc kubenswrapper[4810]: I1201 14:56:34.239444 4810 generic.go:334] "Generic (PLEG): container finished" podID="3039d59e-4377-411d-98e7-9fd84a2d93b6" containerID="ad42623d524fd2d5246b4533a6cc14b8730ace1654df85830c52e33fdeecce05" exitCode=143 Dec 01 14:56:34 crc kubenswrapper[4810]: I1201 14:56:34.239517 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3039d59e-4377-411d-98e7-9fd84a2d93b6","Type":"ContainerDied","Data":"ad42623d524fd2d5246b4533a6cc14b8730ace1654df85830c52e33fdeecce05"} Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.029147 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.030431 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d18c05bc-3e20-4305-b8bf-d45e6ff46308" containerName="glance-log" containerID="cri-o://373d8ee5e28ab65af88de3532b2fd0604c91e4cbc4b1d3f1e4e86653bfc62288" gracePeriod=30 Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.030500 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d18c05bc-3e20-4305-b8bf-d45e6ff46308" containerName="glance-httpd" containerID="cri-o://11cb52869bfafc31bb1d1d214f6a4b58a7e39a377578ec37c708f280ae68b49f" gracePeriod=30 Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.056866 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-2fflr"] Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.058380 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2fflr" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.065128 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-2fflr"] Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.108069 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/916e9520-9e98-42a8-bbb5-cb5f4029daab-operator-scripts\") pod \"nova-api-db-create-2fflr\" (UID: \"916e9520-9e98-42a8-bbb5-cb5f4029daab\") " pod="openstack/nova-api-db-create-2fflr" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.108179 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26vh7\" (UniqueName: \"kubernetes.io/projected/916e9520-9e98-42a8-bbb5-cb5f4029daab-kube-api-access-26vh7\") pod \"nova-api-db-create-2fflr\" (UID: \"916e9520-9e98-42a8-bbb5-cb5f4029daab\") " pod="openstack/nova-api-db-create-2fflr" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.161086 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-c0fd-account-create-update-w9l6n"] Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.162530 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c0fd-account-create-update-w9l6n" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.165812 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.185134 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c0fd-account-create-update-w9l6n"] Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.213761 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/916e9520-9e98-42a8-bbb5-cb5f4029daab-operator-scripts\") pod \"nova-api-db-create-2fflr\" (UID: \"916e9520-9e98-42a8-bbb5-cb5f4029daab\") " pod="openstack/nova-api-db-create-2fflr" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.213827 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75872443-9a8a-4b7b-9c70-036dc558594e-operator-scripts\") pod \"nova-api-c0fd-account-create-update-w9l6n\" (UID: \"75872443-9a8a-4b7b-9c70-036dc558594e\") " pod="openstack/nova-api-c0fd-account-create-update-w9l6n" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.213861 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtjpw\" (UniqueName: \"kubernetes.io/projected/75872443-9a8a-4b7b-9c70-036dc558594e-kube-api-access-qtjpw\") pod \"nova-api-c0fd-account-create-update-w9l6n\" (UID: \"75872443-9a8a-4b7b-9c70-036dc558594e\") " pod="openstack/nova-api-c0fd-account-create-update-w9l6n" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.213914 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26vh7\" (UniqueName: \"kubernetes.io/projected/916e9520-9e98-42a8-bbb5-cb5f4029daab-kube-api-access-26vh7\") pod \"nova-api-db-create-2fflr\" (UID: \"916e9520-9e98-42a8-bbb5-cb5f4029daab\") " pod="openstack/nova-api-db-create-2fflr" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.215668 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/916e9520-9e98-42a8-bbb5-cb5f4029daab-operator-scripts\") pod \"nova-api-db-create-2fflr\" (UID: \"916e9520-9e98-42a8-bbb5-cb5f4029daab\") " pod="openstack/nova-api-db-create-2fflr" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.242743 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26vh7\" (UniqueName: \"kubernetes.io/projected/916e9520-9e98-42a8-bbb5-cb5f4029daab-kube-api-access-26vh7\") pod \"nova-api-db-create-2fflr\" (UID: \"916e9520-9e98-42a8-bbb5-cb5f4029daab\") " pod="openstack/nova-api-db-create-2fflr" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.279343 4810 generic.go:334] "Generic (PLEG): container finished" podID="d18c05bc-3e20-4305-b8bf-d45e6ff46308" containerID="373d8ee5e28ab65af88de3532b2fd0604c91e4cbc4b1d3f1e4e86653bfc62288" exitCode=143 Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.279429 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d18c05bc-3e20-4305-b8bf-d45e6ff46308","Type":"ContainerDied","Data":"373d8ee5e28ab65af88de3532b2fd0604c91e4cbc4b1d3f1e4e86653bfc62288"} Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.315960 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75872443-9a8a-4b7b-9c70-036dc558594e-operator-scripts\") pod \"nova-api-c0fd-account-create-update-w9l6n\" (UID: \"75872443-9a8a-4b7b-9c70-036dc558594e\") " pod="openstack/nova-api-c0fd-account-create-update-w9l6n" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.316012 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtjpw\" (UniqueName: \"kubernetes.io/projected/75872443-9a8a-4b7b-9c70-036dc558594e-kube-api-access-qtjpw\") pod \"nova-api-c0fd-account-create-update-w9l6n\" (UID: \"75872443-9a8a-4b7b-9c70-036dc558594e\") " pod="openstack/nova-api-c0fd-account-create-update-w9l6n" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.317115 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75872443-9a8a-4b7b-9c70-036dc558594e-operator-scripts\") pod \"nova-api-c0fd-account-create-update-w9l6n\" (UID: \"75872443-9a8a-4b7b-9c70-036dc558594e\") " pod="openstack/nova-api-c0fd-account-create-update-w9l6n" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.343095 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtjpw\" (UniqueName: \"kubernetes.io/projected/75872443-9a8a-4b7b-9c70-036dc558594e-kube-api-access-qtjpw\") pod \"nova-api-c0fd-account-create-update-w9l6n\" (UID: \"75872443-9a8a-4b7b-9c70-036dc558594e\") " pod="openstack/nova-api-c0fd-account-create-update-w9l6n" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.353073 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-c258c"] Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.354215 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-c258c" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.362843 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-96e4-account-create-update-jgpfh"] Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.364041 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-96e4-account-create-update-jgpfh" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.366088 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.438432 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2fflr" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.440504 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cglvd\" (UniqueName: \"kubernetes.io/projected/1e760d14-b5f0-48c8-927e-09d086591dce-kube-api-access-cglvd\") pod \"nova-cell0-96e4-account-create-update-jgpfh\" (UID: \"1e760d14-b5f0-48c8-927e-09d086591dce\") " pod="openstack/nova-cell0-96e4-account-create-update-jgpfh" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.440568 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e760d14-b5f0-48c8-927e-09d086591dce-operator-scripts\") pod \"nova-cell0-96e4-account-create-update-jgpfh\" (UID: \"1e760d14-b5f0-48c8-927e-09d086591dce\") " pod="openstack/nova-cell0-96e4-account-create-update-jgpfh" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.453572 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-c258c"] Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.462483 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-96e4-account-create-update-jgpfh"] Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.489890 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c0fd-account-create-update-w9l6n" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.533659 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-r8tkk"] Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.535804 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-r8tkk" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.541861 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cglvd\" (UniqueName: \"kubernetes.io/projected/1e760d14-b5f0-48c8-927e-09d086591dce-kube-api-access-cglvd\") pod \"nova-cell0-96e4-account-create-update-jgpfh\" (UID: \"1e760d14-b5f0-48c8-927e-09d086591dce\") " pod="openstack/nova-cell0-96e4-account-create-update-jgpfh" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.541905 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckbtm\" (UniqueName: \"kubernetes.io/projected/20b33096-654a-4684-971a-bd81b90d0a79-kube-api-access-ckbtm\") pod \"nova-cell1-db-create-r8tkk\" (UID: \"20b33096-654a-4684-971a-bd81b90d0a79\") " pod="openstack/nova-cell1-db-create-r8tkk" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.541934 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e760d14-b5f0-48c8-927e-09d086591dce-operator-scripts\") pod \"nova-cell0-96e4-account-create-update-jgpfh\" (UID: \"1e760d14-b5f0-48c8-927e-09d086591dce\") " pod="openstack/nova-cell0-96e4-account-create-update-jgpfh" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.541976 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b33096-654a-4684-971a-bd81b90d0a79-operator-scripts\") pod \"nova-cell1-db-create-r8tkk\" (UID: \"20b33096-654a-4684-971a-bd81b90d0a79\") " pod="openstack/nova-cell1-db-create-r8tkk" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.542016 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btcdt\" (UniqueName: \"kubernetes.io/projected/fc8eec67-10ea-427f-b6f7-045c1dc7519f-kube-api-access-btcdt\") pod \"nova-cell0-db-create-c258c\" (UID: \"fc8eec67-10ea-427f-b6f7-045c1dc7519f\") " pod="openstack/nova-cell0-db-create-c258c" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.542036 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc8eec67-10ea-427f-b6f7-045c1dc7519f-operator-scripts\") pod \"nova-cell0-db-create-c258c\" (UID: \"fc8eec67-10ea-427f-b6f7-045c1dc7519f\") " pod="openstack/nova-cell0-db-create-c258c" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.544953 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e760d14-b5f0-48c8-927e-09d086591dce-operator-scripts\") pod \"nova-cell0-96e4-account-create-update-jgpfh\" (UID: \"1e760d14-b5f0-48c8-927e-09d086591dce\") " pod="openstack/nova-cell0-96e4-account-create-update-jgpfh" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.566910 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-r8tkk"] Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.570062 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cglvd\" (UniqueName: \"kubernetes.io/projected/1e760d14-b5f0-48c8-927e-09d086591dce-kube-api-access-cglvd\") pod \"nova-cell0-96e4-account-create-update-jgpfh\" (UID: \"1e760d14-b5f0-48c8-927e-09d086591dce\") " pod="openstack/nova-cell0-96e4-account-create-update-jgpfh" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.623495 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-c683-account-create-update-m9bvk"] Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.624620 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c683-account-create-update-m9bvk" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.627372 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.631361 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-c683-account-create-update-m9bvk"] Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.644282 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btcdt\" (UniqueName: \"kubernetes.io/projected/fc8eec67-10ea-427f-b6f7-045c1dc7519f-kube-api-access-btcdt\") pod \"nova-cell0-db-create-c258c\" (UID: \"fc8eec67-10ea-427f-b6f7-045c1dc7519f\") " pod="openstack/nova-cell0-db-create-c258c" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.644347 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc8eec67-10ea-427f-b6f7-045c1dc7519f-operator-scripts\") pod \"nova-cell0-db-create-c258c\" (UID: \"fc8eec67-10ea-427f-b6f7-045c1dc7519f\") " pod="openstack/nova-cell0-db-create-c258c" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.644488 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckbtm\" (UniqueName: \"kubernetes.io/projected/20b33096-654a-4684-971a-bd81b90d0a79-kube-api-access-ckbtm\") pod \"nova-cell1-db-create-r8tkk\" (UID: \"20b33096-654a-4684-971a-bd81b90d0a79\") " pod="openstack/nova-cell1-db-create-r8tkk" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.644571 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6346394-0edb-4b7b-85fd-47907ea08f08-operator-scripts\") pod \"nova-cell1-c683-account-create-update-m9bvk\" (UID: \"f6346394-0edb-4b7b-85fd-47907ea08f08\") " pod="openstack/nova-cell1-c683-account-create-update-m9bvk" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.644597 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b33096-654a-4684-971a-bd81b90d0a79-operator-scripts\") pod \"nova-cell1-db-create-r8tkk\" (UID: \"20b33096-654a-4684-971a-bd81b90d0a79\") " pod="openstack/nova-cell1-db-create-r8tkk" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.644617 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvb7h\" (UniqueName: \"kubernetes.io/projected/f6346394-0edb-4b7b-85fd-47907ea08f08-kube-api-access-hvb7h\") pod \"nova-cell1-c683-account-create-update-m9bvk\" (UID: \"f6346394-0edb-4b7b-85fd-47907ea08f08\") " pod="openstack/nova-cell1-c683-account-create-update-m9bvk" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.645953 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b33096-654a-4684-971a-bd81b90d0a79-operator-scripts\") pod \"nova-cell1-db-create-r8tkk\" (UID: \"20b33096-654a-4684-971a-bd81b90d0a79\") " pod="openstack/nova-cell1-db-create-r8tkk" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.646126 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc8eec67-10ea-427f-b6f7-045c1dc7519f-operator-scripts\") pod \"nova-cell0-db-create-c258c\" (UID: \"fc8eec67-10ea-427f-b6f7-045c1dc7519f\") " pod="openstack/nova-cell0-db-create-c258c" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.672184 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckbtm\" (UniqueName: \"kubernetes.io/projected/20b33096-654a-4684-971a-bd81b90d0a79-kube-api-access-ckbtm\") pod \"nova-cell1-db-create-r8tkk\" (UID: \"20b33096-654a-4684-971a-bd81b90d0a79\") " pod="openstack/nova-cell1-db-create-r8tkk" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.678947 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btcdt\" (UniqueName: \"kubernetes.io/projected/fc8eec67-10ea-427f-b6f7-045c1dc7519f-kube-api-access-btcdt\") pod \"nova-cell0-db-create-c258c\" (UID: \"fc8eec67-10ea-427f-b6f7-045c1dc7519f\") " pod="openstack/nova-cell0-db-create-c258c" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.746550 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6346394-0edb-4b7b-85fd-47907ea08f08-operator-scripts\") pod \"nova-cell1-c683-account-create-update-m9bvk\" (UID: \"f6346394-0edb-4b7b-85fd-47907ea08f08\") " pod="openstack/nova-cell1-c683-account-create-update-m9bvk" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.746625 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvb7h\" (UniqueName: \"kubernetes.io/projected/f6346394-0edb-4b7b-85fd-47907ea08f08-kube-api-access-hvb7h\") pod \"nova-cell1-c683-account-create-update-m9bvk\" (UID: \"f6346394-0edb-4b7b-85fd-47907ea08f08\") " pod="openstack/nova-cell1-c683-account-create-update-m9bvk" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.747199 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6346394-0edb-4b7b-85fd-47907ea08f08-operator-scripts\") pod \"nova-cell1-c683-account-create-update-m9bvk\" (UID: \"f6346394-0edb-4b7b-85fd-47907ea08f08\") " pod="openstack/nova-cell1-c683-account-create-update-m9bvk" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.765388 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvb7h\" (UniqueName: \"kubernetes.io/projected/f6346394-0edb-4b7b-85fd-47907ea08f08-kube-api-access-hvb7h\") pod \"nova-cell1-c683-account-create-update-m9bvk\" (UID: \"f6346394-0edb-4b7b-85fd-47907ea08f08\") " pod="openstack/nova-cell1-c683-account-create-update-m9bvk" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.785040 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-c258c" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.798610 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-96e4-account-create-update-jgpfh" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.866269 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-r8tkk" Dec 01 14:56:36 crc kubenswrapper[4810]: I1201 14:56:36.955202 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c683-account-create-update-m9bvk" Dec 01 14:56:37 crc kubenswrapper[4810]: I1201 14:56:37.315072 4810 generic.go:334] "Generic (PLEG): container finished" podID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerID="6f2250ffa5a8fa4cbd93aff65636f6a04fd9b92fdbf1ffbc4e97507590ff20a3" exitCode=0 Dec 01 14:56:37 crc kubenswrapper[4810]: I1201 14:56:37.315162 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a454ac3-a4dc-41c9-b548-3b57cad82bc5","Type":"ContainerDied","Data":"6f2250ffa5a8fa4cbd93aff65636f6a04fd9b92fdbf1ffbc4e97507590ff20a3"} Dec 01 14:56:37 crc kubenswrapper[4810]: I1201 14:56:37.330205 4810 generic.go:334] "Generic (PLEG): container finished" podID="3039d59e-4377-411d-98e7-9fd84a2d93b6" containerID="0ce22d98360b8f8557cdeab1c36e735725358d0cfe0253144934e0ceb6ba1349" exitCode=0 Dec 01 14:56:37 crc kubenswrapper[4810]: I1201 14:56:37.330257 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3039d59e-4377-411d-98e7-9fd84a2d93b6","Type":"ContainerDied","Data":"0ce22d98360b8f8557cdeab1c36e735725358d0cfe0253144934e0ceb6ba1349"} Dec 01 14:56:38 crc kubenswrapper[4810]: I1201 14:56:38.222150 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.161:3000/\": dial tcp 10.217.0.161:3000: connect: connection refused" Dec 01 14:56:40 crc kubenswrapper[4810]: I1201 14:56:40.366871 4810 generic.go:334] "Generic (PLEG): container finished" podID="d18c05bc-3e20-4305-b8bf-d45e6ff46308" containerID="11cb52869bfafc31bb1d1d214f6a4b58a7e39a377578ec37c708f280ae68b49f" exitCode=0 Dec 01 14:56:40 crc kubenswrapper[4810]: I1201 14:56:40.366914 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d18c05bc-3e20-4305-b8bf-d45e6ff46308","Type":"ContainerDied","Data":"11cb52869bfafc31bb1d1d214f6a4b58a7e39a377578ec37c708f280ae68b49f"} Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.192032 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.318375 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.353356 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-combined-ca-bundle\") pod \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.353459 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-run-httpd\") pod \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.353512 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-config-data\") pod \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.353535 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-log-httpd\") pod \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.353577 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-scripts\") pod \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.353657 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9twtd\" (UniqueName: \"kubernetes.io/projected/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-kube-api-access-9twtd\") pod \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.353706 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-sg-core-conf-yaml\") pod \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\" (UID: \"1a454ac3-a4dc-41c9-b548-3b57cad82bc5\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.357033 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1a454ac3-a4dc-41c9-b548-3b57cad82bc5" (UID: "1a454ac3-a4dc-41c9-b548-3b57cad82bc5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.357425 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1a454ac3-a4dc-41c9-b548-3b57cad82bc5" (UID: "1a454ac3-a4dc-41c9-b548-3b57cad82bc5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.372956 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-kube-api-access-9twtd" (OuterVolumeSpecName: "kube-api-access-9twtd") pod "1a454ac3-a4dc-41c9-b548-3b57cad82bc5" (UID: "1a454ac3-a4dc-41c9-b548-3b57cad82bc5"). InnerVolumeSpecName "kube-api-access-9twtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.376583 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-scripts" (OuterVolumeSpecName: "scripts") pod "1a454ac3-a4dc-41c9-b548-3b57cad82bc5" (UID: "1a454ac3-a4dc-41c9-b548-3b57cad82bc5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.394310 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"c35fc22d-59ab-4396-87c5-fadfc17ecc53","Type":"ContainerStarted","Data":"a303d0364f2f0ad764ac7d54b67330e695ac583a1627abd27d1f6447c9de1ee2"} Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.401683 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1a454ac3-a4dc-41c9-b548-3b57cad82bc5" (UID: "1a454ac3-a4dc-41c9-b548-3b57cad82bc5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.404441 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1a454ac3-a4dc-41c9-b548-3b57cad82bc5","Type":"ContainerDied","Data":"29804ee1f5096e755708f4ef3aa4bd594ed3ebe547d076ebcc7bd7a8474f45c9"} Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.404461 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.404552 4810 scope.go:117] "RemoveContainer" containerID="0720428d31bf19f31e912919435903191f3bcb90677211a516c8c3dbd918a037" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.412601 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3039d59e-4377-411d-98e7-9fd84a2d93b6","Type":"ContainerDied","Data":"8ad63be1f8544e127ea84a20e8394c29ddf23c0d67240904b40820218c663e4d"} Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.412706 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.420090 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.7038875679999999 podStartE2EDuration="14.420073954s" podCreationTimestamp="2025-12-01 14:56:27 +0000 UTC" firstStartedPulling="2025-12-01 14:56:28.313873494 +0000 UTC m=+1354.077383097" lastFinishedPulling="2025-12-01 14:56:41.03005988 +0000 UTC m=+1366.793569483" observedRunningTime="2025-12-01 14:56:41.409737663 +0000 UTC m=+1367.173247266" watchObservedRunningTime="2025-12-01 14:56:41.420073954 +0000 UTC m=+1367.183583557" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.430681 4810 scope.go:117] "RemoveContainer" containerID="5c9a3f70bc821a5731836c8aed647e817a723b5616565d4bf6d518bff2fdb117" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.455989 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"3039d59e-4377-411d-98e7-9fd84a2d93b6\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.456517 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swxpr\" (UniqueName: \"kubernetes.io/projected/3039d59e-4377-411d-98e7-9fd84a2d93b6-kube-api-access-swxpr\") pod \"3039d59e-4377-411d-98e7-9fd84a2d93b6\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.456862 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-public-tls-certs\") pod \"3039d59e-4377-411d-98e7-9fd84a2d93b6\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.456889 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3039d59e-4377-411d-98e7-9fd84a2d93b6-logs\") pod \"3039d59e-4377-411d-98e7-9fd84a2d93b6\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.456934 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-config-data\") pod \"3039d59e-4377-411d-98e7-9fd84a2d93b6\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.456961 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-scripts\") pod \"3039d59e-4377-411d-98e7-9fd84a2d93b6\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.456983 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-combined-ca-bundle\") pod \"3039d59e-4377-411d-98e7-9fd84a2d93b6\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.457300 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3039d59e-4377-411d-98e7-9fd84a2d93b6-httpd-run\") pod \"3039d59e-4377-411d-98e7-9fd84a2d93b6\" (UID: \"3039d59e-4377-411d-98e7-9fd84a2d93b6\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.458020 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3039d59e-4377-411d-98e7-9fd84a2d93b6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3039d59e-4377-411d-98e7-9fd84a2d93b6" (UID: "3039d59e-4377-411d-98e7-9fd84a2d93b6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.457931 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3039d59e-4377-411d-98e7-9fd84a2d93b6-logs" (OuterVolumeSpecName: "logs") pod "3039d59e-4377-411d-98e7-9fd84a2d93b6" (UID: "3039d59e-4377-411d-98e7-9fd84a2d93b6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.458402 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3039d59e-4377-411d-98e7-9fd84a2d93b6-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.458421 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.458429 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.458437 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.458447 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3039d59e-4377-411d-98e7-9fd84a2d93b6-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.458488 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9twtd\" (UniqueName: \"kubernetes.io/projected/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-kube-api-access-9twtd\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.458498 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.462662 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3039d59e-4377-411d-98e7-9fd84a2d93b6-kube-api-access-swxpr" (OuterVolumeSpecName: "kube-api-access-swxpr") pod "3039d59e-4377-411d-98e7-9fd84a2d93b6" (UID: "3039d59e-4377-411d-98e7-9fd84a2d93b6"). InnerVolumeSpecName "kube-api-access-swxpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.466811 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-scripts" (OuterVolumeSpecName: "scripts") pod "3039d59e-4377-411d-98e7-9fd84a2d93b6" (UID: "3039d59e-4377-411d-98e7-9fd84a2d93b6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.468966 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "3039d59e-4377-411d-98e7-9fd84a2d93b6" (UID: "3039d59e-4377-411d-98e7-9fd84a2d93b6"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.471745 4810 scope.go:117] "RemoveContainer" containerID="6f2250ffa5a8fa4cbd93aff65636f6a04fd9b92fdbf1ffbc4e97507590ff20a3" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.492792 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.499958 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.500071 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3039d59e-4377-411d-98e7-9fd84a2d93b6" (UID: "3039d59e-4377-411d-98e7-9fd84a2d93b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.509569 4810 scope.go:117] "RemoveContainer" containerID="040cad2764ca8aa7273ce97e369d59aed9af01a3cc2aef0ac5e4e35d012e003a" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.547551 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-config-data" (OuterVolumeSpecName: "config-data") pod "1a454ac3-a4dc-41c9-b548-3b57cad82bc5" (UID: "1a454ac3-a4dc-41c9-b548-3b57cad82bc5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.554755 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a454ac3-a4dc-41c9-b548-3b57cad82bc5" (UID: "1a454ac3-a4dc-41c9-b548-3b57cad82bc5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.557104 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.560687 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.560724 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a454ac3-a4dc-41c9-b548-3b57cad82bc5-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.560752 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.560767 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swxpr\" (UniqueName: \"kubernetes.io/projected/3039d59e-4377-411d-98e7-9fd84a2d93b6-kube-api-access-swxpr\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.560780 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.560792 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.577998 4810 scope.go:117] "RemoveContainer" containerID="0ce22d98360b8f8557cdeab1c36e735725358d0cfe0253144934e0ceb6ba1349" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.606020 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3039d59e-4377-411d-98e7-9fd84a2d93b6" (UID: "3039d59e-4377-411d-98e7-9fd84a2d93b6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.631411 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.659563 4810 scope.go:117] "RemoveContainer" containerID="ad42623d524fd2d5246b4533a6cc14b8730ace1654df85830c52e33fdeecce05" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.662165 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-scripts\") pod \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.662224 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs2g9\" (UniqueName: \"kubernetes.io/projected/d18c05bc-3e20-4305-b8bf-d45e6ff46308-kube-api-access-rs2g9\") pod \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.662265 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d18c05bc-3e20-4305-b8bf-d45e6ff46308-httpd-run\") pod \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.662318 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.662358 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-config-data\") pod \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.662450 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d18c05bc-3e20-4305-b8bf-d45e6ff46308-logs\") pod \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.662512 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-internal-tls-certs\") pod \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.662580 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-combined-ca-bundle\") pod \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\" (UID: \"d18c05bc-3e20-4305-b8bf-d45e6ff46308\") " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.662947 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d18c05bc-3e20-4305-b8bf-d45e6ff46308-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d18c05bc-3e20-4305-b8bf-d45e6ff46308" (UID: "d18c05bc-3e20-4305-b8bf-d45e6ff46308"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.663302 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.663323 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d18c05bc-3e20-4305-b8bf-d45e6ff46308-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.663336 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.663376 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d18c05bc-3e20-4305-b8bf-d45e6ff46308-logs" (OuterVolumeSpecName: "logs") pod "d18c05bc-3e20-4305-b8bf-d45e6ff46308" (UID: "d18c05bc-3e20-4305-b8bf-d45e6ff46308"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.665965 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-config-data" (OuterVolumeSpecName: "config-data") pod "3039d59e-4377-411d-98e7-9fd84a2d93b6" (UID: "3039d59e-4377-411d-98e7-9fd84a2d93b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.669215 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-c683-account-create-update-m9bvk"] Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.669714 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-scripts" (OuterVolumeSpecName: "scripts") pod "d18c05bc-3e20-4305-b8bf-d45e6ff46308" (UID: "d18c05bc-3e20-4305-b8bf-d45e6ff46308"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.669797 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "d18c05bc-3e20-4305-b8bf-d45e6ff46308" (UID: "d18c05bc-3e20-4305-b8bf-d45e6ff46308"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.669799 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d18c05bc-3e20-4305-b8bf-d45e6ff46308-kube-api-access-rs2g9" (OuterVolumeSpecName: "kube-api-access-rs2g9") pod "d18c05bc-3e20-4305-b8bf-d45e6ff46308" (UID: "d18c05bc-3e20-4305-b8bf-d45e6ff46308"). InnerVolumeSpecName "kube-api-access-rs2g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.736895 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-config-data" (OuterVolumeSpecName: "config-data") pod "d18c05bc-3e20-4305-b8bf-d45e6ff46308" (UID: "d18c05bc-3e20-4305-b8bf-d45e6ff46308"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.753223 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d18c05bc-3e20-4305-b8bf-d45e6ff46308" (UID: "d18c05bc-3e20-4305-b8bf-d45e6ff46308"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.764491 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.764518 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.764530 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs2g9\" (UniqueName: \"kubernetes.io/projected/d18c05bc-3e20-4305-b8bf-d45e6ff46308-kube-api-access-rs2g9\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.764552 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.764563 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.764573 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3039d59e-4377-411d-98e7-9fd84a2d93b6-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.764580 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d18c05bc-3e20-4305-b8bf-d45e6ff46308-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.797356 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.809993 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.827816 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.846541 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:56:41 crc kubenswrapper[4810]: E1201 14:56:41.846939 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d18c05bc-3e20-4305-b8bf-d45e6ff46308" containerName="glance-httpd" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.846955 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d18c05bc-3e20-4305-b8bf-d45e6ff46308" containerName="glance-httpd" Dec 01 14:56:41 crc kubenswrapper[4810]: E1201 14:56:41.846977 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerName="proxy-httpd" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.846987 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerName="proxy-httpd" Dec 01 14:56:41 crc kubenswrapper[4810]: E1201 14:56:41.846996 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d18c05bc-3e20-4305-b8bf-d45e6ff46308" containerName="glance-log" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.847004 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d18c05bc-3e20-4305-b8bf-d45e6ff46308" containerName="glance-log" Dec 01 14:56:41 crc kubenswrapper[4810]: E1201 14:56:41.847024 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerName="sg-core" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.847031 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerName="sg-core" Dec 01 14:56:41 crc kubenswrapper[4810]: E1201 14:56:41.847041 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerName="ceilometer-notification-agent" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.847050 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerName="ceilometer-notification-agent" Dec 01 14:56:41 crc kubenswrapper[4810]: E1201 14:56:41.847059 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3039d59e-4377-411d-98e7-9fd84a2d93b6" containerName="glance-log" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.847065 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3039d59e-4377-411d-98e7-9fd84a2d93b6" containerName="glance-log" Dec 01 14:56:41 crc kubenswrapper[4810]: E1201 14:56:41.847080 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3039d59e-4377-411d-98e7-9fd84a2d93b6" containerName="glance-httpd" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.847086 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3039d59e-4377-411d-98e7-9fd84a2d93b6" containerName="glance-httpd" Dec 01 14:56:41 crc kubenswrapper[4810]: E1201 14:56:41.847110 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerName="ceilometer-central-agent" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.847116 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerName="ceilometer-central-agent" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.847283 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d18c05bc-3e20-4305-b8bf-d45e6ff46308" containerName="glance-httpd" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.847293 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d18c05bc-3e20-4305-b8bf-d45e6ff46308" containerName="glance-log" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.847306 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerName="ceilometer-notification-agent" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.847319 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerName="sg-core" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.847327 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3039d59e-4377-411d-98e7-9fd84a2d93b6" containerName="glance-log" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.847337 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerName="proxy-httpd" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.847349 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3039d59e-4377-411d-98e7-9fd84a2d93b6" containerName="glance-httpd" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.847360 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" containerName="ceilometer-central-agent" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.848889 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.853925 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.854164 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.859988 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.861231 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.870519 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.871603 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.885513 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d18c05bc-3e20-4305-b8bf-d45e6ff46308" (UID: "d18c05bc-3e20-4305-b8bf-d45e6ff46308"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.913522 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.915844 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.934728 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.939647 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.944522 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.974244 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx7mz\" (UniqueName: \"kubernetes.io/projected/9dd705f6-0d8f-49df-aff6-813792212999-kube-api-access-kx7mz\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.974348 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-scripts\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.974390 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-config-data\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.974421 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vp6j\" (UniqueName: \"kubernetes.io/projected/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-kube-api-access-5vp6j\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.982241 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.982395 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.982500 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.992368 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.992414 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9dd705f6-0d8f-49df-aff6-813792212999-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.992436 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd705f6-0d8f-49df-aff6-813792212999-logs\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.992511 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-config-data\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.992560 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-log-httpd\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.992594 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-run-httpd\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.992619 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-scripts\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.992705 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:41 crc kubenswrapper[4810]: I1201 14:56:41.992797 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d18c05bc-3e20-4305-b8bf-d45e6ff46308-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.104039 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-scripts\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.104115 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-config-data\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.104145 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vp6j\" (UniqueName: \"kubernetes.io/projected/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-kube-api-access-5vp6j\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.104232 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.104279 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.104318 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.104347 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.104370 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9dd705f6-0d8f-49df-aff6-813792212999-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.104394 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd705f6-0d8f-49df-aff6-813792212999-logs\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.104459 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-config-data\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.104520 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-log-httpd\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.104562 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-run-httpd\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.104595 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-scripts\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.104697 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.104761 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx7mz\" (UniqueName: \"kubernetes.io/projected/9dd705f6-0d8f-49df-aff6-813792212999-kube-api-access-kx7mz\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.107292 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd705f6-0d8f-49df-aff6-813792212999-logs\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.107306 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-run-httpd\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.110383 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-log-httpd\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.111815 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.112408 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9dd705f6-0d8f-49df-aff6-813792212999-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.112944 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.113357 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.115707 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.115999 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-config-data\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.117669 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-config-data\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.120371 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c0fd-account-create-update-w9l6n"] Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.129636 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-scripts\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.131222 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx7mz\" (UniqueName: \"kubernetes.io/projected/9dd705f6-0d8f-49df-aff6-813792212999-kube-api-access-kx7mz\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.133036 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.133774 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-c258c"] Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.134875 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-scripts\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.144634 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-r8tkk"] Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.155849 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vp6j\" (UniqueName: \"kubernetes.io/projected/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-kube-api-access-5vp6j\") pod \"ceilometer-0\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " pod="openstack/ceilometer-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.163237 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-96e4-account-create-update-jgpfh"] Dec 01 14:56:42 crc kubenswrapper[4810]: W1201 14:56:42.167629 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20b33096_654a_4684_971a_bd81b90d0a79.slice/crio-dfa60d51236878f99097606ea65a6ea1007324216150cf61ae44696d70feb950 WatchSource:0}: Error finding container dfa60d51236878f99097606ea65a6ea1007324216150cf61ae44696d70feb950: Status 404 returned error can't find the container with id dfa60d51236878f99097606ea65a6ea1007324216150cf61ae44696d70feb950 Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.173853 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-2fflr"] Dec 01 14:56:42 crc kubenswrapper[4810]: W1201 14:56:42.179149 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e760d14_b5f0_48c8_927e_09d086591dce.slice/crio-e8bc1f98786fa7b6eea581856d94f9bede1a45c89a52597d328faecbd91c7430 WatchSource:0}: Error finding container e8bc1f98786fa7b6eea581856d94f9bede1a45c89a52597d328faecbd91c7430: Status 404 returned error can't find the container with id e8bc1f98786fa7b6eea581856d94f9bede1a45c89a52597d328faecbd91c7430 Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.184546 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " pod="openstack/glance-default-external-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.317004 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.330338 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.426053 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-96e4-account-create-update-jgpfh" event={"ID":"1e760d14-b5f0-48c8-927e-09d086591dce","Type":"ContainerStarted","Data":"e8bc1f98786fa7b6eea581856d94f9bede1a45c89a52597d328faecbd91c7430"} Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.427633 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-c683-account-create-update-m9bvk" event={"ID":"f6346394-0edb-4b7b-85fd-47907ea08f08","Type":"ContainerStarted","Data":"a6874557a40d912608216311cf27091cc14caad2e17413b248e220dba30dee7e"} Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.427657 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-c683-account-create-update-m9bvk" event={"ID":"f6346394-0edb-4b7b-85fd-47907ea08f08","Type":"ContainerStarted","Data":"e935ce64bb2ff4a6be16306a428aa996337330e5db1d25a49dc20c3cfa431d2e"} Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.436247 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d18c05bc-3e20-4305-b8bf-d45e6ff46308","Type":"ContainerDied","Data":"91fb5799b749f63d23a20589068152c4f2153471a0828e73df67eaee15e98a21"} Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.436305 4810 scope.go:117] "RemoveContainer" containerID="11cb52869bfafc31bb1d1d214f6a4b58a7e39a377578ec37c708f280ae68b49f" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.436256 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.438541 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-r8tkk" event={"ID":"20b33096-654a-4684-971a-bd81b90d0a79","Type":"ContainerStarted","Data":"dfa60d51236878f99097606ea65a6ea1007324216150cf61ae44696d70feb950"} Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.445905 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-c683-account-create-update-m9bvk" podStartSLOduration=6.445866387 podStartE2EDuration="6.445866387s" podCreationTimestamp="2025-12-01 14:56:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:56:42.444739636 +0000 UTC m=+1368.208249249" watchObservedRunningTime="2025-12-01 14:56:42.445866387 +0000 UTC m=+1368.209375990" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.455994 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c0fd-account-create-update-w9l6n" event={"ID":"75872443-9a8a-4b7b-9c70-036dc558594e","Type":"ContainerStarted","Data":"9df09c9053d4dfb917add09ca666c3fcf230de818eac274927a36b5b8de0d7f9"} Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.474208 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-c258c" event={"ID":"fc8eec67-10ea-427f-b6f7-045c1dc7519f","Type":"ContainerStarted","Data":"967ee5ecc888c5fddf98652b2bfce4e0294725b73ee312e79f622b066b46955d"} Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.476841 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2fflr" event={"ID":"916e9520-9e98-42a8-bbb5-cb5f4029daab","Type":"ContainerStarted","Data":"d649a5cbd45a9592477004ddf2db73a35f3ea3c46040ea327c1489c49954ab45"} Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.536281 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a454ac3-a4dc-41c9-b548-3b57cad82bc5" path="/var/lib/kubelet/pods/1a454ac3-a4dc-41c9-b548-3b57cad82bc5/volumes" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.542836 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3039d59e-4377-411d-98e7-9fd84a2d93b6" path="/var/lib/kubelet/pods/3039d59e-4377-411d-98e7-9fd84a2d93b6/volumes" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.607434 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.616216 4810 scope.go:117] "RemoveContainer" containerID="373d8ee5e28ab65af88de3532b2fd0604c91e4cbc4b1d3f1e4e86653bfc62288" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.654719 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.677147 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.678653 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.681986 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.682110 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.715588 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.749422 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-logs\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.749486 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.749512 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.749907 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.750173 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.750263 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.750356 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.750394 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42pvr\" (UniqueName: \"kubernetes.io/projected/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-kube-api-access-42pvr\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.852864 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.852969 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.853281 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.854016 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.854050 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42pvr\" (UniqueName: \"kubernetes.io/projected/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-kube-api-access-42pvr\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.854103 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-logs\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.854137 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.854168 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.854217 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.864383 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-logs\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.864521 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.867917 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.877384 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.877957 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.883725 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.883994 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42pvr\" (UniqueName: \"kubernetes.io/projected/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-kube-api-access-42pvr\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:42 crc kubenswrapper[4810]: I1201 14:56:42.914350 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-internal-api-0\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " pod="openstack/glance-default-internal-api-0" Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.015171 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.036838 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.343004 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 14:56:43 crc kubenswrapper[4810]: W1201 14:56:43.361360 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9dd705f6_0d8f_49df_aff6_813792212999.slice/crio-16375e9d336a56f3aaf7a3394fb2b605a0b1e82cbab8e3ac3b7db8ab2f0f5f80 WatchSource:0}: Error finding container 16375e9d336a56f3aaf7a3394fb2b605a0b1e82cbab8e3ac3b7db8ab2f0f5f80: Status 404 returned error can't find the container with id 16375e9d336a56f3aaf7a3394fb2b605a0b1e82cbab8e3ac3b7db8ab2f0f5f80 Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.494064 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9dd705f6-0d8f-49df-aff6-813792212999","Type":"ContainerStarted","Data":"16375e9d336a56f3aaf7a3394fb2b605a0b1e82cbab8e3ac3b7db8ab2f0f5f80"} Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.497425 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724","Type":"ContainerStarted","Data":"b5799bfe4058fb7f66b50a58eb285d6f89880a8b4cfe1647c268f7505a2fdce0"} Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.499742 4810 generic.go:334] "Generic (PLEG): container finished" podID="fc8eec67-10ea-427f-b6f7-045c1dc7519f" containerID="06071e1b921f7719bea9f6badcd98338a4b9dc1695fc5c83f76c3f50c3c6483d" exitCode=0 Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.499875 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-c258c" event={"ID":"fc8eec67-10ea-427f-b6f7-045c1dc7519f","Type":"ContainerDied","Data":"06071e1b921f7719bea9f6badcd98338a4b9dc1695fc5c83f76c3f50c3c6483d"} Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.507798 4810 generic.go:334] "Generic (PLEG): container finished" podID="20b33096-654a-4684-971a-bd81b90d0a79" containerID="0d35ce8009f60e73fe688ef64e665403a31395b2f3b0dca76964cc0ed35f017b" exitCode=0 Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.507869 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-r8tkk" event={"ID":"20b33096-654a-4684-971a-bd81b90d0a79","Type":"ContainerDied","Data":"0d35ce8009f60e73fe688ef64e665403a31395b2f3b0dca76964cc0ed35f017b"} Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.514147 4810 generic.go:334] "Generic (PLEG): container finished" podID="1e760d14-b5f0-48c8-927e-09d086591dce" containerID="63159094debcf674a5071ab3215f742252e19a2d2a7c8655f727042b942a3233" exitCode=0 Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.514212 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-96e4-account-create-update-jgpfh" event={"ID":"1e760d14-b5f0-48c8-927e-09d086591dce","Type":"ContainerDied","Data":"63159094debcf674a5071ab3215f742252e19a2d2a7c8655f727042b942a3233"} Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.515705 4810 generic.go:334] "Generic (PLEG): container finished" podID="75872443-9a8a-4b7b-9c70-036dc558594e" containerID="a093fc007de4e4dc8464d53799a70b12b4165e5b97136fd03d12179dfe40411d" exitCode=0 Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.515750 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c0fd-account-create-update-w9l6n" event={"ID":"75872443-9a8a-4b7b-9c70-036dc558594e","Type":"ContainerDied","Data":"a093fc007de4e4dc8464d53799a70b12b4165e5b97136fd03d12179dfe40411d"} Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.522109 4810 generic.go:334] "Generic (PLEG): container finished" podID="f6346394-0edb-4b7b-85fd-47907ea08f08" containerID="a6874557a40d912608216311cf27091cc14caad2e17413b248e220dba30dee7e" exitCode=0 Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.522172 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-c683-account-create-update-m9bvk" event={"ID":"f6346394-0edb-4b7b-85fd-47907ea08f08","Type":"ContainerDied","Data":"a6874557a40d912608216311cf27091cc14caad2e17413b248e220dba30dee7e"} Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.555083 4810 generic.go:334] "Generic (PLEG): container finished" podID="916e9520-9e98-42a8-bbb5-cb5f4029daab" containerID="4fa5bae08b56685a2931d6610c251e1527a50f7f4d4f0e483abd703d7161e032" exitCode=0 Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.557193 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2fflr" event={"ID":"916e9520-9e98-42a8-bbb5-cb5f4029daab","Type":"ContainerDied","Data":"4fa5bae08b56685a2931d6610c251e1527a50f7f4d4f0e483abd703d7161e032"} Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.717088 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 14:56:43 crc kubenswrapper[4810]: I1201 14:56:43.774406 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:56:44 crc kubenswrapper[4810]: I1201 14:56:44.512379 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d18c05bc-3e20-4305-b8bf-d45e6ff46308" path="/var/lib/kubelet/pods/d18c05bc-3e20-4305-b8bf-d45e6ff46308/volumes" Dec 01 14:56:44 crc kubenswrapper[4810]: I1201 14:56:44.602887 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc","Type":"ContainerStarted","Data":"630e5f798fca1eedf921c4930a1d5c0f33b14ea41b717a6558b2a6c06e4903f1"} Dec 01 14:56:44 crc kubenswrapper[4810]: I1201 14:56:44.602938 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc","Type":"ContainerStarted","Data":"8b54ceba4a7e600832243250c7d73799a28d24f2dca3d75e7a1f10997db13fb1"} Dec 01 14:56:44 crc kubenswrapper[4810]: I1201 14:56:44.606713 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9dd705f6-0d8f-49df-aff6-813792212999","Type":"ContainerStarted","Data":"1077d64ab29e8e4a5e08c7d81078a1254c035e9de419585bf1944ac1e43a909c"} Dec 01 14:56:44 crc kubenswrapper[4810]: I1201 14:56:44.615612 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724","Type":"ContainerStarted","Data":"a7da4879c3e131a643a2e4f7ef5f6bab666765fd329c8aea28b3e988ca8f0a33"} Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.150152 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-r8tkk" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.218523 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b33096-654a-4684-971a-bd81b90d0a79-operator-scripts\") pod \"20b33096-654a-4684-971a-bd81b90d0a79\" (UID: \"20b33096-654a-4684-971a-bd81b90d0a79\") " Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.218608 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckbtm\" (UniqueName: \"kubernetes.io/projected/20b33096-654a-4684-971a-bd81b90d0a79-kube-api-access-ckbtm\") pod \"20b33096-654a-4684-971a-bd81b90d0a79\" (UID: \"20b33096-654a-4684-971a-bd81b90d0a79\") " Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.220128 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20b33096-654a-4684-971a-bd81b90d0a79-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "20b33096-654a-4684-971a-bd81b90d0a79" (UID: "20b33096-654a-4684-971a-bd81b90d0a79"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.233691 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b33096-654a-4684-971a-bd81b90d0a79-kube-api-access-ckbtm" (OuterVolumeSpecName: "kube-api-access-ckbtm") pod "20b33096-654a-4684-971a-bd81b90d0a79" (UID: "20b33096-654a-4684-971a-bd81b90d0a79"). InnerVolumeSpecName "kube-api-access-ckbtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.295566 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c0fd-account-create-update-w9l6n" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.318520 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2fflr" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.320222 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20b33096-654a-4684-971a-bd81b90d0a79-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.320248 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckbtm\" (UniqueName: \"kubernetes.io/projected/20b33096-654a-4684-971a-bd81b90d0a79-kube-api-access-ckbtm\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.327005 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-96e4-account-create-update-jgpfh" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.342934 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c683-account-create-update-m9bvk" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.343930 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-c258c" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.420981 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cglvd\" (UniqueName: \"kubernetes.io/projected/1e760d14-b5f0-48c8-927e-09d086591dce-kube-api-access-cglvd\") pod \"1e760d14-b5f0-48c8-927e-09d086591dce\" (UID: \"1e760d14-b5f0-48c8-927e-09d086591dce\") " Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.421053 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75872443-9a8a-4b7b-9c70-036dc558594e-operator-scripts\") pod \"75872443-9a8a-4b7b-9c70-036dc558594e\" (UID: \"75872443-9a8a-4b7b-9c70-036dc558594e\") " Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.421097 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btcdt\" (UniqueName: \"kubernetes.io/projected/fc8eec67-10ea-427f-b6f7-045c1dc7519f-kube-api-access-btcdt\") pod \"fc8eec67-10ea-427f-b6f7-045c1dc7519f\" (UID: \"fc8eec67-10ea-427f-b6f7-045c1dc7519f\") " Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.421133 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6346394-0edb-4b7b-85fd-47907ea08f08-operator-scripts\") pod \"f6346394-0edb-4b7b-85fd-47907ea08f08\" (UID: \"f6346394-0edb-4b7b-85fd-47907ea08f08\") " Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.421163 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtjpw\" (UniqueName: \"kubernetes.io/projected/75872443-9a8a-4b7b-9c70-036dc558594e-kube-api-access-qtjpw\") pod \"75872443-9a8a-4b7b-9c70-036dc558594e\" (UID: \"75872443-9a8a-4b7b-9c70-036dc558594e\") " Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.421188 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26vh7\" (UniqueName: \"kubernetes.io/projected/916e9520-9e98-42a8-bbb5-cb5f4029daab-kube-api-access-26vh7\") pod \"916e9520-9e98-42a8-bbb5-cb5f4029daab\" (UID: \"916e9520-9e98-42a8-bbb5-cb5f4029daab\") " Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.421219 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e760d14-b5f0-48c8-927e-09d086591dce-operator-scripts\") pod \"1e760d14-b5f0-48c8-927e-09d086591dce\" (UID: \"1e760d14-b5f0-48c8-927e-09d086591dce\") " Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.421272 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc8eec67-10ea-427f-b6f7-045c1dc7519f-operator-scripts\") pod \"fc8eec67-10ea-427f-b6f7-045c1dc7519f\" (UID: \"fc8eec67-10ea-427f-b6f7-045c1dc7519f\") " Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.421319 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvb7h\" (UniqueName: \"kubernetes.io/projected/f6346394-0edb-4b7b-85fd-47907ea08f08-kube-api-access-hvb7h\") pod \"f6346394-0edb-4b7b-85fd-47907ea08f08\" (UID: \"f6346394-0edb-4b7b-85fd-47907ea08f08\") " Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.421373 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/916e9520-9e98-42a8-bbb5-cb5f4029daab-operator-scripts\") pod \"916e9520-9e98-42a8-bbb5-cb5f4029daab\" (UID: \"916e9520-9e98-42a8-bbb5-cb5f4029daab\") " Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.422188 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/916e9520-9e98-42a8-bbb5-cb5f4029daab-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "916e9520-9e98-42a8-bbb5-cb5f4029daab" (UID: "916e9520-9e98-42a8-bbb5-cb5f4029daab"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.423884 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e760d14-b5f0-48c8-927e-09d086591dce-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1e760d14-b5f0-48c8-927e-09d086591dce" (UID: "1e760d14-b5f0-48c8-927e-09d086591dce"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.424951 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc8eec67-10ea-427f-b6f7-045c1dc7519f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fc8eec67-10ea-427f-b6f7-045c1dc7519f" (UID: "fc8eec67-10ea-427f-b6f7-045c1dc7519f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.428067 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75872443-9a8a-4b7b-9c70-036dc558594e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "75872443-9a8a-4b7b-9c70-036dc558594e" (UID: "75872443-9a8a-4b7b-9c70-036dc558594e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.430672 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6346394-0edb-4b7b-85fd-47907ea08f08-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f6346394-0edb-4b7b-85fd-47907ea08f08" (UID: "f6346394-0edb-4b7b-85fd-47907ea08f08"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.432752 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75872443-9a8a-4b7b-9c70-036dc558594e-kube-api-access-qtjpw" (OuterVolumeSpecName: "kube-api-access-qtjpw") pod "75872443-9a8a-4b7b-9c70-036dc558594e" (UID: "75872443-9a8a-4b7b-9c70-036dc558594e"). InnerVolumeSpecName "kube-api-access-qtjpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.433591 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/916e9520-9e98-42a8-bbb5-cb5f4029daab-kube-api-access-26vh7" (OuterVolumeSpecName: "kube-api-access-26vh7") pod "916e9520-9e98-42a8-bbb5-cb5f4029daab" (UID: "916e9520-9e98-42a8-bbb5-cb5f4029daab"). InnerVolumeSpecName "kube-api-access-26vh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.434667 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc8eec67-10ea-427f-b6f7-045c1dc7519f-kube-api-access-btcdt" (OuterVolumeSpecName: "kube-api-access-btcdt") pod "fc8eec67-10ea-427f-b6f7-045c1dc7519f" (UID: "fc8eec67-10ea-427f-b6f7-045c1dc7519f"). InnerVolumeSpecName "kube-api-access-btcdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.436820 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6346394-0edb-4b7b-85fd-47907ea08f08-kube-api-access-hvb7h" (OuterVolumeSpecName: "kube-api-access-hvb7h") pod "f6346394-0edb-4b7b-85fd-47907ea08f08" (UID: "f6346394-0edb-4b7b-85fd-47907ea08f08"). InnerVolumeSpecName "kube-api-access-hvb7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.440429 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e760d14-b5f0-48c8-927e-09d086591dce-kube-api-access-cglvd" (OuterVolumeSpecName: "kube-api-access-cglvd") pod "1e760d14-b5f0-48c8-927e-09d086591dce" (UID: "1e760d14-b5f0-48c8-927e-09d086591dce"). InnerVolumeSpecName "kube-api-access-cglvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.523886 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtjpw\" (UniqueName: \"kubernetes.io/projected/75872443-9a8a-4b7b-9c70-036dc558594e-kube-api-access-qtjpw\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.523920 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26vh7\" (UniqueName: \"kubernetes.io/projected/916e9520-9e98-42a8-bbb5-cb5f4029daab-kube-api-access-26vh7\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.523930 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e760d14-b5f0-48c8-927e-09d086591dce-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.523938 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fc8eec67-10ea-427f-b6f7-045c1dc7519f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.523947 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvb7h\" (UniqueName: \"kubernetes.io/projected/f6346394-0edb-4b7b-85fd-47907ea08f08-kube-api-access-hvb7h\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.523955 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/916e9520-9e98-42a8-bbb5-cb5f4029daab-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.523963 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cglvd\" (UniqueName: \"kubernetes.io/projected/1e760d14-b5f0-48c8-927e-09d086591dce-kube-api-access-cglvd\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.523971 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75872443-9a8a-4b7b-9c70-036dc558594e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.523980 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btcdt\" (UniqueName: \"kubernetes.io/projected/fc8eec67-10ea-427f-b6f7-045c1dc7519f-kube-api-access-btcdt\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.523988 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6346394-0edb-4b7b-85fd-47907ea08f08-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.625624 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c683-account-create-update-m9bvk" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.625664 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-c683-account-create-update-m9bvk" event={"ID":"f6346394-0edb-4b7b-85fd-47907ea08f08","Type":"ContainerDied","Data":"e935ce64bb2ff4a6be16306a428aa996337330e5db1d25a49dc20c3cfa431d2e"} Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.625702 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e935ce64bb2ff4a6be16306a428aa996337330e5db1d25a49dc20c3cfa431d2e" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.627432 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-c258c" event={"ID":"fc8eec67-10ea-427f-b6f7-045c1dc7519f","Type":"ContainerDied","Data":"967ee5ecc888c5fddf98652b2bfce4e0294725b73ee312e79f622b066b46955d"} Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.627451 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-c258c" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.627488 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="967ee5ecc888c5fddf98652b2bfce4e0294725b73ee312e79f622b066b46955d" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.629156 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-r8tkk" event={"ID":"20b33096-654a-4684-971a-bd81b90d0a79","Type":"ContainerDied","Data":"dfa60d51236878f99097606ea65a6ea1007324216150cf61ae44696d70feb950"} Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.629261 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfa60d51236878f99097606ea65a6ea1007324216150cf61ae44696d70feb950" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.629175 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-r8tkk" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.631233 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc","Type":"ContainerStarted","Data":"aba75285c97143bda5c120fd54711afa0a88b59eeac5b761fa88578cc1f0f3b2"} Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.649497 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c0fd-account-create-update-w9l6n" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.649620 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c0fd-account-create-update-w9l6n" event={"ID":"75872443-9a8a-4b7b-9c70-036dc558594e","Type":"ContainerDied","Data":"9df09c9053d4dfb917add09ca666c3fcf230de818eac274927a36b5b8de0d7f9"} Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.649663 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9df09c9053d4dfb917add09ca666c3fcf230de818eac274927a36b5b8de0d7f9" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.651172 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2fflr" event={"ID":"916e9520-9e98-42a8-bbb5-cb5f4029daab","Type":"ContainerDied","Data":"d649a5cbd45a9592477004ddf2db73a35f3ea3c46040ea327c1489c49954ab45"} Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.651218 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d649a5cbd45a9592477004ddf2db73a35f3ea3c46040ea327c1489c49954ab45" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.651300 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2fflr" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.653880 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9dd705f6-0d8f-49df-aff6-813792212999","Type":"ContainerStarted","Data":"ca8c16cdb56a677709c0ac603398c85e43e5a2f9e1cb3036f753b35de804a1f8"} Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.657884 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724","Type":"ContainerStarted","Data":"0d1323d6de251253b53f35bf7bdec9ffa1cdce6ba5012e959dbda2effdf8cccd"} Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.667892 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-96e4-account-create-update-jgpfh" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.667899 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-96e4-account-create-update-jgpfh" event={"ID":"1e760d14-b5f0-48c8-927e-09d086591dce","Type":"ContainerDied","Data":"e8bc1f98786fa7b6eea581856d94f9bede1a45c89a52597d328faecbd91c7430"} Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.668260 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8bc1f98786fa7b6eea581856d94f9bede1a45c89a52597d328faecbd91c7430" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.673760 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.673735166 podStartE2EDuration="3.673735166s" podCreationTimestamp="2025-12-01 14:56:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:56:45.660925477 +0000 UTC m=+1371.424435080" watchObservedRunningTime="2025-12-01 14:56:45.673735166 +0000 UTC m=+1371.437244769" Dec 01 14:56:45 crc kubenswrapper[4810]: I1201 14:56:45.708413 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.708392046 podStartE2EDuration="4.708392046s" podCreationTimestamp="2025-12-01 14:56:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:56:45.707601445 +0000 UTC m=+1371.471111068" watchObservedRunningTime="2025-12-01 14:56:45.708392046 +0000 UTC m=+1371.471901649" Dec 01 14:56:46 crc kubenswrapper[4810]: I1201 14:56:46.676757 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724","Type":"ContainerStarted","Data":"a67320d759667df001e48575d2383e891be36019eb6ca2c8c64fb1442e2271c9"} Dec 01 14:56:48 crc kubenswrapper[4810]: I1201 14:56:48.701607 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724","Type":"ContainerStarted","Data":"a97dee02d446ebf05a7be83e77ce57975282cb30345d0ea6b22b0bac66e86288"} Dec 01 14:56:48 crc kubenswrapper[4810]: I1201 14:56:48.702128 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 14:56:48 crc kubenswrapper[4810]: I1201 14:56:48.701785 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerName="ceilometer-central-agent" containerID="cri-o://a7da4879c3e131a643a2e4f7ef5f6bab666765fd329c8aea28b3e988ca8f0a33" gracePeriod=30 Dec 01 14:56:48 crc kubenswrapper[4810]: I1201 14:56:48.702223 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerName="proxy-httpd" containerID="cri-o://a97dee02d446ebf05a7be83e77ce57975282cb30345d0ea6b22b0bac66e86288" gracePeriod=30 Dec 01 14:56:48 crc kubenswrapper[4810]: I1201 14:56:48.702289 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerName="ceilometer-notification-agent" containerID="cri-o://0d1323d6de251253b53f35bf7bdec9ffa1cdce6ba5012e959dbda2effdf8cccd" gracePeriod=30 Dec 01 14:56:48 crc kubenswrapper[4810]: I1201 14:56:48.702325 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerName="sg-core" containerID="cri-o://a67320d759667df001e48575d2383e891be36019eb6ca2c8c64fb1442e2271c9" gracePeriod=30 Dec 01 14:56:48 crc kubenswrapper[4810]: I1201 14:56:48.733666 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.142347889 podStartE2EDuration="7.733644561s" podCreationTimestamp="2025-12-01 14:56:41 +0000 UTC" firstStartedPulling="2025-12-01 14:56:43.157613601 +0000 UTC m=+1368.921123204" lastFinishedPulling="2025-12-01 14:56:47.748910273 +0000 UTC m=+1373.512419876" observedRunningTime="2025-12-01 14:56:48.728005428 +0000 UTC m=+1374.491515031" watchObservedRunningTime="2025-12-01 14:56:48.733644561 +0000 UTC m=+1374.497154164" Dec 01 14:56:49 crc kubenswrapper[4810]: I1201 14:56:49.712833 4810 generic.go:334] "Generic (PLEG): container finished" podID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerID="a97dee02d446ebf05a7be83e77ce57975282cb30345d0ea6b22b0bac66e86288" exitCode=0 Dec 01 14:56:49 crc kubenswrapper[4810]: I1201 14:56:49.713611 4810 generic.go:334] "Generic (PLEG): container finished" podID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerID="a67320d759667df001e48575d2383e891be36019eb6ca2c8c64fb1442e2271c9" exitCode=2 Dec 01 14:56:49 crc kubenswrapper[4810]: I1201 14:56:49.713676 4810 generic.go:334] "Generic (PLEG): container finished" podID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerID="0d1323d6de251253b53f35bf7bdec9ffa1cdce6ba5012e959dbda2effdf8cccd" exitCode=0 Dec 01 14:56:49 crc kubenswrapper[4810]: I1201 14:56:49.712873 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724","Type":"ContainerDied","Data":"a97dee02d446ebf05a7be83e77ce57975282cb30345d0ea6b22b0bac66e86288"} Dec 01 14:56:49 crc kubenswrapper[4810]: I1201 14:56:49.713834 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724","Type":"ContainerDied","Data":"a67320d759667df001e48575d2383e891be36019eb6ca2c8c64fb1442e2271c9"} Dec 01 14:56:49 crc kubenswrapper[4810]: I1201 14:56:49.713895 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724","Type":"ContainerDied","Data":"0d1323d6de251253b53f35bf7bdec9ffa1cdce6ba5012e959dbda2effdf8cccd"} Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.676922 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.725406 4810 generic.go:334] "Generic (PLEG): container finished" podID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerID="a7da4879c3e131a643a2e4f7ef5f6bab666765fd329c8aea28b3e988ca8f0a33" exitCode=0 Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.725453 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724","Type":"ContainerDied","Data":"a7da4879c3e131a643a2e4f7ef5f6bab666765fd329c8aea28b3e988ca8f0a33"} Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.725502 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724","Type":"ContainerDied","Data":"b5799bfe4058fb7f66b50a58eb285d6f89880a8b4cfe1647c268f7505a2fdce0"} Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.725523 4810 scope.go:117] "RemoveContainer" containerID="a97dee02d446ebf05a7be83e77ce57975282cb30345d0ea6b22b0bac66e86288" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.725614 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.749438 4810 scope.go:117] "RemoveContainer" containerID="a67320d759667df001e48575d2383e891be36019eb6ca2c8c64fb1442e2271c9" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.770029 4810 scope.go:117] "RemoveContainer" containerID="0d1323d6de251253b53f35bf7bdec9ffa1cdce6ba5012e959dbda2effdf8cccd" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.794174 4810 scope.go:117] "RemoveContainer" containerID="a7da4879c3e131a643a2e4f7ef5f6bab666765fd329c8aea28b3e988ca8f0a33" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.813209 4810 scope.go:117] "RemoveContainer" containerID="a97dee02d446ebf05a7be83e77ce57975282cb30345d0ea6b22b0bac66e86288" Dec 01 14:56:50 crc kubenswrapper[4810]: E1201 14:56:50.813797 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a97dee02d446ebf05a7be83e77ce57975282cb30345d0ea6b22b0bac66e86288\": container with ID starting with a97dee02d446ebf05a7be83e77ce57975282cb30345d0ea6b22b0bac66e86288 not found: ID does not exist" containerID="a97dee02d446ebf05a7be83e77ce57975282cb30345d0ea6b22b0bac66e86288" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.813837 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a97dee02d446ebf05a7be83e77ce57975282cb30345d0ea6b22b0bac66e86288"} err="failed to get container status \"a97dee02d446ebf05a7be83e77ce57975282cb30345d0ea6b22b0bac66e86288\": rpc error: code = NotFound desc = could not find container \"a97dee02d446ebf05a7be83e77ce57975282cb30345d0ea6b22b0bac66e86288\": container with ID starting with a97dee02d446ebf05a7be83e77ce57975282cb30345d0ea6b22b0bac66e86288 not found: ID does not exist" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.813861 4810 scope.go:117] "RemoveContainer" containerID="a67320d759667df001e48575d2383e891be36019eb6ca2c8c64fb1442e2271c9" Dec 01 14:56:50 crc kubenswrapper[4810]: E1201 14:56:50.814181 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a67320d759667df001e48575d2383e891be36019eb6ca2c8c64fb1442e2271c9\": container with ID starting with a67320d759667df001e48575d2383e891be36019eb6ca2c8c64fb1442e2271c9 not found: ID does not exist" containerID="a67320d759667df001e48575d2383e891be36019eb6ca2c8c64fb1442e2271c9" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.814208 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a67320d759667df001e48575d2383e891be36019eb6ca2c8c64fb1442e2271c9"} err="failed to get container status \"a67320d759667df001e48575d2383e891be36019eb6ca2c8c64fb1442e2271c9\": rpc error: code = NotFound desc = could not find container \"a67320d759667df001e48575d2383e891be36019eb6ca2c8c64fb1442e2271c9\": container with ID starting with a67320d759667df001e48575d2383e891be36019eb6ca2c8c64fb1442e2271c9 not found: ID does not exist" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.814226 4810 scope.go:117] "RemoveContainer" containerID="0d1323d6de251253b53f35bf7bdec9ffa1cdce6ba5012e959dbda2effdf8cccd" Dec 01 14:56:50 crc kubenswrapper[4810]: E1201 14:56:50.814688 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d1323d6de251253b53f35bf7bdec9ffa1cdce6ba5012e959dbda2effdf8cccd\": container with ID starting with 0d1323d6de251253b53f35bf7bdec9ffa1cdce6ba5012e959dbda2effdf8cccd not found: ID does not exist" containerID="0d1323d6de251253b53f35bf7bdec9ffa1cdce6ba5012e959dbda2effdf8cccd" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.814706 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d1323d6de251253b53f35bf7bdec9ffa1cdce6ba5012e959dbda2effdf8cccd"} err="failed to get container status \"0d1323d6de251253b53f35bf7bdec9ffa1cdce6ba5012e959dbda2effdf8cccd\": rpc error: code = NotFound desc = could not find container \"0d1323d6de251253b53f35bf7bdec9ffa1cdce6ba5012e959dbda2effdf8cccd\": container with ID starting with 0d1323d6de251253b53f35bf7bdec9ffa1cdce6ba5012e959dbda2effdf8cccd not found: ID does not exist" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.814720 4810 scope.go:117] "RemoveContainer" containerID="a7da4879c3e131a643a2e4f7ef5f6bab666765fd329c8aea28b3e988ca8f0a33" Dec 01 14:56:50 crc kubenswrapper[4810]: E1201 14:56:50.814919 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7da4879c3e131a643a2e4f7ef5f6bab666765fd329c8aea28b3e988ca8f0a33\": container with ID starting with a7da4879c3e131a643a2e4f7ef5f6bab666765fd329c8aea28b3e988ca8f0a33 not found: ID does not exist" containerID="a7da4879c3e131a643a2e4f7ef5f6bab666765fd329c8aea28b3e988ca8f0a33" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.814936 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7da4879c3e131a643a2e4f7ef5f6bab666765fd329c8aea28b3e988ca8f0a33"} err="failed to get container status \"a7da4879c3e131a643a2e4f7ef5f6bab666765fd329c8aea28b3e988ca8f0a33\": rpc error: code = NotFound desc = could not find container \"a7da4879c3e131a643a2e4f7ef5f6bab666765fd329c8aea28b3e988ca8f0a33\": container with ID starting with a7da4879c3e131a643a2e4f7ef5f6bab666765fd329c8aea28b3e988ca8f0a33 not found: ID does not exist" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.826384 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-log-httpd\") pod \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.826587 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-combined-ca-bundle\") pod \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.826648 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-sg-core-conf-yaml\") pod \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.826684 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vp6j\" (UniqueName: \"kubernetes.io/projected/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-kube-api-access-5vp6j\") pod \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.826736 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-config-data\") pod \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.826766 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-scripts\") pod \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.826801 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-run-httpd\") pod \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\" (UID: \"fe7b36f4-9b50-4c3b-b11b-34cb6b5da724\") " Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.827756 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" (UID: "fe7b36f4-9b50-4c3b-b11b-34cb6b5da724"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.827878 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" (UID: "fe7b36f4-9b50-4c3b-b11b-34cb6b5da724"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.832565 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-kube-api-access-5vp6j" (OuterVolumeSpecName: "kube-api-access-5vp6j") pod "fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" (UID: "fe7b36f4-9b50-4c3b-b11b-34cb6b5da724"). InnerVolumeSpecName "kube-api-access-5vp6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.833426 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-scripts" (OuterVolumeSpecName: "scripts") pod "fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" (UID: "fe7b36f4-9b50-4c3b-b11b-34cb6b5da724"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.860191 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" (UID: "fe7b36f4-9b50-4c3b-b11b-34cb6b5da724"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.913449 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" (UID: "fe7b36f4-9b50-4c3b-b11b-34cb6b5da724"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.928680 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vp6j\" (UniqueName: \"kubernetes.io/projected/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-kube-api-access-5vp6j\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.928721 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.928732 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.928742 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.928753 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.928765 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:50 crc kubenswrapper[4810]: I1201 14:56:50.935573 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-config-data" (OuterVolumeSpecName: "config-data") pod "fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" (UID: "fe7b36f4-9b50-4c3b-b11b-34cb6b5da724"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.030653 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.058741 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.072101 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.092017 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:56:51 crc kubenswrapper[4810]: E1201 14:56:51.092485 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc8eec67-10ea-427f-b6f7-045c1dc7519f" containerName="mariadb-database-create" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.092511 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc8eec67-10ea-427f-b6f7-045c1dc7519f" containerName="mariadb-database-create" Dec 01 14:56:51 crc kubenswrapper[4810]: E1201 14:56:51.092527 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerName="sg-core" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.092537 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerName="sg-core" Dec 01 14:56:51 crc kubenswrapper[4810]: E1201 14:56:51.092555 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerName="proxy-httpd" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.092564 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerName="proxy-httpd" Dec 01 14:56:51 crc kubenswrapper[4810]: E1201 14:56:51.092577 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerName="ceilometer-central-agent" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.092585 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerName="ceilometer-central-agent" Dec 01 14:56:51 crc kubenswrapper[4810]: E1201 14:56:51.092596 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e760d14-b5f0-48c8-927e-09d086591dce" containerName="mariadb-account-create-update" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.092604 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e760d14-b5f0-48c8-927e-09d086591dce" containerName="mariadb-account-create-update" Dec 01 14:56:51 crc kubenswrapper[4810]: E1201 14:56:51.092619 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6346394-0edb-4b7b-85fd-47907ea08f08" containerName="mariadb-account-create-update" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.092628 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6346394-0edb-4b7b-85fd-47907ea08f08" containerName="mariadb-account-create-update" Dec 01 14:56:51 crc kubenswrapper[4810]: E1201 14:56:51.092644 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20b33096-654a-4684-971a-bd81b90d0a79" containerName="mariadb-database-create" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.092653 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="20b33096-654a-4684-971a-bd81b90d0a79" containerName="mariadb-database-create" Dec 01 14:56:51 crc kubenswrapper[4810]: E1201 14:56:51.092666 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="916e9520-9e98-42a8-bbb5-cb5f4029daab" containerName="mariadb-database-create" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.092675 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="916e9520-9e98-42a8-bbb5-cb5f4029daab" containerName="mariadb-database-create" Dec 01 14:56:51 crc kubenswrapper[4810]: E1201 14:56:51.092687 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerName="ceilometer-notification-agent" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.092694 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerName="ceilometer-notification-agent" Dec 01 14:56:51 crc kubenswrapper[4810]: E1201 14:56:51.092713 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75872443-9a8a-4b7b-9c70-036dc558594e" containerName="mariadb-account-create-update" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.092721 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="75872443-9a8a-4b7b-9c70-036dc558594e" containerName="mariadb-account-create-update" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.092933 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e760d14-b5f0-48c8-927e-09d086591dce" containerName="mariadb-account-create-update" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.092952 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="916e9520-9e98-42a8-bbb5-cb5f4029daab" containerName="mariadb-database-create" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.092968 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6346394-0edb-4b7b-85fd-47907ea08f08" containerName="mariadb-account-create-update" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.092981 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerName="sg-core" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.092999 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerName="ceilometer-central-agent" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.093012 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="20b33096-654a-4684-971a-bd81b90d0a79" containerName="mariadb-database-create" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.093023 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerName="proxy-httpd" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.093033 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc8eec67-10ea-427f-b6f7-045c1dc7519f" containerName="mariadb-database-create" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.093046 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" containerName="ceilometer-notification-agent" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.093065 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="75872443-9a8a-4b7b-9c70-036dc558594e" containerName="mariadb-account-create-update" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.095292 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.099127 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.099634 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.102878 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.233782 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-config-data\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.233923 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.234084 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfcr4\" (UniqueName: \"kubernetes.io/projected/aab1a1ee-52fd-4c95-a15a-23a8b5766385-kube-api-access-qfcr4\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.234166 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab1a1ee-52fd-4c95-a15a-23a8b5766385-log-httpd\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.234339 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab1a1ee-52fd-4c95-a15a-23a8b5766385-run-httpd\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.234414 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-scripts\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.234603 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.335898 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfcr4\" (UniqueName: \"kubernetes.io/projected/aab1a1ee-52fd-4c95-a15a-23a8b5766385-kube-api-access-qfcr4\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.336230 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab1a1ee-52fd-4c95-a15a-23a8b5766385-log-httpd\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.336274 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab1a1ee-52fd-4c95-a15a-23a8b5766385-run-httpd\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.336303 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-scripts\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.336354 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.336371 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-config-data\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.336390 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.337262 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab1a1ee-52fd-4c95-a15a-23a8b5766385-run-httpd\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.337823 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab1a1ee-52fd-4c95-a15a-23a8b5766385-log-httpd\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.340835 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.341977 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.346182 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-scripts\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.351625 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-config-data\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.363377 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfcr4\" (UniqueName: \"kubernetes.io/projected/aab1a1ee-52fd-4c95-a15a-23a8b5766385-kube-api-access-qfcr4\") pod \"ceilometer-0\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.411133 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.594420 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gnmm4"] Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.596206 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gnmm4" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.600011 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-68sxs" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.603487 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.603682 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.611507 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gnmm4"] Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.745199 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lplt\" (UniqueName: \"kubernetes.io/projected/21fcb95f-afd8-43e5-90b2-afa5346fcd57-kube-api-access-2lplt\") pod \"nova-cell0-conductor-db-sync-gnmm4\" (UID: \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\") " pod="openstack/nova-cell0-conductor-db-sync-gnmm4" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.745350 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21fcb95f-afd8-43e5-90b2-afa5346fcd57-config-data\") pod \"nova-cell0-conductor-db-sync-gnmm4\" (UID: \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\") " pod="openstack/nova-cell0-conductor-db-sync-gnmm4" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.745387 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21fcb95f-afd8-43e5-90b2-afa5346fcd57-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-gnmm4\" (UID: \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\") " pod="openstack/nova-cell0-conductor-db-sync-gnmm4" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.745421 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21fcb95f-afd8-43e5-90b2-afa5346fcd57-scripts\") pod \"nova-cell0-conductor-db-sync-gnmm4\" (UID: \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\") " pod="openstack/nova-cell0-conductor-db-sync-gnmm4" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.847507 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21fcb95f-afd8-43e5-90b2-afa5346fcd57-config-data\") pod \"nova-cell0-conductor-db-sync-gnmm4\" (UID: \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\") " pod="openstack/nova-cell0-conductor-db-sync-gnmm4" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.847557 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21fcb95f-afd8-43e5-90b2-afa5346fcd57-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-gnmm4\" (UID: \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\") " pod="openstack/nova-cell0-conductor-db-sync-gnmm4" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.847586 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21fcb95f-afd8-43e5-90b2-afa5346fcd57-scripts\") pod \"nova-cell0-conductor-db-sync-gnmm4\" (UID: \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\") " pod="openstack/nova-cell0-conductor-db-sync-gnmm4" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.847650 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lplt\" (UniqueName: \"kubernetes.io/projected/21fcb95f-afd8-43e5-90b2-afa5346fcd57-kube-api-access-2lplt\") pod \"nova-cell0-conductor-db-sync-gnmm4\" (UID: \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\") " pod="openstack/nova-cell0-conductor-db-sync-gnmm4" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.854111 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21fcb95f-afd8-43e5-90b2-afa5346fcd57-config-data\") pod \"nova-cell0-conductor-db-sync-gnmm4\" (UID: \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\") " pod="openstack/nova-cell0-conductor-db-sync-gnmm4" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.861592 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21fcb95f-afd8-43e5-90b2-afa5346fcd57-scripts\") pod \"nova-cell0-conductor-db-sync-gnmm4\" (UID: \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\") " pod="openstack/nova-cell0-conductor-db-sync-gnmm4" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.875121 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lplt\" (UniqueName: \"kubernetes.io/projected/21fcb95f-afd8-43e5-90b2-afa5346fcd57-kube-api-access-2lplt\") pod \"nova-cell0-conductor-db-sync-gnmm4\" (UID: \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\") " pod="openstack/nova-cell0-conductor-db-sync-gnmm4" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.878999 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21fcb95f-afd8-43e5-90b2-afa5346fcd57-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-gnmm4\" (UID: \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\") " pod="openstack/nova-cell0-conductor-db-sync-gnmm4" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.922944 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gnmm4" Dec 01 14:56:51 crc kubenswrapper[4810]: I1201 14:56:51.977538 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:56:51 crc kubenswrapper[4810]: W1201 14:56:51.983868 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaab1a1ee_52fd_4c95_a15a_23a8b5766385.slice/crio-b141be6be2c52ce8cc1c317f0d800dd5c81718b031ddc68541f12f6ae2997b5c WatchSource:0}: Error finding container b141be6be2c52ce8cc1c317f0d800dd5c81718b031ddc68541f12f6ae2997b5c: Status 404 returned error can't find the container with id b141be6be2c52ce8cc1c317f0d800dd5c81718b031ddc68541f12f6ae2997b5c Dec 01 14:56:52 crc kubenswrapper[4810]: I1201 14:56:52.338681 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 14:56:52 crc kubenswrapper[4810]: I1201 14:56:52.338730 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 14:56:52 crc kubenswrapper[4810]: I1201 14:56:52.488629 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gnmm4"] Dec 01 14:56:52 crc kubenswrapper[4810]: I1201 14:56:52.497861 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 14:56:52 crc kubenswrapper[4810]: I1201 14:56:52.545626 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe7b36f4-9b50-4c3b-b11b-34cb6b5da724" path="/var/lib/kubelet/pods/fe7b36f4-9b50-4c3b-b11b-34cb6b5da724/volumes" Dec 01 14:56:52 crc kubenswrapper[4810]: I1201 14:56:52.552569 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 14:56:52 crc kubenswrapper[4810]: I1201 14:56:52.744848 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab1a1ee-52fd-4c95-a15a-23a8b5766385","Type":"ContainerStarted","Data":"b141be6be2c52ce8cc1c317f0d800dd5c81718b031ddc68541f12f6ae2997b5c"} Dec 01 14:56:52 crc kubenswrapper[4810]: I1201 14:56:52.746639 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gnmm4" event={"ID":"21fcb95f-afd8-43e5-90b2-afa5346fcd57","Type":"ContainerStarted","Data":"863196f2d4fdac01ce0e09b6dc45e85c419b7dc200a5980e800d70f1cfdc61ac"} Dec 01 14:56:52 crc kubenswrapper[4810]: I1201 14:56:52.746950 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 14:56:52 crc kubenswrapper[4810]: I1201 14:56:52.747345 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 14:56:53 crc kubenswrapper[4810]: I1201 14:56:53.015673 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 14:56:53 crc kubenswrapper[4810]: I1201 14:56:53.015730 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 14:56:53 crc kubenswrapper[4810]: I1201 14:56:53.049076 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 14:56:53 crc kubenswrapper[4810]: I1201 14:56:53.059755 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 14:56:53 crc kubenswrapper[4810]: I1201 14:56:53.798199 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab1a1ee-52fd-4c95-a15a-23a8b5766385","Type":"ContainerStarted","Data":"4411a834043bfbd4c6eebbfb6f9aefd3cccebe1398e3665943b43e9bfd8949b6"} Dec 01 14:56:53 crc kubenswrapper[4810]: I1201 14:56:53.799485 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 14:56:53 crc kubenswrapper[4810]: I1201 14:56:53.799507 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 14:56:54 crc kubenswrapper[4810]: I1201 14:56:54.803810 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 14:56:54 crc kubenswrapper[4810]: I1201 14:56:54.815585 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 14:56:54 crc kubenswrapper[4810]: I1201 14:56:54.815795 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab1a1ee-52fd-4c95-a15a-23a8b5766385","Type":"ContainerStarted","Data":"cd9ed3a7eea9c764a163d1a66f52a918a08ff86b5d992b9a8c69e03959355c2a"} Dec 01 14:56:55 crc kubenswrapper[4810]: I1201 14:56:55.345682 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:56:55 crc kubenswrapper[4810]: I1201 14:56:55.828034 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 14:56:55 crc kubenswrapper[4810]: I1201 14:56:55.828064 4810 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 14:56:55 crc kubenswrapper[4810]: I1201 14:56:55.829064 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab1a1ee-52fd-4c95-a15a-23a8b5766385","Type":"ContainerStarted","Data":"c393f75663b47fca078d96f2d48cea2b4c160dec49669eb8abe649f5407d5e69"} Dec 01 14:56:56 crc kubenswrapper[4810]: I1201 14:56:56.332425 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 14:56:56 crc kubenswrapper[4810]: I1201 14:56:56.336501 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 14:56:56 crc kubenswrapper[4810]: I1201 14:56:56.842531 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab1a1ee-52fd-4c95-a15a-23a8b5766385","Type":"ContainerStarted","Data":"eb41d25cc4e3ac12bc042cf93e3dba592f6e438c3585a09bcf8d0651ce2c591d"} Dec 01 14:56:56 crc kubenswrapper[4810]: I1201 14:56:56.842824 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerName="ceilometer-central-agent" containerID="cri-o://4411a834043bfbd4c6eebbfb6f9aefd3cccebe1398e3665943b43e9bfd8949b6" gracePeriod=30 Dec 01 14:56:56 crc kubenswrapper[4810]: I1201 14:56:56.842972 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerName="ceilometer-notification-agent" containerID="cri-o://cd9ed3a7eea9c764a163d1a66f52a918a08ff86b5d992b9a8c69e03959355c2a" gracePeriod=30 Dec 01 14:56:56 crc kubenswrapper[4810]: I1201 14:56:56.842936 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerName="proxy-httpd" containerID="cri-o://eb41d25cc4e3ac12bc042cf93e3dba592f6e438c3585a09bcf8d0651ce2c591d" gracePeriod=30 Dec 01 14:56:56 crc kubenswrapper[4810]: I1201 14:56:56.842920 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerName="sg-core" containerID="cri-o://c393f75663b47fca078d96f2d48cea2b4c160dec49669eb8abe649f5407d5e69" gracePeriod=30 Dec 01 14:56:57 crc kubenswrapper[4810]: I1201 14:56:57.853402 4810 generic.go:334] "Generic (PLEG): container finished" podID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerID="eb41d25cc4e3ac12bc042cf93e3dba592f6e438c3585a09bcf8d0651ce2c591d" exitCode=0 Dec 01 14:56:57 crc kubenswrapper[4810]: I1201 14:56:57.853750 4810 generic.go:334] "Generic (PLEG): container finished" podID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerID="c393f75663b47fca078d96f2d48cea2b4c160dec49669eb8abe649f5407d5e69" exitCode=2 Dec 01 14:56:57 crc kubenswrapper[4810]: I1201 14:56:57.853761 4810 generic.go:334] "Generic (PLEG): container finished" podID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerID="cd9ed3a7eea9c764a163d1a66f52a918a08ff86b5d992b9a8c69e03959355c2a" exitCode=0 Dec 01 14:56:57 crc kubenswrapper[4810]: I1201 14:56:57.853555 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab1a1ee-52fd-4c95-a15a-23a8b5766385","Type":"ContainerDied","Data":"eb41d25cc4e3ac12bc042cf93e3dba592f6e438c3585a09bcf8d0651ce2c591d"} Dec 01 14:56:57 crc kubenswrapper[4810]: I1201 14:56:57.854081 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab1a1ee-52fd-4c95-a15a-23a8b5766385","Type":"ContainerDied","Data":"c393f75663b47fca078d96f2d48cea2b4c160dec49669eb8abe649f5407d5e69"} Dec 01 14:56:57 crc kubenswrapper[4810]: I1201 14:56:57.854105 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab1a1ee-52fd-4c95-a15a-23a8b5766385","Type":"ContainerDied","Data":"cd9ed3a7eea9c764a163d1a66f52a918a08ff86b5d992b9a8c69e03959355c2a"} Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.605249 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.792110 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab1a1ee-52fd-4c95-a15a-23a8b5766385-log-httpd\") pod \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.792196 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab1a1ee-52fd-4c95-a15a-23a8b5766385-run-httpd\") pod \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.792324 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-combined-ca-bundle\") pod \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.792672 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aab1a1ee-52fd-4c95-a15a-23a8b5766385-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "aab1a1ee-52fd-4c95-a15a-23a8b5766385" (UID: "aab1a1ee-52fd-4c95-a15a-23a8b5766385"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.792761 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aab1a1ee-52fd-4c95-a15a-23a8b5766385-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "aab1a1ee-52fd-4c95-a15a-23a8b5766385" (UID: "aab1a1ee-52fd-4c95-a15a-23a8b5766385"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.793069 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfcr4\" (UniqueName: \"kubernetes.io/projected/aab1a1ee-52fd-4c95-a15a-23a8b5766385-kube-api-access-qfcr4\") pod \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.793367 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-sg-core-conf-yaml\") pod \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.793407 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-config-data\") pod \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.793454 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-scripts\") pod \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\" (UID: \"aab1a1ee-52fd-4c95-a15a-23a8b5766385\") " Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.794063 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab1a1ee-52fd-4c95-a15a-23a8b5766385-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.794086 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab1a1ee-52fd-4c95-a15a-23a8b5766385-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.799185 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aab1a1ee-52fd-4c95-a15a-23a8b5766385-kube-api-access-qfcr4" (OuterVolumeSpecName: "kube-api-access-qfcr4") pod "aab1a1ee-52fd-4c95-a15a-23a8b5766385" (UID: "aab1a1ee-52fd-4c95-a15a-23a8b5766385"). InnerVolumeSpecName "kube-api-access-qfcr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.800598 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-scripts" (OuterVolumeSpecName: "scripts") pod "aab1a1ee-52fd-4c95-a15a-23a8b5766385" (UID: "aab1a1ee-52fd-4c95-a15a-23a8b5766385"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.833432 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "aab1a1ee-52fd-4c95-a15a-23a8b5766385" (UID: "aab1a1ee-52fd-4c95-a15a-23a8b5766385"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.892853 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-config-data" (OuterVolumeSpecName: "config-data") pod "aab1a1ee-52fd-4c95-a15a-23a8b5766385" (UID: "aab1a1ee-52fd-4c95-a15a-23a8b5766385"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.896028 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfcr4\" (UniqueName: \"kubernetes.io/projected/aab1a1ee-52fd-4c95-a15a-23a8b5766385-kube-api-access-qfcr4\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.896060 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.896070 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.896078 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.899460 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aab1a1ee-52fd-4c95-a15a-23a8b5766385" (UID: "aab1a1ee-52fd-4c95-a15a-23a8b5766385"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.926958 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gnmm4" event={"ID":"21fcb95f-afd8-43e5-90b2-afa5346fcd57","Type":"ContainerStarted","Data":"cd15449ff36f15ba79faeb70f743c8df8b5661e48cc14348d7fb12694847e714"} Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.929365 4810 generic.go:334] "Generic (PLEG): container finished" podID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerID="4411a834043bfbd4c6eebbfb6f9aefd3cccebe1398e3665943b43e9bfd8949b6" exitCode=0 Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.929401 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab1a1ee-52fd-4c95-a15a-23a8b5766385","Type":"ContainerDied","Data":"4411a834043bfbd4c6eebbfb6f9aefd3cccebe1398e3665943b43e9bfd8949b6"} Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.929421 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab1a1ee-52fd-4c95-a15a-23a8b5766385","Type":"ContainerDied","Data":"b141be6be2c52ce8cc1c317f0d800dd5c81718b031ddc68541f12f6ae2997b5c"} Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.929436 4810 scope.go:117] "RemoveContainer" containerID="eb41d25cc4e3ac12bc042cf93e3dba592f6e438c3585a09bcf8d0651ce2c591d" Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.929543 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.949380 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-gnmm4" podStartSLOduration=2.562402223 podStartE2EDuration="14.949358518s" podCreationTimestamp="2025-12-01 14:56:51 +0000 UTC" firstStartedPulling="2025-12-01 14:56:52.499267235 +0000 UTC m=+1378.262776838" lastFinishedPulling="2025-12-01 14:57:04.88622353 +0000 UTC m=+1390.649733133" observedRunningTime="2025-12-01 14:57:05.943762296 +0000 UTC m=+1391.707271899" watchObservedRunningTime="2025-12-01 14:57:05.949358518 +0000 UTC m=+1391.712868121" Dec 01 14:57:05 crc kubenswrapper[4810]: I1201 14:57:05.998371 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab1a1ee-52fd-4c95-a15a-23a8b5766385-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.028153 4810 scope.go:117] "RemoveContainer" containerID="c393f75663b47fca078d96f2d48cea2b4c160dec49669eb8abe649f5407d5e69" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.042584 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.051694 4810 scope.go:117] "RemoveContainer" containerID="cd9ed3a7eea9c764a163d1a66f52a918a08ff86b5d992b9a8c69e03959355c2a" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.057150 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.070276 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:57:06 crc kubenswrapper[4810]: E1201 14:57:06.070732 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerName="proxy-httpd" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.070755 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerName="proxy-httpd" Dec 01 14:57:06 crc kubenswrapper[4810]: E1201 14:57:06.070794 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerName="ceilometer-central-agent" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.070802 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerName="ceilometer-central-agent" Dec 01 14:57:06 crc kubenswrapper[4810]: E1201 14:57:06.070823 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerName="sg-core" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.070831 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerName="sg-core" Dec 01 14:57:06 crc kubenswrapper[4810]: E1201 14:57:06.070842 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerName="ceilometer-notification-agent" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.070850 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerName="ceilometer-notification-agent" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.071050 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerName="sg-core" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.071071 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerName="ceilometer-central-agent" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.071103 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerName="ceilometer-notification-agent" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.071116 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" containerName="proxy-httpd" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.072899 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.076176 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.076543 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.079413 4810 scope.go:117] "RemoveContainer" containerID="4411a834043bfbd4c6eebbfb6f9aefd3cccebe1398e3665943b43e9bfd8949b6" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.105298 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.110664 4810 scope.go:117] "RemoveContainer" containerID="eb41d25cc4e3ac12bc042cf93e3dba592f6e438c3585a09bcf8d0651ce2c591d" Dec 01 14:57:06 crc kubenswrapper[4810]: E1201 14:57:06.114404 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb41d25cc4e3ac12bc042cf93e3dba592f6e438c3585a09bcf8d0651ce2c591d\": container with ID starting with eb41d25cc4e3ac12bc042cf93e3dba592f6e438c3585a09bcf8d0651ce2c591d not found: ID does not exist" containerID="eb41d25cc4e3ac12bc042cf93e3dba592f6e438c3585a09bcf8d0651ce2c591d" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.114449 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb41d25cc4e3ac12bc042cf93e3dba592f6e438c3585a09bcf8d0651ce2c591d"} err="failed to get container status \"eb41d25cc4e3ac12bc042cf93e3dba592f6e438c3585a09bcf8d0651ce2c591d\": rpc error: code = NotFound desc = could not find container \"eb41d25cc4e3ac12bc042cf93e3dba592f6e438c3585a09bcf8d0651ce2c591d\": container with ID starting with eb41d25cc4e3ac12bc042cf93e3dba592f6e438c3585a09bcf8d0651ce2c591d not found: ID does not exist" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.114494 4810 scope.go:117] "RemoveContainer" containerID="c393f75663b47fca078d96f2d48cea2b4c160dec49669eb8abe649f5407d5e69" Dec 01 14:57:06 crc kubenswrapper[4810]: E1201 14:57:06.115127 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c393f75663b47fca078d96f2d48cea2b4c160dec49669eb8abe649f5407d5e69\": container with ID starting with c393f75663b47fca078d96f2d48cea2b4c160dec49669eb8abe649f5407d5e69 not found: ID does not exist" containerID="c393f75663b47fca078d96f2d48cea2b4c160dec49669eb8abe649f5407d5e69" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.115151 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c393f75663b47fca078d96f2d48cea2b4c160dec49669eb8abe649f5407d5e69"} err="failed to get container status \"c393f75663b47fca078d96f2d48cea2b4c160dec49669eb8abe649f5407d5e69\": rpc error: code = NotFound desc = could not find container \"c393f75663b47fca078d96f2d48cea2b4c160dec49669eb8abe649f5407d5e69\": container with ID starting with c393f75663b47fca078d96f2d48cea2b4c160dec49669eb8abe649f5407d5e69 not found: ID does not exist" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.115167 4810 scope.go:117] "RemoveContainer" containerID="cd9ed3a7eea9c764a163d1a66f52a918a08ff86b5d992b9a8c69e03959355c2a" Dec 01 14:57:06 crc kubenswrapper[4810]: E1201 14:57:06.115408 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd9ed3a7eea9c764a163d1a66f52a918a08ff86b5d992b9a8c69e03959355c2a\": container with ID starting with cd9ed3a7eea9c764a163d1a66f52a918a08ff86b5d992b9a8c69e03959355c2a not found: ID does not exist" containerID="cd9ed3a7eea9c764a163d1a66f52a918a08ff86b5d992b9a8c69e03959355c2a" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.115429 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd9ed3a7eea9c764a163d1a66f52a918a08ff86b5d992b9a8c69e03959355c2a"} err="failed to get container status \"cd9ed3a7eea9c764a163d1a66f52a918a08ff86b5d992b9a8c69e03959355c2a\": rpc error: code = NotFound desc = could not find container \"cd9ed3a7eea9c764a163d1a66f52a918a08ff86b5d992b9a8c69e03959355c2a\": container with ID starting with cd9ed3a7eea9c764a163d1a66f52a918a08ff86b5d992b9a8c69e03959355c2a not found: ID does not exist" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.115445 4810 scope.go:117] "RemoveContainer" containerID="4411a834043bfbd4c6eebbfb6f9aefd3cccebe1398e3665943b43e9bfd8949b6" Dec 01 14:57:06 crc kubenswrapper[4810]: E1201 14:57:06.115663 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4411a834043bfbd4c6eebbfb6f9aefd3cccebe1398e3665943b43e9bfd8949b6\": container with ID starting with 4411a834043bfbd4c6eebbfb6f9aefd3cccebe1398e3665943b43e9bfd8949b6 not found: ID does not exist" containerID="4411a834043bfbd4c6eebbfb6f9aefd3cccebe1398e3665943b43e9bfd8949b6" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.115684 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4411a834043bfbd4c6eebbfb6f9aefd3cccebe1398e3665943b43e9bfd8949b6"} err="failed to get container status \"4411a834043bfbd4c6eebbfb6f9aefd3cccebe1398e3665943b43e9bfd8949b6\": rpc error: code = NotFound desc = could not find container \"4411a834043bfbd4c6eebbfb6f9aefd3cccebe1398e3665943b43e9bfd8949b6\": container with ID starting with 4411a834043bfbd4c6eebbfb6f9aefd3cccebe1398e3665943b43e9bfd8949b6 not found: ID does not exist" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.201301 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-scripts\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.201627 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.201785 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-run-httpd\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.201926 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.202020 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nl2zj\" (UniqueName: \"kubernetes.io/projected/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-kube-api-access-nl2zj\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.202047 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-config-data\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.202156 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-log-httpd\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.304120 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-run-httpd\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.304194 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.304245 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nl2zj\" (UniqueName: \"kubernetes.io/projected/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-kube-api-access-nl2zj\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.304280 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-config-data\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.304675 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-log-httpd\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.305055 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-log-httpd\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.304699 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-scripts\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.305186 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.305873 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-run-httpd\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.308328 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-scripts\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.311818 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-config-data\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.315396 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.325988 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.329164 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nl2zj\" (UniqueName: \"kubernetes.io/projected/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-kube-api-access-nl2zj\") pod \"ceilometer-0\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.410829 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.505341 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aab1a1ee-52fd-4c95-a15a-23a8b5766385" path="/var/lib/kubelet/pods/aab1a1ee-52fd-4c95-a15a-23a8b5766385/volumes" Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.871213 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:57:06 crc kubenswrapper[4810]: I1201 14:57:06.938333 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba866fb-11bf-4edf-a0a9-1e7534cae07f","Type":"ContainerStarted","Data":"0e208fba9fb7dcfaa22465c42f0df0caa69e2703c19481e23e4ca76e8263d78c"} Dec 01 14:57:07 crc kubenswrapper[4810]: I1201 14:57:07.948801 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba866fb-11bf-4edf-a0a9-1e7534cae07f","Type":"ContainerStarted","Data":"f741f8a0ce24ec2347b6ff71f1e6d56ee41b153add26bf402baca5a5b8eff0c1"} Dec 01 14:57:08 crc kubenswrapper[4810]: I1201 14:57:08.958305 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba866fb-11bf-4edf-a0a9-1e7534cae07f","Type":"ContainerStarted","Data":"0828594d163132cb7db5ec6c8956f48822a6bab045cc6ab9d7c946f86312ac7f"} Dec 01 14:57:09 crc kubenswrapper[4810]: I1201 14:57:09.971840 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba866fb-11bf-4edf-a0a9-1e7534cae07f","Type":"ContainerStarted","Data":"95ea32388466bbc69b18495263c46fffd9ccc7121f3e7e29fe8e9c0b05de57e4"} Dec 01 14:57:11 crc kubenswrapper[4810]: I1201 14:57:11.464513 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="d18c05bc-3e20-4305-b8bf-d45e6ff46308" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.146:9292/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 01 14:57:11 crc kubenswrapper[4810]: I1201 14:57:11.465827 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="d18c05bc-3e20-4305-b8bf-d45e6ff46308" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.146:9292/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 01 14:57:11 crc kubenswrapper[4810]: I1201 14:57:11.991867 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba866fb-11bf-4edf-a0a9-1e7534cae07f","Type":"ContainerStarted","Data":"0d8769d7c7d03b825251d3d117c83522bf3d8aa26d1c14009417406e1c583738"} Dec 01 14:57:11 crc kubenswrapper[4810]: I1201 14:57:11.992588 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 14:57:12 crc kubenswrapper[4810]: I1201 14:57:12.023860 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.486010667 podStartE2EDuration="6.023837797s" podCreationTimestamp="2025-12-01 14:57:06 +0000 UTC" firstStartedPulling="2025-12-01 14:57:06.874927499 +0000 UTC m=+1392.638437102" lastFinishedPulling="2025-12-01 14:57:11.412754629 +0000 UTC m=+1397.176264232" observedRunningTime="2025-12-01 14:57:12.020109695 +0000 UTC m=+1397.783619308" watchObservedRunningTime="2025-12-01 14:57:12.023837797 +0000 UTC m=+1397.787347400" Dec 01 14:57:18 crc kubenswrapper[4810]: I1201 14:57:18.041588 4810 generic.go:334] "Generic (PLEG): container finished" podID="21fcb95f-afd8-43e5-90b2-afa5346fcd57" containerID="cd15449ff36f15ba79faeb70f743c8df8b5661e48cc14348d7fb12694847e714" exitCode=0 Dec 01 14:57:18 crc kubenswrapper[4810]: I1201 14:57:18.041689 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gnmm4" event={"ID":"21fcb95f-afd8-43e5-90b2-afa5346fcd57","Type":"ContainerDied","Data":"cd15449ff36f15ba79faeb70f743c8df8b5661e48cc14348d7fb12694847e714"} Dec 01 14:57:19 crc kubenswrapper[4810]: I1201 14:57:19.433291 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gnmm4" Dec 01 14:57:19 crc kubenswrapper[4810]: I1201 14:57:19.545792 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21fcb95f-afd8-43e5-90b2-afa5346fcd57-combined-ca-bundle\") pod \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\" (UID: \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\") " Dec 01 14:57:19 crc kubenswrapper[4810]: I1201 14:57:19.545892 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21fcb95f-afd8-43e5-90b2-afa5346fcd57-scripts\") pod \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\" (UID: \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\") " Dec 01 14:57:19 crc kubenswrapper[4810]: I1201 14:57:19.545935 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lplt\" (UniqueName: \"kubernetes.io/projected/21fcb95f-afd8-43e5-90b2-afa5346fcd57-kube-api-access-2lplt\") pod \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\" (UID: \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\") " Dec 01 14:57:19 crc kubenswrapper[4810]: I1201 14:57:19.546050 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21fcb95f-afd8-43e5-90b2-afa5346fcd57-config-data\") pod \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\" (UID: \"21fcb95f-afd8-43e5-90b2-afa5346fcd57\") " Dec 01 14:57:19 crc kubenswrapper[4810]: I1201 14:57:19.553197 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21fcb95f-afd8-43e5-90b2-afa5346fcd57-kube-api-access-2lplt" (OuterVolumeSpecName: "kube-api-access-2lplt") pod "21fcb95f-afd8-43e5-90b2-afa5346fcd57" (UID: "21fcb95f-afd8-43e5-90b2-afa5346fcd57"). InnerVolumeSpecName "kube-api-access-2lplt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:57:19 crc kubenswrapper[4810]: I1201 14:57:19.563850 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21fcb95f-afd8-43e5-90b2-afa5346fcd57-scripts" (OuterVolumeSpecName: "scripts") pod "21fcb95f-afd8-43e5-90b2-afa5346fcd57" (UID: "21fcb95f-afd8-43e5-90b2-afa5346fcd57"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:19 crc kubenswrapper[4810]: I1201 14:57:19.575730 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21fcb95f-afd8-43e5-90b2-afa5346fcd57-config-data" (OuterVolumeSpecName: "config-data") pod "21fcb95f-afd8-43e5-90b2-afa5346fcd57" (UID: "21fcb95f-afd8-43e5-90b2-afa5346fcd57"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:19 crc kubenswrapper[4810]: I1201 14:57:19.587505 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21fcb95f-afd8-43e5-90b2-afa5346fcd57-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21fcb95f-afd8-43e5-90b2-afa5346fcd57" (UID: "21fcb95f-afd8-43e5-90b2-afa5346fcd57"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:19 crc kubenswrapper[4810]: I1201 14:57:19.647958 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21fcb95f-afd8-43e5-90b2-afa5346fcd57-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:19 crc kubenswrapper[4810]: I1201 14:57:19.648571 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21fcb95f-afd8-43e5-90b2-afa5346fcd57-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:19 crc kubenswrapper[4810]: I1201 14:57:19.648603 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lplt\" (UniqueName: \"kubernetes.io/projected/21fcb95f-afd8-43e5-90b2-afa5346fcd57-kube-api-access-2lplt\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:19 crc kubenswrapper[4810]: I1201 14:57:19.648618 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21fcb95f-afd8-43e5-90b2-afa5346fcd57-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.059450 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gnmm4" event={"ID":"21fcb95f-afd8-43e5-90b2-afa5346fcd57","Type":"ContainerDied","Data":"863196f2d4fdac01ce0e09b6dc45e85c419b7dc200a5980e800d70f1cfdc61ac"} Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.059773 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="863196f2d4fdac01ce0e09b6dc45e85c419b7dc200a5980e800d70f1cfdc61ac" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.059526 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gnmm4" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.162652 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 14:57:20 crc kubenswrapper[4810]: E1201 14:57:20.163053 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21fcb95f-afd8-43e5-90b2-afa5346fcd57" containerName="nova-cell0-conductor-db-sync" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.163071 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="21fcb95f-afd8-43e5-90b2-afa5346fcd57" containerName="nova-cell0-conductor-db-sync" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.163251 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="21fcb95f-afd8-43e5-90b2-afa5346fcd57" containerName="nova-cell0-conductor-db-sync" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.163826 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.173335 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-68sxs" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.173628 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.188415 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.258625 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c73a61da-0c40-4a26-888c-ab5b994417d4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c73a61da-0c40-4a26-888c-ab5b994417d4\") " pod="openstack/nova-cell0-conductor-0" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.258668 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c73a61da-0c40-4a26-888c-ab5b994417d4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c73a61da-0c40-4a26-888c-ab5b994417d4\") " pod="openstack/nova-cell0-conductor-0" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.258780 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9jdn\" (UniqueName: \"kubernetes.io/projected/c73a61da-0c40-4a26-888c-ab5b994417d4-kube-api-access-w9jdn\") pod \"nova-cell0-conductor-0\" (UID: \"c73a61da-0c40-4a26-888c-ab5b994417d4\") " pod="openstack/nova-cell0-conductor-0" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.359983 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c73a61da-0c40-4a26-888c-ab5b994417d4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c73a61da-0c40-4a26-888c-ab5b994417d4\") " pod="openstack/nova-cell0-conductor-0" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.360037 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c73a61da-0c40-4a26-888c-ab5b994417d4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c73a61da-0c40-4a26-888c-ab5b994417d4\") " pod="openstack/nova-cell0-conductor-0" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.360175 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9jdn\" (UniqueName: \"kubernetes.io/projected/c73a61da-0c40-4a26-888c-ab5b994417d4-kube-api-access-w9jdn\") pod \"nova-cell0-conductor-0\" (UID: \"c73a61da-0c40-4a26-888c-ab5b994417d4\") " pod="openstack/nova-cell0-conductor-0" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.371861 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c73a61da-0c40-4a26-888c-ab5b994417d4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c73a61da-0c40-4a26-888c-ab5b994417d4\") " pod="openstack/nova-cell0-conductor-0" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.371924 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c73a61da-0c40-4a26-888c-ab5b994417d4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c73a61da-0c40-4a26-888c-ab5b994417d4\") " pod="openstack/nova-cell0-conductor-0" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.375616 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9jdn\" (UniqueName: \"kubernetes.io/projected/c73a61da-0c40-4a26-888c-ab5b994417d4-kube-api-access-w9jdn\") pod \"nova-cell0-conductor-0\" (UID: \"c73a61da-0c40-4a26-888c-ab5b994417d4\") " pod="openstack/nova-cell0-conductor-0" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.491099 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 14:57:20 crc kubenswrapper[4810]: I1201 14:57:20.929002 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 14:57:21 crc kubenswrapper[4810]: I1201 14:57:21.070671 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c73a61da-0c40-4a26-888c-ab5b994417d4","Type":"ContainerStarted","Data":"0427c38a149f10233ec14f720e735356ae66eeb0a15ee032d631eff0f84f5475"} Dec 01 14:57:22 crc kubenswrapper[4810]: I1201 14:57:22.079730 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c73a61da-0c40-4a26-888c-ab5b994417d4","Type":"ContainerStarted","Data":"6cffd8cec4aa7e3f42f1a3b9cff8d7da52efafaece6f09b30bae42093876b093"} Dec 01 14:57:22 crc kubenswrapper[4810]: I1201 14:57:22.080888 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 01 14:57:22 crc kubenswrapper[4810]: I1201 14:57:22.097852 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.097833855 podStartE2EDuration="2.097833855s" podCreationTimestamp="2025-12-01 14:57:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:57:22.092780527 +0000 UTC m=+1407.856290120" watchObservedRunningTime="2025-12-01 14:57:22.097833855 +0000 UTC m=+1407.861343458" Dec 01 14:57:30 crc kubenswrapper[4810]: I1201 14:57:30.528797 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 01 14:57:30 crc kubenswrapper[4810]: I1201 14:57:30.958810 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-78w6c"] Dec 01 14:57:30 crc kubenswrapper[4810]: I1201 14:57:30.960179 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-78w6c" Dec 01 14:57:30 crc kubenswrapper[4810]: I1201 14:57:30.961932 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 01 14:57:30 crc kubenswrapper[4810]: I1201 14:57:30.962191 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 01 14:57:30 crc kubenswrapper[4810]: I1201 14:57:30.971939 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-78w6c"] Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.047180 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bbc4102-9734-42a7-8bd5-9487e83d768b-scripts\") pod \"nova-cell0-cell-mapping-78w6c\" (UID: \"8bbc4102-9734-42a7-8bd5-9487e83d768b\") " pod="openstack/nova-cell0-cell-mapping-78w6c" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.047558 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bbc4102-9734-42a7-8bd5-9487e83d768b-config-data\") pod \"nova-cell0-cell-mapping-78w6c\" (UID: \"8bbc4102-9734-42a7-8bd5-9487e83d768b\") " pod="openstack/nova-cell0-cell-mapping-78w6c" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.047587 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs2gw\" (UniqueName: \"kubernetes.io/projected/8bbc4102-9734-42a7-8bd5-9487e83d768b-kube-api-access-rs2gw\") pod \"nova-cell0-cell-mapping-78w6c\" (UID: \"8bbc4102-9734-42a7-8bd5-9487e83d768b\") " pod="openstack/nova-cell0-cell-mapping-78w6c" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.047606 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bbc4102-9734-42a7-8bd5-9487e83d768b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-78w6c\" (UID: \"8bbc4102-9734-42a7-8bd5-9487e83d768b\") " pod="openstack/nova-cell0-cell-mapping-78w6c" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.150453 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bbc4102-9734-42a7-8bd5-9487e83d768b-config-data\") pod \"nova-cell0-cell-mapping-78w6c\" (UID: \"8bbc4102-9734-42a7-8bd5-9487e83d768b\") " pod="openstack/nova-cell0-cell-mapping-78w6c" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.150519 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs2gw\" (UniqueName: \"kubernetes.io/projected/8bbc4102-9734-42a7-8bd5-9487e83d768b-kube-api-access-rs2gw\") pod \"nova-cell0-cell-mapping-78w6c\" (UID: \"8bbc4102-9734-42a7-8bd5-9487e83d768b\") " pod="openstack/nova-cell0-cell-mapping-78w6c" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.150539 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bbc4102-9734-42a7-8bd5-9487e83d768b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-78w6c\" (UID: \"8bbc4102-9734-42a7-8bd5-9487e83d768b\") " pod="openstack/nova-cell0-cell-mapping-78w6c" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.150657 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bbc4102-9734-42a7-8bd5-9487e83d768b-scripts\") pod \"nova-cell0-cell-mapping-78w6c\" (UID: \"8bbc4102-9734-42a7-8bd5-9487e83d768b\") " pod="openstack/nova-cell0-cell-mapping-78w6c" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.160317 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.164644 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.173166 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.187098 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bbc4102-9734-42a7-8bd5-9487e83d768b-scripts\") pod \"nova-cell0-cell-mapping-78w6c\" (UID: \"8bbc4102-9734-42a7-8bd5-9487e83d768b\") " pod="openstack/nova-cell0-cell-mapping-78w6c" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.187646 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs2gw\" (UniqueName: \"kubernetes.io/projected/8bbc4102-9734-42a7-8bd5-9487e83d768b-kube-api-access-rs2gw\") pod \"nova-cell0-cell-mapping-78w6c\" (UID: \"8bbc4102-9734-42a7-8bd5-9487e83d768b\") " pod="openstack/nova-cell0-cell-mapping-78w6c" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.195417 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bbc4102-9734-42a7-8bd5-9487e83d768b-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-78w6c\" (UID: \"8bbc4102-9734-42a7-8bd5-9487e83d768b\") " pod="openstack/nova-cell0-cell-mapping-78w6c" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.195943 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bbc4102-9734-42a7-8bd5-9487e83d768b-config-data\") pod \"nova-cell0-cell-mapping-78w6c\" (UID: \"8bbc4102-9734-42a7-8bd5-9487e83d768b\") " pod="openstack/nova-cell0-cell-mapping-78w6c" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.196563 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.254593 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2ftl\" (UniqueName: \"kubernetes.io/projected/55df027d-d8ed-477f-85e8-9c5548846f2b-kube-api-access-g2ftl\") pod \"nova-api-0\" (UID: \"55df027d-d8ed-477f-85e8-9c5548846f2b\") " pod="openstack/nova-api-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.254797 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55df027d-d8ed-477f-85e8-9c5548846f2b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"55df027d-d8ed-477f-85e8-9c5548846f2b\") " pod="openstack/nova-api-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.254821 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55df027d-d8ed-477f-85e8-9c5548846f2b-config-data\") pod \"nova-api-0\" (UID: \"55df027d-d8ed-477f-85e8-9c5548846f2b\") " pod="openstack/nova-api-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.254855 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55df027d-d8ed-477f-85e8-9c5548846f2b-logs\") pod \"nova-api-0\" (UID: \"55df027d-d8ed-477f-85e8-9c5548846f2b\") " pod="openstack/nova-api-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.261241 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.263233 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.266012 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.283896 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-78w6c" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.296085 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.356666 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92e83ed1-117e-4772-a9d5-f7888a952207-logs\") pod \"nova-metadata-0\" (UID: \"92e83ed1-117e-4772-a9d5-f7888a952207\") " pod="openstack/nova-metadata-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.356729 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55df027d-d8ed-477f-85e8-9c5548846f2b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"55df027d-d8ed-477f-85e8-9c5548846f2b\") " pod="openstack/nova-api-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.356748 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55df027d-d8ed-477f-85e8-9c5548846f2b-config-data\") pod \"nova-api-0\" (UID: \"55df027d-d8ed-477f-85e8-9c5548846f2b\") " pod="openstack/nova-api-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.356775 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55df027d-d8ed-477f-85e8-9c5548846f2b-logs\") pod \"nova-api-0\" (UID: \"55df027d-d8ed-477f-85e8-9c5548846f2b\") " pod="openstack/nova-api-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.356797 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92e83ed1-117e-4772-a9d5-f7888a952207-config-data\") pod \"nova-metadata-0\" (UID: \"92e83ed1-117e-4772-a9d5-f7888a952207\") " pod="openstack/nova-metadata-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.356847 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4hs2\" (UniqueName: \"kubernetes.io/projected/92e83ed1-117e-4772-a9d5-f7888a952207-kube-api-access-l4hs2\") pod \"nova-metadata-0\" (UID: \"92e83ed1-117e-4772-a9d5-f7888a952207\") " pod="openstack/nova-metadata-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.356873 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92e83ed1-117e-4772-a9d5-f7888a952207-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"92e83ed1-117e-4772-a9d5-f7888a952207\") " pod="openstack/nova-metadata-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.356892 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2ftl\" (UniqueName: \"kubernetes.io/projected/55df027d-d8ed-477f-85e8-9c5548846f2b-kube-api-access-g2ftl\") pod \"nova-api-0\" (UID: \"55df027d-d8ed-477f-85e8-9c5548846f2b\") " pod="openstack/nova-api-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.358433 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55df027d-d8ed-477f-85e8-9c5548846f2b-logs\") pod \"nova-api-0\" (UID: \"55df027d-d8ed-477f-85e8-9c5548846f2b\") " pod="openstack/nova-api-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.360816 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.365555 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.367083 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55df027d-d8ed-477f-85e8-9c5548846f2b-config-data\") pod \"nova-api-0\" (UID: \"55df027d-d8ed-477f-85e8-9c5548846f2b\") " pod="openstack/nova-api-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.368194 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55df027d-d8ed-477f-85e8-9c5548846f2b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"55df027d-d8ed-477f-85e8-9c5548846f2b\") " pod="openstack/nova-api-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.388901 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.406206 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2ftl\" (UniqueName: \"kubernetes.io/projected/55df027d-d8ed-477f-85e8-9c5548846f2b-kube-api-access-g2ftl\") pod \"nova-api-0\" (UID: \"55df027d-d8ed-477f-85e8-9c5548846f2b\") " pod="openstack/nova-api-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.434551 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.435640 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.438721 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.456898 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.457878 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92e83ed1-117e-4772-a9d5-f7888a952207-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"92e83ed1-117e-4772-a9d5-f7888a952207\") " pod="openstack/nova-metadata-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.457946 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a18952e-ee54-4510-b149-e9e5cb873af5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1a18952e-ee54-4510-b149-e9e5cb873af5\") " pod="openstack/nova-scheduler-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.458001 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92e83ed1-117e-4772-a9d5-f7888a952207-logs\") pod \"nova-metadata-0\" (UID: \"92e83ed1-117e-4772-a9d5-f7888a952207\") " pod="openstack/nova-metadata-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.458025 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a18952e-ee54-4510-b149-e9e5cb873af5-config-data\") pod \"nova-scheduler-0\" (UID: \"1a18952e-ee54-4510-b149-e9e5cb873af5\") " pod="openstack/nova-scheduler-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.458059 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f54zq\" (UniqueName: \"kubernetes.io/projected/1a18952e-ee54-4510-b149-e9e5cb873af5-kube-api-access-f54zq\") pod \"nova-scheduler-0\" (UID: \"1a18952e-ee54-4510-b149-e9e5cb873af5\") " pod="openstack/nova-scheduler-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.458086 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92e83ed1-117e-4772-a9d5-f7888a952207-config-data\") pod \"nova-metadata-0\" (UID: \"92e83ed1-117e-4772-a9d5-f7888a952207\") " pod="openstack/nova-metadata-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.458116 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4hs2\" (UniqueName: \"kubernetes.io/projected/92e83ed1-117e-4772-a9d5-f7888a952207-kube-api-access-l4hs2\") pod \"nova-metadata-0\" (UID: \"92e83ed1-117e-4772-a9d5-f7888a952207\") " pod="openstack/nova-metadata-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.460939 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92e83ed1-117e-4772-a9d5-f7888a952207-logs\") pod \"nova-metadata-0\" (UID: \"92e83ed1-117e-4772-a9d5-f7888a952207\") " pod="openstack/nova-metadata-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.467931 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92e83ed1-117e-4772-a9d5-f7888a952207-config-data\") pod \"nova-metadata-0\" (UID: \"92e83ed1-117e-4772-a9d5-f7888a952207\") " pod="openstack/nova-metadata-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.474062 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92e83ed1-117e-4772-a9d5-f7888a952207-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"92e83ed1-117e-4772-a9d5-f7888a952207\") " pod="openstack/nova-metadata-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.500085 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.513859 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4hs2\" (UniqueName: \"kubernetes.io/projected/92e83ed1-117e-4772-a9d5-f7888a952207-kube-api-access-l4hs2\") pod \"nova-metadata-0\" (UID: \"92e83ed1-117e-4772-a9d5-f7888a952207\") " pod="openstack/nova-metadata-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.516549 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5594d9b959-wrzd5"] Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.518437 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.561684 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a18952e-ee54-4510-b149-e9e5cb873af5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1a18952e-ee54-4510-b149-e9e5cb873af5\") " pod="openstack/nova-scheduler-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.561822 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t97q7\" (UniqueName: \"kubernetes.io/projected/20bf5931-cfe6-4607-89dc-164eef1386a9-kube-api-access-t97q7\") pod \"nova-cell1-novncproxy-0\" (UID: \"20bf5931-cfe6-4607-89dc-164eef1386a9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.561853 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a18952e-ee54-4510-b149-e9e5cb873af5-config-data\") pod \"nova-scheduler-0\" (UID: \"1a18952e-ee54-4510-b149-e9e5cb873af5\") " pod="openstack/nova-scheduler-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.561905 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f54zq\" (UniqueName: \"kubernetes.io/projected/1a18952e-ee54-4510-b149-e9e5cb873af5-kube-api-access-f54zq\") pod \"nova-scheduler-0\" (UID: \"1a18952e-ee54-4510-b149-e9e5cb873af5\") " pod="openstack/nova-scheduler-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.562033 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20bf5931-cfe6-4607-89dc-164eef1386a9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"20bf5931-cfe6-4607-89dc-164eef1386a9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.562056 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20bf5931-cfe6-4607-89dc-164eef1386a9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"20bf5931-cfe6-4607-89dc-164eef1386a9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.571031 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a18952e-ee54-4510-b149-e9e5cb873af5-config-data\") pod \"nova-scheduler-0\" (UID: \"1a18952e-ee54-4510-b149-e9e5cb873af5\") " pod="openstack/nova-scheduler-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.571836 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5594d9b959-wrzd5"] Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.574158 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a18952e-ee54-4510-b149-e9e5cb873af5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1a18952e-ee54-4510-b149-e9e5cb873af5\") " pod="openstack/nova-scheduler-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.579917 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.599313 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.622128 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f54zq\" (UniqueName: \"kubernetes.io/projected/1a18952e-ee54-4510-b149-e9e5cb873af5-kube-api-access-f54zq\") pod \"nova-scheduler-0\" (UID: \"1a18952e-ee54-4510-b149-e9e5cb873af5\") " pod="openstack/nova-scheduler-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.663523 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-dns-svc\") pod \"dnsmasq-dns-5594d9b959-wrzd5\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.663952 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2j48\" (UniqueName: \"kubernetes.io/projected/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-kube-api-access-d2j48\") pod \"dnsmasq-dns-5594d9b959-wrzd5\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.664079 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20bf5931-cfe6-4607-89dc-164eef1386a9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"20bf5931-cfe6-4607-89dc-164eef1386a9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.664161 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20bf5931-cfe6-4607-89dc-164eef1386a9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"20bf5931-cfe6-4607-89dc-164eef1386a9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.664277 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-ovsdbserver-nb\") pod \"dnsmasq-dns-5594d9b959-wrzd5\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.664384 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-config\") pod \"dnsmasq-dns-5594d9b959-wrzd5\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.665244 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-dns-swift-storage-0\") pod \"dnsmasq-dns-5594d9b959-wrzd5\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.665885 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t97q7\" (UniqueName: \"kubernetes.io/projected/20bf5931-cfe6-4607-89dc-164eef1386a9-kube-api-access-t97q7\") pod \"nova-cell1-novncproxy-0\" (UID: \"20bf5931-cfe6-4607-89dc-164eef1386a9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.666263 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-ovsdbserver-sb\") pod \"dnsmasq-dns-5594d9b959-wrzd5\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.687675 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20bf5931-cfe6-4607-89dc-164eef1386a9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"20bf5931-cfe6-4607-89dc-164eef1386a9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.688184 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20bf5931-cfe6-4607-89dc-164eef1386a9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"20bf5931-cfe6-4607-89dc-164eef1386a9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.716715 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t97q7\" (UniqueName: \"kubernetes.io/projected/20bf5931-cfe6-4607-89dc-164eef1386a9-kube-api-access-t97q7\") pod \"nova-cell1-novncproxy-0\" (UID: \"20bf5931-cfe6-4607-89dc-164eef1386a9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.767759 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-ovsdbserver-nb\") pod \"dnsmasq-dns-5594d9b959-wrzd5\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.768120 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-config\") pod \"dnsmasq-dns-5594d9b959-wrzd5\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.768224 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-dns-swift-storage-0\") pod \"dnsmasq-dns-5594d9b959-wrzd5\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.768299 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-ovsdbserver-sb\") pod \"dnsmasq-dns-5594d9b959-wrzd5\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.768507 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-dns-svc\") pod \"dnsmasq-dns-5594d9b959-wrzd5\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.768635 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2j48\" (UniqueName: \"kubernetes.io/projected/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-kube-api-access-d2j48\") pod \"dnsmasq-dns-5594d9b959-wrzd5\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.769005 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-ovsdbserver-nb\") pod \"dnsmasq-dns-5594d9b959-wrzd5\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.769627 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-dns-swift-storage-0\") pod \"dnsmasq-dns-5594d9b959-wrzd5\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.769996 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-config\") pod \"dnsmasq-dns-5594d9b959-wrzd5\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.770328 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-dns-svc\") pod \"dnsmasq-dns-5594d9b959-wrzd5\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.770789 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-ovsdbserver-sb\") pod \"dnsmasq-dns-5594d9b959-wrzd5\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.795718 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2j48\" (UniqueName: \"kubernetes.io/projected/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-kube-api-access-d2j48\") pod \"dnsmasq-dns-5594d9b959-wrzd5\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.866174 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.924938 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:57:31 crc kubenswrapper[4810]: I1201 14:57:31.959984 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.279860 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-78w6c"] Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.410591 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.429112 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:57:32 crc kubenswrapper[4810]: W1201 14:57:32.440890 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55df027d_d8ed_477f_85e8_9c5548846f2b.slice/crio-493f249b8651b93539f277c48d95aad7d598ab0cb025e2d891579188eeb70b12 WatchSource:0}: Error finding container 493f249b8651b93539f277c48d95aad7d598ab0cb025e2d891579188eeb70b12: Status 404 returned error can't find the container with id 493f249b8651b93539f277c48d95aad7d598ab0cb025e2d891579188eeb70b12 Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.544888 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.629171 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-f27f6"] Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.630601 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-f27f6" Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.643064 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.643380 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.653103 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-f27f6"] Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.710904 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.718689 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5594d9b959-wrzd5"] Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.793639 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-f27f6\" (UID: \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\") " pod="openstack/nova-cell1-conductor-db-sync-f27f6" Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.793718 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-config-data\") pod \"nova-cell1-conductor-db-sync-f27f6\" (UID: \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\") " pod="openstack/nova-cell1-conductor-db-sync-f27f6" Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.793814 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmsfs\" (UniqueName: \"kubernetes.io/projected/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-kube-api-access-kmsfs\") pod \"nova-cell1-conductor-db-sync-f27f6\" (UID: \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\") " pod="openstack/nova-cell1-conductor-db-sync-f27f6" Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.793923 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-scripts\") pod \"nova-cell1-conductor-db-sync-f27f6\" (UID: \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\") " pod="openstack/nova-cell1-conductor-db-sync-f27f6" Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.895612 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-config-data\") pod \"nova-cell1-conductor-db-sync-f27f6\" (UID: \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\") " pod="openstack/nova-cell1-conductor-db-sync-f27f6" Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.896387 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmsfs\" (UniqueName: \"kubernetes.io/projected/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-kube-api-access-kmsfs\") pod \"nova-cell1-conductor-db-sync-f27f6\" (UID: \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\") " pod="openstack/nova-cell1-conductor-db-sync-f27f6" Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.896533 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-scripts\") pod \"nova-cell1-conductor-db-sync-f27f6\" (UID: \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\") " pod="openstack/nova-cell1-conductor-db-sync-f27f6" Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.896972 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-f27f6\" (UID: \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\") " pod="openstack/nova-cell1-conductor-db-sync-f27f6" Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.901312 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-scripts\") pod \"nova-cell1-conductor-db-sync-f27f6\" (UID: \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\") " pod="openstack/nova-cell1-conductor-db-sync-f27f6" Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.901453 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-config-data\") pod \"nova-cell1-conductor-db-sync-f27f6\" (UID: \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\") " pod="openstack/nova-cell1-conductor-db-sync-f27f6" Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.901635 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-f27f6\" (UID: \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\") " pod="openstack/nova-cell1-conductor-db-sync-f27f6" Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.920043 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmsfs\" (UniqueName: \"kubernetes.io/projected/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-kube-api-access-kmsfs\") pod \"nova-cell1-conductor-db-sync-f27f6\" (UID: \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\") " pod="openstack/nova-cell1-conductor-db-sync-f27f6" Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.965931 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-f27f6" Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.973679 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:57:32 crc kubenswrapper[4810]: I1201 14:57:32.973739 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:57:33 crc kubenswrapper[4810]: I1201 14:57:33.170189 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-78w6c" event={"ID":"8bbc4102-9734-42a7-8bd5-9487e83d768b","Type":"ContainerStarted","Data":"bcb6d4eef8012d35e7a2581c37d3de047e5d6c95182ecb8fe94d92a03fdbccd4"} Dec 01 14:57:33 crc kubenswrapper[4810]: I1201 14:57:33.170502 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-78w6c" event={"ID":"8bbc4102-9734-42a7-8bd5-9487e83d768b","Type":"ContainerStarted","Data":"b4f6b4670e37a86fc2f55dbfdc350f5c2b8a2b95244231daf8c5e0ba2da33ce9"} Dec 01 14:57:33 crc kubenswrapper[4810]: I1201 14:57:33.180394 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55df027d-d8ed-477f-85e8-9c5548846f2b","Type":"ContainerStarted","Data":"493f249b8651b93539f277c48d95aad7d598ab0cb025e2d891579188eeb70b12"} Dec 01 14:57:33 crc kubenswrapper[4810]: I1201 14:57:33.181816 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"20bf5931-cfe6-4607-89dc-164eef1386a9","Type":"ContainerStarted","Data":"fc2a927f8ae77e44ae3a010953178caf95ee8be5e41c8346e9dc086340771dfe"} Dec 01 14:57:33 crc kubenswrapper[4810]: I1201 14:57:33.183352 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1a18952e-ee54-4510-b149-e9e5cb873af5","Type":"ContainerStarted","Data":"e2eeb1c7160e1a72ac6978e9c38e9af08db57280618b87f96c3f19f337a4ae45"} Dec 01 14:57:33 crc kubenswrapper[4810]: I1201 14:57:33.184758 4810 generic.go:334] "Generic (PLEG): container finished" podID="af2b1a14-3eb7-41b7-90f7-08c4b93525ec" containerID="b21f2a7ba3de1521ea49e1c9aac6a01c8cd63406e9ead0864ac59de0a527ef05" exitCode=0 Dec 01 14:57:33 crc kubenswrapper[4810]: I1201 14:57:33.184802 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" event={"ID":"af2b1a14-3eb7-41b7-90f7-08c4b93525ec","Type":"ContainerDied","Data":"b21f2a7ba3de1521ea49e1c9aac6a01c8cd63406e9ead0864ac59de0a527ef05"} Dec 01 14:57:33 crc kubenswrapper[4810]: I1201 14:57:33.184816 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" event={"ID":"af2b1a14-3eb7-41b7-90f7-08c4b93525ec","Type":"ContainerStarted","Data":"b16040d54efd5cc544057757aad240a38c99cf6dbaf529b13945b0a84863072d"} Dec 01 14:57:33 crc kubenswrapper[4810]: I1201 14:57:33.188119 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"92e83ed1-117e-4772-a9d5-f7888a952207","Type":"ContainerStarted","Data":"5788c0333bea83a21e3a71e513e16dd8f39a1b0336dc89c18621c8d7a803f01c"} Dec 01 14:57:33 crc kubenswrapper[4810]: I1201 14:57:33.207211 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-78w6c" podStartSLOduration=3.207189007 podStartE2EDuration="3.207189007s" podCreationTimestamp="2025-12-01 14:57:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:57:33.198210922 +0000 UTC m=+1418.961720535" watchObservedRunningTime="2025-12-01 14:57:33.207189007 +0000 UTC m=+1418.970698610" Dec 01 14:57:33 crc kubenswrapper[4810]: I1201 14:57:33.442506 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-f27f6"] Dec 01 14:57:34 crc kubenswrapper[4810]: I1201 14:57:34.204418 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-f27f6" event={"ID":"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88","Type":"ContainerStarted","Data":"9a9cb3e6a77baeced59069d7dc5d260e53122dd4e38d62c9a1293deed78602d5"} Dec 01 14:57:34 crc kubenswrapper[4810]: I1201 14:57:34.204784 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-f27f6" event={"ID":"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88","Type":"ContainerStarted","Data":"c5fc07d8e9dbe9a76f702de2bbbd99d9d9ab8da5d433469d5cac9d4c92448907"} Dec 01 14:57:34 crc kubenswrapper[4810]: I1201 14:57:34.214627 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" event={"ID":"af2b1a14-3eb7-41b7-90f7-08c4b93525ec","Type":"ContainerStarted","Data":"7a2e28f7ee7d517895c84a1fdb6b11d1364cb74ed5a8484dd188be059f519de4"} Dec 01 14:57:34 crc kubenswrapper[4810]: I1201 14:57:34.215580 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:34 crc kubenswrapper[4810]: I1201 14:57:34.253381 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-f27f6" podStartSLOduration=2.253361984 podStartE2EDuration="2.253361984s" podCreationTimestamp="2025-12-01 14:57:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:57:34.24145924 +0000 UTC m=+1420.004968843" watchObservedRunningTime="2025-12-01 14:57:34.253361984 +0000 UTC m=+1420.016871597" Dec 01 14:57:34 crc kubenswrapper[4810]: I1201 14:57:34.281630 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" podStartSLOduration=3.281612071 podStartE2EDuration="3.281612071s" podCreationTimestamp="2025-12-01 14:57:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:57:34.273846979 +0000 UTC m=+1420.037356582" watchObservedRunningTime="2025-12-01 14:57:34.281612071 +0000 UTC m=+1420.045121674" Dec 01 14:57:35 crc kubenswrapper[4810]: I1201 14:57:35.307539 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 14:57:35 crc kubenswrapper[4810]: I1201 14:57:35.314303 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:57:36 crc kubenswrapper[4810]: I1201 14:57:36.418667 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 14:57:37 crc kubenswrapper[4810]: I1201 14:57:37.238580 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"20bf5931-cfe6-4607-89dc-164eef1386a9","Type":"ContainerStarted","Data":"f246c42d90ab5616b2e473b63be7d3941dc56405f046e81b96754d4fc5182551"} Dec 01 14:57:37 crc kubenswrapper[4810]: I1201 14:57:37.238630 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="20bf5931-cfe6-4607-89dc-164eef1386a9" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://f246c42d90ab5616b2e473b63be7d3941dc56405f046e81b96754d4fc5182551" gracePeriod=30 Dec 01 14:57:37 crc kubenswrapper[4810]: I1201 14:57:37.242598 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1a18952e-ee54-4510-b149-e9e5cb873af5","Type":"ContainerStarted","Data":"90196ba74428e571cbc8ee8f9ebb13d1d28fa0a9224cc71182c6382ec51bfcb5"} Dec 01 14:57:37 crc kubenswrapper[4810]: I1201 14:57:37.247946 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"92e83ed1-117e-4772-a9d5-f7888a952207","Type":"ContainerStarted","Data":"d6ccca98c0a412f87b04880e016bc64f47878835f452ea9d761fa6a323b16375"} Dec 01 14:57:37 crc kubenswrapper[4810]: I1201 14:57:37.247997 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="92e83ed1-117e-4772-a9d5-f7888a952207" containerName="nova-metadata-log" containerID="cri-o://4e19c28e3e53471d12fce3344985c52bc2aa5ce57e9fb61552e544c9d3e843a4" gracePeriod=30 Dec 01 14:57:37 crc kubenswrapper[4810]: I1201 14:57:37.248016 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"92e83ed1-117e-4772-a9d5-f7888a952207","Type":"ContainerStarted","Data":"4e19c28e3e53471d12fce3344985c52bc2aa5ce57e9fb61552e544c9d3e843a4"} Dec 01 14:57:37 crc kubenswrapper[4810]: I1201 14:57:37.248028 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="92e83ed1-117e-4772-a9d5-f7888a952207" containerName="nova-metadata-metadata" containerID="cri-o://d6ccca98c0a412f87b04880e016bc64f47878835f452ea9d761fa6a323b16375" gracePeriod=30 Dec 01 14:57:37 crc kubenswrapper[4810]: I1201 14:57:37.264102 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55df027d-d8ed-477f-85e8-9c5548846f2b","Type":"ContainerStarted","Data":"03dd105231934f6f599d549d1c6952103e38d93f40910a0c330ff1bc6b92dee3"} Dec 01 14:57:37 crc kubenswrapper[4810]: I1201 14:57:37.264155 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55df027d-d8ed-477f-85e8-9c5548846f2b","Type":"ContainerStarted","Data":"2c0ceed3036077fcb735cc5df4ddd017139fa9df3cde3ab8a62b54c5fba97c37"} Dec 01 14:57:37 crc kubenswrapper[4810]: I1201 14:57:37.270749 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.9533155730000002 podStartE2EDuration="6.270726204s" podCreationTimestamp="2025-12-01 14:57:31 +0000 UTC" firstStartedPulling="2025-12-01 14:57:32.724048872 +0000 UTC m=+1418.487558475" lastFinishedPulling="2025-12-01 14:57:36.041459503 +0000 UTC m=+1421.804969106" observedRunningTime="2025-12-01 14:57:37.261334828 +0000 UTC m=+1423.024844441" watchObservedRunningTime="2025-12-01 14:57:37.270726204 +0000 UTC m=+1423.034235807" Dec 01 14:57:37 crc kubenswrapper[4810]: I1201 14:57:37.292098 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.711280928 podStartE2EDuration="6.292072963s" podCreationTimestamp="2025-12-01 14:57:31 +0000 UTC" firstStartedPulling="2025-12-01 14:57:32.460666088 +0000 UTC m=+1418.224175691" lastFinishedPulling="2025-12-01 14:57:36.041458123 +0000 UTC m=+1421.804967726" observedRunningTime="2025-12-01 14:57:37.283549822 +0000 UTC m=+1423.047059425" watchObservedRunningTime="2025-12-01 14:57:37.292072963 +0000 UTC m=+1423.055582566" Dec 01 14:57:37 crc kubenswrapper[4810]: I1201 14:57:37.331166 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.850672655 podStartE2EDuration="6.331130194s" podCreationTimestamp="2025-12-01 14:57:31 +0000 UTC" firstStartedPulling="2025-12-01 14:57:32.561058096 +0000 UTC m=+1418.324567699" lastFinishedPulling="2025-12-01 14:57:36.041515635 +0000 UTC m=+1421.805025238" observedRunningTime="2025-12-01 14:57:37.321890903 +0000 UTC m=+1423.085400506" watchObservedRunningTime="2025-12-01 14:57:37.331130194 +0000 UTC m=+1423.094639797" Dec 01 14:57:37 crc kubenswrapper[4810]: I1201 14:57:37.386005 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.794054526 podStartE2EDuration="6.385980424s" podCreationTimestamp="2025-12-01 14:57:31 +0000 UTC" firstStartedPulling="2025-12-01 14:57:32.452498746 +0000 UTC m=+1418.216008349" lastFinishedPulling="2025-12-01 14:57:36.044424634 +0000 UTC m=+1421.807934247" observedRunningTime="2025-12-01 14:57:37.378024477 +0000 UTC m=+1423.141534080" watchObservedRunningTime="2025-12-01 14:57:37.385980424 +0000 UTC m=+1423.149490027" Dec 01 14:57:37 crc kubenswrapper[4810]: I1201 14:57:37.875037 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.021085 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92e83ed1-117e-4772-a9d5-f7888a952207-logs\") pod \"92e83ed1-117e-4772-a9d5-f7888a952207\" (UID: \"92e83ed1-117e-4772-a9d5-f7888a952207\") " Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.021148 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92e83ed1-117e-4772-a9d5-f7888a952207-combined-ca-bundle\") pod \"92e83ed1-117e-4772-a9d5-f7888a952207\" (UID: \"92e83ed1-117e-4772-a9d5-f7888a952207\") " Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.021186 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4hs2\" (UniqueName: \"kubernetes.io/projected/92e83ed1-117e-4772-a9d5-f7888a952207-kube-api-access-l4hs2\") pod \"92e83ed1-117e-4772-a9d5-f7888a952207\" (UID: \"92e83ed1-117e-4772-a9d5-f7888a952207\") " Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.021244 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92e83ed1-117e-4772-a9d5-f7888a952207-config-data\") pod \"92e83ed1-117e-4772-a9d5-f7888a952207\" (UID: \"92e83ed1-117e-4772-a9d5-f7888a952207\") " Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.021640 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92e83ed1-117e-4772-a9d5-f7888a952207-logs" (OuterVolumeSpecName: "logs") pod "92e83ed1-117e-4772-a9d5-f7888a952207" (UID: "92e83ed1-117e-4772-a9d5-f7888a952207"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.021758 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92e83ed1-117e-4772-a9d5-f7888a952207-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.027133 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92e83ed1-117e-4772-a9d5-f7888a952207-kube-api-access-l4hs2" (OuterVolumeSpecName: "kube-api-access-l4hs2") pod "92e83ed1-117e-4772-a9d5-f7888a952207" (UID: "92e83ed1-117e-4772-a9d5-f7888a952207"). InnerVolumeSpecName "kube-api-access-l4hs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.053266 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92e83ed1-117e-4772-a9d5-f7888a952207-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92e83ed1-117e-4772-a9d5-f7888a952207" (UID: "92e83ed1-117e-4772-a9d5-f7888a952207"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.053985 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92e83ed1-117e-4772-a9d5-f7888a952207-config-data" (OuterVolumeSpecName: "config-data") pod "92e83ed1-117e-4772-a9d5-f7888a952207" (UID: "92e83ed1-117e-4772-a9d5-f7888a952207"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.124097 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92e83ed1-117e-4772-a9d5-f7888a952207-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.124139 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4hs2\" (UniqueName: \"kubernetes.io/projected/92e83ed1-117e-4772-a9d5-f7888a952207-kube-api-access-l4hs2\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.124154 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92e83ed1-117e-4772-a9d5-f7888a952207-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.285865 4810 generic.go:334] "Generic (PLEG): container finished" podID="92e83ed1-117e-4772-a9d5-f7888a952207" containerID="d6ccca98c0a412f87b04880e016bc64f47878835f452ea9d761fa6a323b16375" exitCode=0 Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.285898 4810 generic.go:334] "Generic (PLEG): container finished" podID="92e83ed1-117e-4772-a9d5-f7888a952207" containerID="4e19c28e3e53471d12fce3344985c52bc2aa5ce57e9fb61552e544c9d3e843a4" exitCode=143 Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.285916 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.285939 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"92e83ed1-117e-4772-a9d5-f7888a952207","Type":"ContainerDied","Data":"d6ccca98c0a412f87b04880e016bc64f47878835f452ea9d761fa6a323b16375"} Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.286003 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"92e83ed1-117e-4772-a9d5-f7888a952207","Type":"ContainerDied","Data":"4e19c28e3e53471d12fce3344985c52bc2aa5ce57e9fb61552e544c9d3e843a4"} Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.286018 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"92e83ed1-117e-4772-a9d5-f7888a952207","Type":"ContainerDied","Data":"5788c0333bea83a21e3a71e513e16dd8f39a1b0336dc89c18621c8d7a803f01c"} Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.286039 4810 scope.go:117] "RemoveContainer" containerID="d6ccca98c0a412f87b04880e016bc64f47878835f452ea9d761fa6a323b16375" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.316576 4810 scope.go:117] "RemoveContainer" containerID="4e19c28e3e53471d12fce3344985c52bc2aa5ce57e9fb61552e544c9d3e843a4" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.343956 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.359709 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.381672 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:57:38 crc kubenswrapper[4810]: E1201 14:57:38.382172 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92e83ed1-117e-4772-a9d5-f7888a952207" containerName="nova-metadata-log" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.382193 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="92e83ed1-117e-4772-a9d5-f7888a952207" containerName="nova-metadata-log" Dec 01 14:57:38 crc kubenswrapper[4810]: E1201 14:57:38.382207 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92e83ed1-117e-4772-a9d5-f7888a952207" containerName="nova-metadata-metadata" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.382214 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="92e83ed1-117e-4772-a9d5-f7888a952207" containerName="nova-metadata-metadata" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.382422 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="92e83ed1-117e-4772-a9d5-f7888a952207" containerName="nova-metadata-log" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.382454 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="92e83ed1-117e-4772-a9d5-f7888a952207" containerName="nova-metadata-metadata" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.383842 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.386753 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.387099 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.403574 4810 scope.go:117] "RemoveContainer" containerID="d6ccca98c0a412f87b04880e016bc64f47878835f452ea9d761fa6a323b16375" Dec 01 14:57:38 crc kubenswrapper[4810]: E1201 14:57:38.405344 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6ccca98c0a412f87b04880e016bc64f47878835f452ea9d761fa6a323b16375\": container with ID starting with d6ccca98c0a412f87b04880e016bc64f47878835f452ea9d761fa6a323b16375 not found: ID does not exist" containerID="d6ccca98c0a412f87b04880e016bc64f47878835f452ea9d761fa6a323b16375" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.405379 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6ccca98c0a412f87b04880e016bc64f47878835f452ea9d761fa6a323b16375"} err="failed to get container status \"d6ccca98c0a412f87b04880e016bc64f47878835f452ea9d761fa6a323b16375\": rpc error: code = NotFound desc = could not find container \"d6ccca98c0a412f87b04880e016bc64f47878835f452ea9d761fa6a323b16375\": container with ID starting with d6ccca98c0a412f87b04880e016bc64f47878835f452ea9d761fa6a323b16375 not found: ID does not exist" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.405405 4810 scope.go:117] "RemoveContainer" containerID="4e19c28e3e53471d12fce3344985c52bc2aa5ce57e9fb61552e544c9d3e843a4" Dec 01 14:57:38 crc kubenswrapper[4810]: E1201 14:57:38.405762 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e19c28e3e53471d12fce3344985c52bc2aa5ce57e9fb61552e544c9d3e843a4\": container with ID starting with 4e19c28e3e53471d12fce3344985c52bc2aa5ce57e9fb61552e544c9d3e843a4 not found: ID does not exist" containerID="4e19c28e3e53471d12fce3344985c52bc2aa5ce57e9fb61552e544c9d3e843a4" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.405788 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e19c28e3e53471d12fce3344985c52bc2aa5ce57e9fb61552e544c9d3e843a4"} err="failed to get container status \"4e19c28e3e53471d12fce3344985c52bc2aa5ce57e9fb61552e544c9d3e843a4\": rpc error: code = NotFound desc = could not find container \"4e19c28e3e53471d12fce3344985c52bc2aa5ce57e9fb61552e544c9d3e843a4\": container with ID starting with 4e19c28e3e53471d12fce3344985c52bc2aa5ce57e9fb61552e544c9d3e843a4 not found: ID does not exist" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.405806 4810 scope.go:117] "RemoveContainer" containerID="d6ccca98c0a412f87b04880e016bc64f47878835f452ea9d761fa6a323b16375" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.406046 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6ccca98c0a412f87b04880e016bc64f47878835f452ea9d761fa6a323b16375"} err="failed to get container status \"d6ccca98c0a412f87b04880e016bc64f47878835f452ea9d761fa6a323b16375\": rpc error: code = NotFound desc = could not find container \"d6ccca98c0a412f87b04880e016bc64f47878835f452ea9d761fa6a323b16375\": container with ID starting with d6ccca98c0a412f87b04880e016bc64f47878835f452ea9d761fa6a323b16375 not found: ID does not exist" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.406085 4810 scope.go:117] "RemoveContainer" containerID="4e19c28e3e53471d12fce3344985c52bc2aa5ce57e9fb61552e544c9d3e843a4" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.406300 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e19c28e3e53471d12fce3344985c52bc2aa5ce57e9fb61552e544c9d3e843a4"} err="failed to get container status \"4e19c28e3e53471d12fce3344985c52bc2aa5ce57e9fb61552e544c9d3e843a4\": rpc error: code = NotFound desc = could not find container \"4e19c28e3e53471d12fce3344985c52bc2aa5ce57e9fb61552e544c9d3e843a4\": container with ID starting with 4e19c28e3e53471d12fce3344985c52bc2aa5ce57e9fb61552e544c9d3e843a4 not found: ID does not exist" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.411839 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.502656 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92e83ed1-117e-4772-a9d5-f7888a952207" path="/var/lib/kubelet/pods/92e83ed1-117e-4772-a9d5-f7888a952207/volumes" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.531724 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdmzn\" (UniqueName: \"kubernetes.io/projected/a6abe987-223f-4d84-ae4d-8c89282f4b18-kube-api-access-gdmzn\") pod \"nova-metadata-0\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " pod="openstack/nova-metadata-0" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.532140 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6abe987-223f-4d84-ae4d-8c89282f4b18-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " pod="openstack/nova-metadata-0" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.532183 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6abe987-223f-4d84-ae4d-8c89282f4b18-config-data\") pod \"nova-metadata-0\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " pod="openstack/nova-metadata-0" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.532208 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6abe987-223f-4d84-ae4d-8c89282f4b18-logs\") pod \"nova-metadata-0\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " pod="openstack/nova-metadata-0" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.532264 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6abe987-223f-4d84-ae4d-8c89282f4b18-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " pod="openstack/nova-metadata-0" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.634340 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdmzn\" (UniqueName: \"kubernetes.io/projected/a6abe987-223f-4d84-ae4d-8c89282f4b18-kube-api-access-gdmzn\") pod \"nova-metadata-0\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " pod="openstack/nova-metadata-0" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.634385 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6abe987-223f-4d84-ae4d-8c89282f4b18-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " pod="openstack/nova-metadata-0" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.634449 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6abe987-223f-4d84-ae4d-8c89282f4b18-config-data\") pod \"nova-metadata-0\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " pod="openstack/nova-metadata-0" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.634489 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6abe987-223f-4d84-ae4d-8c89282f4b18-logs\") pod \"nova-metadata-0\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " pod="openstack/nova-metadata-0" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.635626 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6abe987-223f-4d84-ae4d-8c89282f4b18-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " pod="openstack/nova-metadata-0" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.636483 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6abe987-223f-4d84-ae4d-8c89282f4b18-logs\") pod \"nova-metadata-0\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " pod="openstack/nova-metadata-0" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.639422 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6abe987-223f-4d84-ae4d-8c89282f4b18-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " pod="openstack/nova-metadata-0" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.639662 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6abe987-223f-4d84-ae4d-8c89282f4b18-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " pod="openstack/nova-metadata-0" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.647056 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6abe987-223f-4d84-ae4d-8c89282f4b18-config-data\") pod \"nova-metadata-0\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " pod="openstack/nova-metadata-0" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.671682 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdmzn\" (UniqueName: \"kubernetes.io/projected/a6abe987-223f-4d84-ae4d-8c89282f4b18-kube-api-access-gdmzn\") pod \"nova-metadata-0\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " pod="openstack/nova-metadata-0" Dec 01 14:57:38 crc kubenswrapper[4810]: I1201 14:57:38.711209 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 14:57:39 crc kubenswrapper[4810]: I1201 14:57:39.195723 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:57:39 crc kubenswrapper[4810]: I1201 14:57:39.295872 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a6abe987-223f-4d84-ae4d-8c89282f4b18","Type":"ContainerStarted","Data":"bcbc0556f7cbc8370f4f1e7fcc83405f3ccbfcfb7925fc587e9176946fb48843"} Dec 01 14:57:40 crc kubenswrapper[4810]: I1201 14:57:40.306252 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a6abe987-223f-4d84-ae4d-8c89282f4b18","Type":"ContainerStarted","Data":"ad22eaa19e88e697a80abeb04365ae185673de285c03735e4c2038238c47ef75"} Dec 01 14:57:40 crc kubenswrapper[4810]: I1201 14:57:40.306548 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a6abe987-223f-4d84-ae4d-8c89282f4b18","Type":"ContainerStarted","Data":"d2563a8c77321339ba89e24bec49f1370326419ad7015bd8fbd386dd91563a24"} Dec 01 14:57:40 crc kubenswrapper[4810]: I1201 14:57:40.330488 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.330456944 podStartE2EDuration="2.330456944s" podCreationTimestamp="2025-12-01 14:57:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:57:40.326123967 +0000 UTC m=+1426.089633570" watchObservedRunningTime="2025-12-01 14:57:40.330456944 +0000 UTC m=+1426.093966547" Dec 01 14:57:40 crc kubenswrapper[4810]: I1201 14:57:40.944692 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 14:57:40 crc kubenswrapper[4810]: I1201 14:57:40.945093 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="563bcc28-df47-4e24-a246-9b713ad5cbb3" containerName="kube-state-metrics" containerID="cri-o://c8a73c8466b9b60faf6e449bb4dd1db5af7054b925a232c56b49850420483b2a" gracePeriod=30 Dec 01 14:57:41 crc kubenswrapper[4810]: I1201 14:57:41.321204 4810 generic.go:334] "Generic (PLEG): container finished" podID="563bcc28-df47-4e24-a246-9b713ad5cbb3" containerID="c8a73c8466b9b60faf6e449bb4dd1db5af7054b925a232c56b49850420483b2a" exitCode=2 Dec 01 14:57:41 crc kubenswrapper[4810]: I1201 14:57:41.322545 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"563bcc28-df47-4e24-a246-9b713ad5cbb3","Type":"ContainerDied","Data":"c8a73c8466b9b60faf6e449bb4dd1db5af7054b925a232c56b49850420483b2a"} Dec 01 14:57:41 crc kubenswrapper[4810]: I1201 14:57:41.470985 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 14:57:41 crc kubenswrapper[4810]: I1201 14:57:41.581828 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 14:57:41 crc kubenswrapper[4810]: I1201 14:57:41.581872 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 14:57:41 crc kubenswrapper[4810]: I1201 14:57:41.594693 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxgq5\" (UniqueName: \"kubernetes.io/projected/563bcc28-df47-4e24-a246-9b713ad5cbb3-kube-api-access-hxgq5\") pod \"563bcc28-df47-4e24-a246-9b713ad5cbb3\" (UID: \"563bcc28-df47-4e24-a246-9b713ad5cbb3\") " Dec 01 14:57:41 crc kubenswrapper[4810]: I1201 14:57:41.616058 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/563bcc28-df47-4e24-a246-9b713ad5cbb3-kube-api-access-hxgq5" (OuterVolumeSpecName: "kube-api-access-hxgq5") pod "563bcc28-df47-4e24-a246-9b713ad5cbb3" (UID: "563bcc28-df47-4e24-a246-9b713ad5cbb3"). InnerVolumeSpecName "kube-api-access-hxgq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:57:41 crc kubenswrapper[4810]: I1201 14:57:41.697273 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxgq5\" (UniqueName: \"kubernetes.io/projected/563bcc28-df47-4e24-a246-9b713ad5cbb3-kube-api-access-hxgq5\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:41 crc kubenswrapper[4810]: I1201 14:57:41.866997 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 14:57:41 crc kubenswrapper[4810]: I1201 14:57:41.867105 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 14:57:41 crc kubenswrapper[4810]: I1201 14:57:41.896070 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 14:57:41 crc kubenswrapper[4810]: I1201 14:57:41.925603 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:57:41 crc kubenswrapper[4810]: I1201 14:57:41.961727 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.048987 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8ccb5c7cf-r4dnc"] Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.049245 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" podUID="4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed" containerName="dnsmasq-dns" containerID="cri-o://fe264e2791d0c18cc3286e99e91fe1842646747df8709e9962de684167859864" gracePeriod=10 Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.420127 4810 generic.go:334] "Generic (PLEG): container finished" podID="fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88" containerID="9a9cb3e6a77baeced59069d7dc5d260e53122dd4e38d62c9a1293deed78602d5" exitCode=0 Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.420465 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-f27f6" event={"ID":"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88","Type":"ContainerDied","Data":"9a9cb3e6a77baeced59069d7dc5d260e53122dd4e38d62c9a1293deed78602d5"} Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.447099 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"563bcc28-df47-4e24-a246-9b713ad5cbb3","Type":"ContainerDied","Data":"f3c2704bbef3d9d23a3b4ae03adb993df9e67183586c306521f4db470f8720c9"} Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.447147 4810 scope.go:117] "RemoveContainer" containerID="c8a73c8466b9b60faf6e449bb4dd1db5af7054b925a232c56b49850420483b2a" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.447261 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.466003 4810 generic.go:334] "Generic (PLEG): container finished" podID="8bbc4102-9734-42a7-8bd5-9487e83d768b" containerID="bcb6d4eef8012d35e7a2581c37d3de047e5d6c95182ecb8fe94d92a03fdbccd4" exitCode=0 Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.466066 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-78w6c" event={"ID":"8bbc4102-9734-42a7-8bd5-9487e83d768b","Type":"ContainerDied","Data":"bcb6d4eef8012d35e7a2581c37d3de047e5d6c95182ecb8fe94d92a03fdbccd4"} Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.495873 4810 generic.go:334] "Generic (PLEG): container finished" podID="4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed" containerID="fe264e2791d0c18cc3286e99e91fe1842646747df8709e9962de684167859864" exitCode=0 Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.534980 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" event={"ID":"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed","Type":"ContainerDied","Data":"fe264e2791d0c18cc3286e99e91fe1842646747df8709e9962de684167859864"} Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.559893 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.602289 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.614127 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 14:57:42 crc kubenswrapper[4810]: E1201 14:57:42.614587 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="563bcc28-df47-4e24-a246-9b713ad5cbb3" containerName="kube-state-metrics" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.614605 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="563bcc28-df47-4e24-a246-9b713ad5cbb3" containerName="kube-state-metrics" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.614763 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="563bcc28-df47-4e24-a246-9b713ad5cbb3" containerName="kube-state-metrics" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.615376 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.616003 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.621635 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.621824 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.622911 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="55df027d-d8ed-477f-85e8-9c5548846f2b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.180:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.623523 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.664305 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="55df027d-d8ed-477f-85e8-9c5548846f2b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.180:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.703226 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.732809 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9sct\" (UniqueName: \"kubernetes.io/projected/39ce8411-05ee-4022-bd7a-87e46c2cee8d-kube-api-access-h9sct\") pod \"kube-state-metrics-0\" (UID: \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\") " pod="openstack/kube-state-metrics-0" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.732862 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/39ce8411-05ee-4022-bd7a-87e46c2cee8d-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\") " pod="openstack/kube-state-metrics-0" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.732908 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/39ce8411-05ee-4022-bd7a-87e46c2cee8d-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\") " pod="openstack/kube-state-metrics-0" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.732934 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39ce8411-05ee-4022-bd7a-87e46c2cee8d-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\") " pod="openstack/kube-state-metrics-0" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.834418 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-dns-swift-storage-0\") pod \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.835026 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-config\") pod \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.835066 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-ovsdbserver-sb\") pod \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.835099 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-dns-svc\") pod \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.835201 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-ovsdbserver-nb\") pod \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.835238 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zh8rk\" (UniqueName: \"kubernetes.io/projected/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-kube-api-access-zh8rk\") pod \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\" (UID: \"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed\") " Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.835568 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9sct\" (UniqueName: \"kubernetes.io/projected/39ce8411-05ee-4022-bd7a-87e46c2cee8d-kube-api-access-h9sct\") pod \"kube-state-metrics-0\" (UID: \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\") " pod="openstack/kube-state-metrics-0" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.835618 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/39ce8411-05ee-4022-bd7a-87e46c2cee8d-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\") " pod="openstack/kube-state-metrics-0" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.835677 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/39ce8411-05ee-4022-bd7a-87e46c2cee8d-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\") " pod="openstack/kube-state-metrics-0" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.835711 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39ce8411-05ee-4022-bd7a-87e46c2cee8d-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\") " pod="openstack/kube-state-metrics-0" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.843645 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-kube-api-access-zh8rk" (OuterVolumeSpecName: "kube-api-access-zh8rk") pod "4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed" (UID: "4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed"). InnerVolumeSpecName "kube-api-access-zh8rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.844703 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39ce8411-05ee-4022-bd7a-87e46c2cee8d-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\") " pod="openstack/kube-state-metrics-0" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.845143 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/39ce8411-05ee-4022-bd7a-87e46c2cee8d-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\") " pod="openstack/kube-state-metrics-0" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.845358 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/39ce8411-05ee-4022-bd7a-87e46c2cee8d-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\") " pod="openstack/kube-state-metrics-0" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.852530 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9sct\" (UniqueName: \"kubernetes.io/projected/39ce8411-05ee-4022-bd7a-87e46c2cee8d-kube-api-access-h9sct\") pod \"kube-state-metrics-0\" (UID: \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\") " pod="openstack/kube-state-metrics-0" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.903887 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed" (UID: "4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.925694 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-config" (OuterVolumeSpecName: "config") pod "4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed" (UID: "4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.938065 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.938096 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.938106 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zh8rk\" (UniqueName: \"kubernetes.io/projected/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-kube-api-access-zh8rk\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.940910 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed" (UID: "4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.940425 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed" (UID: "4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:57:42 crc kubenswrapper[4810]: I1201 14:57:42.948806 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed" (UID: "4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.016261 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.039606 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.039842 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.039919 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.498952 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.502853 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerName="ceilometer-central-agent" containerID="cri-o://f741f8a0ce24ec2347b6ff71f1e6d56ee41b153add26bf402baca5a5b8eff0c1" gracePeriod=30 Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.503674 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerName="proxy-httpd" containerID="cri-o://0d8769d7c7d03b825251d3d117c83522bf3d8aa26d1c14009417406e1c583738" gracePeriod=30 Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.503765 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerName="sg-core" containerID="cri-o://95ea32388466bbc69b18495263c46fffd9ccc7121f3e7e29fe8e9c0b05de57e4" gracePeriod=30 Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.503847 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerName="ceilometer-notification-agent" containerID="cri-o://0828594d163132cb7db5ec6c8956f48822a6bab045cc6ab9d7c946f86312ac7f" gracePeriod=30 Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.532013 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" event={"ID":"4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed","Type":"ContainerDied","Data":"317bfd153a31d9dcdad645a673394f82bb72293881221f6bce7bc109f6db70fc"} Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.532072 4810 scope.go:117] "RemoveContainer" containerID="fe264e2791d0c18cc3286e99e91fe1842646747df8709e9962de684167859864" Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.532259 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8ccb5c7cf-r4dnc" Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.579220 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.607952 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8ccb5c7cf-r4dnc"] Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.610622 4810 scope.go:117] "RemoveContainer" containerID="0ba04bb2acced6ad016122958c11c701353a9196ca7ece5be1e1a74b3b65fc40" Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.616069 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8ccb5c7cf-r4dnc"] Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.712100 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.712963 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 14:57:43 crc kubenswrapper[4810]: I1201 14:57:43.987048 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-78w6c" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.049019 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-f27f6" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.075514 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs2gw\" (UniqueName: \"kubernetes.io/projected/8bbc4102-9734-42a7-8bd5-9487e83d768b-kube-api-access-rs2gw\") pod \"8bbc4102-9734-42a7-8bd5-9487e83d768b\" (UID: \"8bbc4102-9734-42a7-8bd5-9487e83d768b\") " Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.075588 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bbc4102-9734-42a7-8bd5-9487e83d768b-config-data\") pod \"8bbc4102-9734-42a7-8bd5-9487e83d768b\" (UID: \"8bbc4102-9734-42a7-8bd5-9487e83d768b\") " Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.075712 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bbc4102-9734-42a7-8bd5-9487e83d768b-combined-ca-bundle\") pod \"8bbc4102-9734-42a7-8bd5-9487e83d768b\" (UID: \"8bbc4102-9734-42a7-8bd5-9487e83d768b\") " Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.075754 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bbc4102-9734-42a7-8bd5-9487e83d768b-scripts\") pod \"8bbc4102-9734-42a7-8bd5-9487e83d768b\" (UID: \"8bbc4102-9734-42a7-8bd5-9487e83d768b\") " Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.087588 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bbc4102-9734-42a7-8bd5-9487e83d768b-scripts" (OuterVolumeSpecName: "scripts") pod "8bbc4102-9734-42a7-8bd5-9487e83d768b" (UID: "8bbc4102-9734-42a7-8bd5-9487e83d768b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.087912 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bbc4102-9734-42a7-8bd5-9487e83d768b-kube-api-access-rs2gw" (OuterVolumeSpecName: "kube-api-access-rs2gw") pod "8bbc4102-9734-42a7-8bd5-9487e83d768b" (UID: "8bbc4102-9734-42a7-8bd5-9487e83d768b"). InnerVolumeSpecName "kube-api-access-rs2gw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.126650 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bbc4102-9734-42a7-8bd5-9487e83d768b-config-data" (OuterVolumeSpecName: "config-data") pod "8bbc4102-9734-42a7-8bd5-9487e83d768b" (UID: "8bbc4102-9734-42a7-8bd5-9487e83d768b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.127264 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bbc4102-9734-42a7-8bd5-9487e83d768b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8bbc4102-9734-42a7-8bd5-9487e83d768b" (UID: "8bbc4102-9734-42a7-8bd5-9487e83d768b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.177569 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmsfs\" (UniqueName: \"kubernetes.io/projected/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-kube-api-access-kmsfs\") pod \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\" (UID: \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\") " Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.177613 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-scripts\") pod \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\" (UID: \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\") " Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.177694 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-config-data\") pod \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\" (UID: \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\") " Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.177868 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-combined-ca-bundle\") pod \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\" (UID: \"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88\") " Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.178376 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs2gw\" (UniqueName: \"kubernetes.io/projected/8bbc4102-9734-42a7-8bd5-9487e83d768b-kube-api-access-rs2gw\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.178393 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bbc4102-9734-42a7-8bd5-9487e83d768b-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.178404 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bbc4102-9734-42a7-8bd5-9487e83d768b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.178414 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bbc4102-9734-42a7-8bd5-9487e83d768b-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.181685 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-scripts" (OuterVolumeSpecName: "scripts") pod "fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88" (UID: "fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.183142 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-kube-api-access-kmsfs" (OuterVolumeSpecName: "kube-api-access-kmsfs") pod "fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88" (UID: "fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88"). InnerVolumeSpecName "kube-api-access-kmsfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.205430 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88" (UID: "fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.230553 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-config-data" (OuterVolumeSpecName: "config-data") pod "fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88" (UID: "fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.281193 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.281443 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmsfs\" (UniqueName: \"kubernetes.io/projected/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-kube-api-access-kmsfs\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.281543 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.281602 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.510674 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed" path="/var/lib/kubelet/pods/4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed/volumes" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.512229 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="563bcc28-df47-4e24-a246-9b713ad5cbb3" path="/var/lib/kubelet/pods/563bcc28-df47-4e24-a246-9b713ad5cbb3/volumes" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.567742 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"39ce8411-05ee-4022-bd7a-87e46c2cee8d","Type":"ContainerStarted","Data":"030a775eb17300c273b82602d862651690419af2f0861ccfe4b4d6bdc64e3a5e"} Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.567801 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"39ce8411-05ee-4022-bd7a-87e46c2cee8d","Type":"ContainerStarted","Data":"5b4a8d4df900f9b4fede27035950a845b76dd2849a241b354a524387ce483969"} Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.569204 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.578730 4810 generic.go:334] "Generic (PLEG): container finished" podID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerID="0d8769d7c7d03b825251d3d117c83522bf3d8aa26d1c14009417406e1c583738" exitCode=0 Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.578766 4810 generic.go:334] "Generic (PLEG): container finished" podID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerID="95ea32388466bbc69b18495263c46fffd9ccc7121f3e7e29fe8e9c0b05de57e4" exitCode=2 Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.578776 4810 generic.go:334] "Generic (PLEG): container finished" podID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerID="f741f8a0ce24ec2347b6ff71f1e6d56ee41b153add26bf402baca5a5b8eff0c1" exitCode=0 Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.578858 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba866fb-11bf-4edf-a0a9-1e7534cae07f","Type":"ContainerDied","Data":"0d8769d7c7d03b825251d3d117c83522bf3d8aa26d1c14009417406e1c583738"} Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.578883 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba866fb-11bf-4edf-a0a9-1e7534cae07f","Type":"ContainerDied","Data":"95ea32388466bbc69b18495263c46fffd9ccc7121f3e7e29fe8e9c0b05de57e4"} Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.578895 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba866fb-11bf-4edf-a0a9-1e7534cae07f","Type":"ContainerDied","Data":"f741f8a0ce24ec2347b6ff71f1e6d56ee41b153add26bf402baca5a5b8eff0c1"} Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.586908 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-f27f6" event={"ID":"fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88","Type":"ContainerDied","Data":"c5fc07d8e9dbe9a76f702de2bbbd99d9d9ab8da5d433469d5cac9d4c92448907"} Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.588938 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5fc07d8e9dbe9a76f702de2bbbd99d9d9ab8da5d433469d5cac9d4c92448907" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.587034 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-f27f6" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.599563 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-78w6c" event={"ID":"8bbc4102-9734-42a7-8bd5-9487e83d768b","Type":"ContainerDied","Data":"b4f6b4670e37a86fc2f55dbfdc350f5c2b8a2b95244231daf8c5e0ba2da33ce9"} Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.601549 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4f6b4670e37a86fc2f55dbfdc350f5c2b8a2b95244231daf8c5e0ba2da33ce9" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.599698 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-78w6c" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.610570 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 14:57:44 crc kubenswrapper[4810]: E1201 14:57:44.611060 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88" containerName="nova-cell1-conductor-db-sync" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.611083 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88" containerName="nova-cell1-conductor-db-sync" Dec 01 14:57:44 crc kubenswrapper[4810]: E1201 14:57:44.611111 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed" containerName="dnsmasq-dns" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.611119 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed" containerName="dnsmasq-dns" Dec 01 14:57:44 crc kubenswrapper[4810]: E1201 14:57:44.611159 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bbc4102-9734-42a7-8bd5-9487e83d768b" containerName="nova-manage" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.611168 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bbc4102-9734-42a7-8bd5-9487e83d768b" containerName="nova-manage" Dec 01 14:57:44 crc kubenswrapper[4810]: E1201 14:57:44.611180 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed" containerName="init" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.611187 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed" containerName="init" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.611415 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bbc4102-9734-42a7-8bd5-9487e83d768b" containerName="nova-manage" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.611451 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88" containerName="nova-cell1-conductor-db-sync" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.611485 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ef69bb1-54d2-47b5-bcbd-5d0f3b60d2ed" containerName="dnsmasq-dns" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.612220 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.620679 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.622001 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.624176 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.258667721 podStartE2EDuration="2.624155688s" podCreationTimestamp="2025-12-01 14:57:42 +0000 UTC" firstStartedPulling="2025-12-01 14:57:43.560707343 +0000 UTC m=+1429.324216946" lastFinishedPulling="2025-12-01 14:57:43.92619529 +0000 UTC m=+1429.689704913" observedRunningTime="2025-12-01 14:57:44.599638831 +0000 UTC m=+1430.363148434" watchObservedRunningTime="2025-12-01 14:57:44.624155688 +0000 UTC m=+1430.387665291" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.690959 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c\") " pod="openstack/nova-cell1-conductor-0" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.691068 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c\") " pod="openstack/nova-cell1-conductor-0" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.691104 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxs28\" (UniqueName: \"kubernetes.io/projected/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c-kube-api-access-dxs28\") pod \"nova-cell1-conductor-0\" (UID: \"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c\") " pod="openstack/nova-cell1-conductor-0" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.697313 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.697667 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="55df027d-d8ed-477f-85e8-9c5548846f2b" containerName="nova-api-log" containerID="cri-o://2c0ceed3036077fcb735cc5df4ddd017139fa9df3cde3ab8a62b54c5fba97c37" gracePeriod=30 Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.697788 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="55df027d-d8ed-477f-85e8-9c5548846f2b" containerName="nova-api-api" containerID="cri-o://03dd105231934f6f599d549d1c6952103e38d93f40910a0c330ff1bc6b92dee3" gracePeriod=30 Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.717598 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.788825 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.793142 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c\") " pod="openstack/nova-cell1-conductor-0" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.793196 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxs28\" (UniqueName: \"kubernetes.io/projected/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c-kube-api-access-dxs28\") pod \"nova-cell1-conductor-0\" (UID: \"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c\") " pod="openstack/nova-cell1-conductor-0" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.793323 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c\") " pod="openstack/nova-cell1-conductor-0" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.804304 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c\") " pod="openstack/nova-cell1-conductor-0" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.804346 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c\") " pod="openstack/nova-cell1-conductor-0" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.815906 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxs28\" (UniqueName: \"kubernetes.io/projected/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c-kube-api-access-dxs28\") pod \"nova-cell1-conductor-0\" (UID: \"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c\") " pod="openstack/nova-cell1-conductor-0" Dec 01 14:57:44 crc kubenswrapper[4810]: I1201 14:57:44.936831 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.454438 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.510710 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-config-data\") pod \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.510986 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-combined-ca-bundle\") pod \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.511038 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-sg-core-conf-yaml\") pod \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.511300 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-log-httpd\") pod \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.511330 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-scripts\") pod \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.511408 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nl2zj\" (UniqueName: \"kubernetes.io/projected/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-kube-api-access-nl2zj\") pod \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.511454 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-run-httpd\") pod \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\" (UID: \"4ba866fb-11bf-4edf-a0a9-1e7534cae07f\") " Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.526525 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4ba866fb-11bf-4edf-a0a9-1e7534cae07f" (UID: "4ba866fb-11bf-4edf-a0a9-1e7534cae07f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.529076 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4ba866fb-11bf-4edf-a0a9-1e7534cae07f" (UID: "4ba866fb-11bf-4edf-a0a9-1e7534cae07f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.531925 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-scripts" (OuterVolumeSpecName: "scripts") pod "4ba866fb-11bf-4edf-a0a9-1e7534cae07f" (UID: "4ba866fb-11bf-4edf-a0a9-1e7534cae07f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.537657 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-kube-api-access-nl2zj" (OuterVolumeSpecName: "kube-api-access-nl2zj") pod "4ba866fb-11bf-4edf-a0a9-1e7534cae07f" (UID: "4ba866fb-11bf-4edf-a0a9-1e7534cae07f"). InnerVolumeSpecName "kube-api-access-nl2zj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.620028 4810 generic.go:334] "Generic (PLEG): container finished" podID="55df027d-d8ed-477f-85e8-9c5548846f2b" containerID="2c0ceed3036077fcb735cc5df4ddd017139fa9df3cde3ab8a62b54c5fba97c37" exitCode=143 Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.620185 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55df027d-d8ed-477f-85e8-9c5548846f2b","Type":"ContainerDied","Data":"2c0ceed3036077fcb735cc5df4ddd017139fa9df3cde3ab8a62b54c5fba97c37"} Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.620667 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.620699 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nl2zj\" (UniqueName: \"kubernetes.io/projected/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-kube-api-access-nl2zj\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.620735 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.620767 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.625639 4810 generic.go:334] "Generic (PLEG): container finished" podID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerID="0828594d163132cb7db5ec6c8956f48822a6bab045cc6ab9d7c946f86312ac7f" exitCode=0 Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.625884 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="1a18952e-ee54-4510-b149-e9e5cb873af5" containerName="nova-scheduler-scheduler" containerID="cri-o://90196ba74428e571cbc8ee8f9ebb13d1d28fa0a9224cc71182c6382ec51bfcb5" gracePeriod=30 Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.625981 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.626443 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba866fb-11bf-4edf-a0a9-1e7534cae07f","Type":"ContainerDied","Data":"0828594d163132cb7db5ec6c8956f48822a6bab045cc6ab9d7c946f86312ac7f"} Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.626480 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba866fb-11bf-4edf-a0a9-1e7534cae07f","Type":"ContainerDied","Data":"0e208fba9fb7dcfaa22465c42f0df0caa69e2703c19481e23e4ca76e8263d78c"} Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.626497 4810 scope.go:117] "RemoveContainer" containerID="0d8769d7c7d03b825251d3d117c83522bf3d8aa26d1c14009417406e1c583738" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.626861 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a6abe987-223f-4d84-ae4d-8c89282f4b18" containerName="nova-metadata-log" containerID="cri-o://d2563a8c77321339ba89e24bec49f1370326419ad7015bd8fbd386dd91563a24" gracePeriod=30 Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.627686 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a6abe987-223f-4d84-ae4d-8c89282f4b18" containerName="nova-metadata-metadata" containerID="cri-o://ad22eaa19e88e697a80abeb04365ae185673de285c03735e4c2038238c47ef75" gracePeriod=30 Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.637780 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.653796 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4ba866fb-11bf-4edf-a0a9-1e7534cae07f" (UID: "4ba866fb-11bf-4edf-a0a9-1e7534cae07f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.663123 4810 scope.go:117] "RemoveContainer" containerID="95ea32388466bbc69b18495263c46fffd9ccc7121f3e7e29fe8e9c0b05de57e4" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.681661 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ba866fb-11bf-4edf-a0a9-1e7534cae07f" (UID: "4ba866fb-11bf-4edf-a0a9-1e7534cae07f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.689426 4810 scope.go:117] "RemoveContainer" containerID="0828594d163132cb7db5ec6c8956f48822a6bab045cc6ab9d7c946f86312ac7f" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.722797 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.722846 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.728939 4810 scope.go:117] "RemoveContainer" containerID="f741f8a0ce24ec2347b6ff71f1e6d56ee41b153add26bf402baca5a5b8eff0c1" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.741904 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-config-data" (OuterVolumeSpecName: "config-data") pod "4ba866fb-11bf-4edf-a0a9-1e7534cae07f" (UID: "4ba866fb-11bf-4edf-a0a9-1e7534cae07f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.766111 4810 scope.go:117] "RemoveContainer" containerID="0d8769d7c7d03b825251d3d117c83522bf3d8aa26d1c14009417406e1c583738" Dec 01 14:57:45 crc kubenswrapper[4810]: E1201 14:57:45.767371 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d8769d7c7d03b825251d3d117c83522bf3d8aa26d1c14009417406e1c583738\": container with ID starting with 0d8769d7c7d03b825251d3d117c83522bf3d8aa26d1c14009417406e1c583738 not found: ID does not exist" containerID="0d8769d7c7d03b825251d3d117c83522bf3d8aa26d1c14009417406e1c583738" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.767419 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d8769d7c7d03b825251d3d117c83522bf3d8aa26d1c14009417406e1c583738"} err="failed to get container status \"0d8769d7c7d03b825251d3d117c83522bf3d8aa26d1c14009417406e1c583738\": rpc error: code = NotFound desc = could not find container \"0d8769d7c7d03b825251d3d117c83522bf3d8aa26d1c14009417406e1c583738\": container with ID starting with 0d8769d7c7d03b825251d3d117c83522bf3d8aa26d1c14009417406e1c583738 not found: ID does not exist" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.767448 4810 scope.go:117] "RemoveContainer" containerID="95ea32388466bbc69b18495263c46fffd9ccc7121f3e7e29fe8e9c0b05de57e4" Dec 01 14:57:45 crc kubenswrapper[4810]: E1201 14:57:45.767801 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95ea32388466bbc69b18495263c46fffd9ccc7121f3e7e29fe8e9c0b05de57e4\": container with ID starting with 95ea32388466bbc69b18495263c46fffd9ccc7121f3e7e29fe8e9c0b05de57e4 not found: ID does not exist" containerID="95ea32388466bbc69b18495263c46fffd9ccc7121f3e7e29fe8e9c0b05de57e4" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.767825 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95ea32388466bbc69b18495263c46fffd9ccc7121f3e7e29fe8e9c0b05de57e4"} err="failed to get container status \"95ea32388466bbc69b18495263c46fffd9ccc7121f3e7e29fe8e9c0b05de57e4\": rpc error: code = NotFound desc = could not find container \"95ea32388466bbc69b18495263c46fffd9ccc7121f3e7e29fe8e9c0b05de57e4\": container with ID starting with 95ea32388466bbc69b18495263c46fffd9ccc7121f3e7e29fe8e9c0b05de57e4 not found: ID does not exist" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.767840 4810 scope.go:117] "RemoveContainer" containerID="0828594d163132cb7db5ec6c8956f48822a6bab045cc6ab9d7c946f86312ac7f" Dec 01 14:57:45 crc kubenswrapper[4810]: E1201 14:57:45.768116 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0828594d163132cb7db5ec6c8956f48822a6bab045cc6ab9d7c946f86312ac7f\": container with ID starting with 0828594d163132cb7db5ec6c8956f48822a6bab045cc6ab9d7c946f86312ac7f not found: ID does not exist" containerID="0828594d163132cb7db5ec6c8956f48822a6bab045cc6ab9d7c946f86312ac7f" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.768160 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0828594d163132cb7db5ec6c8956f48822a6bab045cc6ab9d7c946f86312ac7f"} err="failed to get container status \"0828594d163132cb7db5ec6c8956f48822a6bab045cc6ab9d7c946f86312ac7f\": rpc error: code = NotFound desc = could not find container \"0828594d163132cb7db5ec6c8956f48822a6bab045cc6ab9d7c946f86312ac7f\": container with ID starting with 0828594d163132cb7db5ec6c8956f48822a6bab045cc6ab9d7c946f86312ac7f not found: ID does not exist" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.768189 4810 scope.go:117] "RemoveContainer" containerID="f741f8a0ce24ec2347b6ff71f1e6d56ee41b153add26bf402baca5a5b8eff0c1" Dec 01 14:57:45 crc kubenswrapper[4810]: E1201 14:57:45.768464 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f741f8a0ce24ec2347b6ff71f1e6d56ee41b153add26bf402baca5a5b8eff0c1\": container with ID starting with f741f8a0ce24ec2347b6ff71f1e6d56ee41b153add26bf402baca5a5b8eff0c1 not found: ID does not exist" containerID="f741f8a0ce24ec2347b6ff71f1e6d56ee41b153add26bf402baca5a5b8eff0c1" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.768590 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f741f8a0ce24ec2347b6ff71f1e6d56ee41b153add26bf402baca5a5b8eff0c1"} err="failed to get container status \"f741f8a0ce24ec2347b6ff71f1e6d56ee41b153add26bf402baca5a5b8eff0c1\": rpc error: code = NotFound desc = could not find container \"f741f8a0ce24ec2347b6ff71f1e6d56ee41b153add26bf402baca5a5b8eff0c1\": container with ID starting with f741f8a0ce24ec2347b6ff71f1e6d56ee41b153add26bf402baca5a5b8eff0c1 not found: ID does not exist" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.824860 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ba866fb-11bf-4edf-a0a9-1e7534cae07f-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:45 crc kubenswrapper[4810]: I1201 14:57:45.993864 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.006132 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.019562 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:57:46 crc kubenswrapper[4810]: E1201 14:57:46.022759 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerName="ceilometer-central-agent" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.022791 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerName="ceilometer-central-agent" Dec 01 14:57:46 crc kubenswrapper[4810]: E1201 14:57:46.022824 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerName="ceilometer-notification-agent" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.022832 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerName="ceilometer-notification-agent" Dec 01 14:57:46 crc kubenswrapper[4810]: E1201 14:57:46.022840 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerName="proxy-httpd" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.022847 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerName="proxy-httpd" Dec 01 14:57:46 crc kubenswrapper[4810]: E1201 14:57:46.022869 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerName="sg-core" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.022910 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerName="sg-core" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.023339 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerName="ceilometer-central-agent" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.023354 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerName="ceilometer-notification-agent" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.023372 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerName="sg-core" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.023389 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" containerName="proxy-httpd" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.025493 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.028574 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.030799 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.030977 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.031108 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.132462 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/baf9222d-90d1-4719-9724-2fcc56f8a274-log-httpd\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.132583 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.132614 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/baf9222d-90d1-4719-9724-2fcc56f8a274-run-httpd\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.132653 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-config-data\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.132674 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcsjj\" (UniqueName: \"kubernetes.io/projected/baf9222d-90d1-4719-9724-2fcc56f8a274-kube-api-access-lcsjj\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.132710 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.132738 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-scripts\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.132762 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.160333 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.234043 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdmzn\" (UniqueName: \"kubernetes.io/projected/a6abe987-223f-4d84-ae4d-8c89282f4b18-kube-api-access-gdmzn\") pod \"a6abe987-223f-4d84-ae4d-8c89282f4b18\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.234085 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6abe987-223f-4d84-ae4d-8c89282f4b18-config-data\") pod \"a6abe987-223f-4d84-ae4d-8c89282f4b18\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.234127 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6abe987-223f-4d84-ae4d-8c89282f4b18-combined-ca-bundle\") pod \"a6abe987-223f-4d84-ae4d-8c89282f4b18\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.234160 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6abe987-223f-4d84-ae4d-8c89282f4b18-logs\") pod \"a6abe987-223f-4d84-ae4d-8c89282f4b18\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.234541 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6abe987-223f-4d84-ae4d-8c89282f4b18-nova-metadata-tls-certs\") pod \"a6abe987-223f-4d84-ae4d-8c89282f4b18\" (UID: \"a6abe987-223f-4d84-ae4d-8c89282f4b18\") " Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.234913 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6abe987-223f-4d84-ae4d-8c89282f4b18-logs" (OuterVolumeSpecName: "logs") pod "a6abe987-223f-4d84-ae4d-8c89282f4b18" (UID: "a6abe987-223f-4d84-ae4d-8c89282f4b18"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.234972 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/baf9222d-90d1-4719-9724-2fcc56f8a274-log-httpd\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.235033 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.235064 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/baf9222d-90d1-4719-9724-2fcc56f8a274-run-httpd\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.235108 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-config-data\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.235131 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcsjj\" (UniqueName: \"kubernetes.io/projected/baf9222d-90d1-4719-9724-2fcc56f8a274-kube-api-access-lcsjj\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.235162 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.235215 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-scripts\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.235247 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.235717 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6abe987-223f-4d84-ae4d-8c89282f4b18-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.235745 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/baf9222d-90d1-4719-9724-2fcc56f8a274-log-httpd\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.236353 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/baf9222d-90d1-4719-9724-2fcc56f8a274-run-httpd\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.240458 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.241998 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.248808 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6abe987-223f-4d84-ae4d-8c89282f4b18-kube-api-access-gdmzn" (OuterVolumeSpecName: "kube-api-access-gdmzn") pod "a6abe987-223f-4d84-ae4d-8c89282f4b18" (UID: "a6abe987-223f-4d84-ae4d-8c89282f4b18"). InnerVolumeSpecName "kube-api-access-gdmzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.250132 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-config-data\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.250248 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.253047 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcsjj\" (UniqueName: \"kubernetes.io/projected/baf9222d-90d1-4719-9724-2fcc56f8a274-kube-api-access-lcsjj\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.253513 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-scripts\") pod \"ceilometer-0\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.282572 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6abe987-223f-4d84-ae4d-8c89282f4b18-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a6abe987-223f-4d84-ae4d-8c89282f4b18" (UID: "a6abe987-223f-4d84-ae4d-8c89282f4b18"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.293171 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6abe987-223f-4d84-ae4d-8c89282f4b18-config-data" (OuterVolumeSpecName: "config-data") pod "a6abe987-223f-4d84-ae4d-8c89282f4b18" (UID: "a6abe987-223f-4d84-ae4d-8c89282f4b18"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.304683 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6abe987-223f-4d84-ae4d-8c89282f4b18-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "a6abe987-223f-4d84-ae4d-8c89282f4b18" (UID: "a6abe987-223f-4d84-ae4d-8c89282f4b18"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.337726 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdmzn\" (UniqueName: \"kubernetes.io/projected/a6abe987-223f-4d84-ae4d-8c89282f4b18-kube-api-access-gdmzn\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.337760 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6abe987-223f-4d84-ae4d-8c89282f4b18-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.337769 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6abe987-223f-4d84-ae4d-8c89282f4b18-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.337778 4810 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6abe987-223f-4d84-ae4d-8c89282f4b18-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.346903 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.517979 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ba866fb-11bf-4edf-a0a9-1e7534cae07f" path="/var/lib/kubelet/pods/4ba866fb-11bf-4edf-a0a9-1e7534cae07f/volumes" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.634646 4810 generic.go:334] "Generic (PLEG): container finished" podID="a6abe987-223f-4d84-ae4d-8c89282f4b18" containerID="ad22eaa19e88e697a80abeb04365ae185673de285c03735e4c2038238c47ef75" exitCode=0 Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.634685 4810 generic.go:334] "Generic (PLEG): container finished" podID="a6abe987-223f-4d84-ae4d-8c89282f4b18" containerID="d2563a8c77321339ba89e24bec49f1370326419ad7015bd8fbd386dd91563a24" exitCode=143 Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.634691 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a6abe987-223f-4d84-ae4d-8c89282f4b18","Type":"ContainerDied","Data":"ad22eaa19e88e697a80abeb04365ae185673de285c03735e4c2038238c47ef75"} Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.634724 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.634742 4810 scope.go:117] "RemoveContainer" containerID="ad22eaa19e88e697a80abeb04365ae185673de285c03735e4c2038238c47ef75" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.634728 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a6abe987-223f-4d84-ae4d-8c89282f4b18","Type":"ContainerDied","Data":"d2563a8c77321339ba89e24bec49f1370326419ad7015bd8fbd386dd91563a24"} Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.634833 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a6abe987-223f-4d84-ae4d-8c89282f4b18","Type":"ContainerDied","Data":"bcbc0556f7cbc8370f4f1e7fcc83405f3ccbfcfb7925fc587e9176946fb48843"} Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.640257 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c","Type":"ContainerStarted","Data":"c0a6acaf5c3f0108d3d50cb630c0e2cb9470e93a6443905cbd9058d2ef094517"} Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.640325 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c","Type":"ContainerStarted","Data":"2fb0aac5102f60817c90c753fa782cb17744750246e5c4e2c759566245a5b0d6"} Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.640387 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.665343 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.676955 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.686843 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:57:46 crc kubenswrapper[4810]: E1201 14:57:46.687364 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6abe987-223f-4d84-ae4d-8c89282f4b18" containerName="nova-metadata-metadata" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.687389 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6abe987-223f-4d84-ae4d-8c89282f4b18" containerName="nova-metadata-metadata" Dec 01 14:57:46 crc kubenswrapper[4810]: E1201 14:57:46.687413 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6abe987-223f-4d84-ae4d-8c89282f4b18" containerName="nova-metadata-log" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.687422 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6abe987-223f-4d84-ae4d-8c89282f4b18" containerName="nova-metadata-log" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.687755 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6abe987-223f-4d84-ae4d-8c89282f4b18" containerName="nova-metadata-metadata" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.687781 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6abe987-223f-4d84-ae4d-8c89282f4b18" containerName="nova-metadata-log" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.688234 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.688220899 podStartE2EDuration="2.688220899s" podCreationTimestamp="2025-12-01 14:57:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:57:46.665901172 +0000 UTC m=+1432.429410775" watchObservedRunningTime="2025-12-01 14:57:46.688220899 +0000 UTC m=+1432.451730502" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.690081 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.694169 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.694373 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.701762 4810 scope.go:117] "RemoveContainer" containerID="d2563a8c77321339ba89e24bec49f1370326419ad7015bd8fbd386dd91563a24" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.725556 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.744198 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fed2100e-43c2-4cef-923b-74ce03403318-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " pod="openstack/nova-metadata-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.744366 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fed2100e-43c2-4cef-923b-74ce03403318-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " pod="openstack/nova-metadata-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.744450 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn44x\" (UniqueName: \"kubernetes.io/projected/fed2100e-43c2-4cef-923b-74ce03403318-kube-api-access-dn44x\") pod \"nova-metadata-0\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " pod="openstack/nova-metadata-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.744525 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fed2100e-43c2-4cef-923b-74ce03403318-config-data\") pod \"nova-metadata-0\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " pod="openstack/nova-metadata-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.751042 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fed2100e-43c2-4cef-923b-74ce03403318-logs\") pod \"nova-metadata-0\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " pod="openstack/nova-metadata-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.759778 4810 scope.go:117] "RemoveContainer" containerID="ad22eaa19e88e697a80abeb04365ae185673de285c03735e4c2038238c47ef75" Dec 01 14:57:46 crc kubenswrapper[4810]: E1201 14:57:46.763742 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad22eaa19e88e697a80abeb04365ae185673de285c03735e4c2038238c47ef75\": container with ID starting with ad22eaa19e88e697a80abeb04365ae185673de285c03735e4c2038238c47ef75 not found: ID does not exist" containerID="ad22eaa19e88e697a80abeb04365ae185673de285c03735e4c2038238c47ef75" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.763819 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad22eaa19e88e697a80abeb04365ae185673de285c03735e4c2038238c47ef75"} err="failed to get container status \"ad22eaa19e88e697a80abeb04365ae185673de285c03735e4c2038238c47ef75\": rpc error: code = NotFound desc = could not find container \"ad22eaa19e88e697a80abeb04365ae185673de285c03735e4c2038238c47ef75\": container with ID starting with ad22eaa19e88e697a80abeb04365ae185673de285c03735e4c2038238c47ef75 not found: ID does not exist" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.763860 4810 scope.go:117] "RemoveContainer" containerID="d2563a8c77321339ba89e24bec49f1370326419ad7015bd8fbd386dd91563a24" Dec 01 14:57:46 crc kubenswrapper[4810]: E1201 14:57:46.764997 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2563a8c77321339ba89e24bec49f1370326419ad7015bd8fbd386dd91563a24\": container with ID starting with d2563a8c77321339ba89e24bec49f1370326419ad7015bd8fbd386dd91563a24 not found: ID does not exist" containerID="d2563a8c77321339ba89e24bec49f1370326419ad7015bd8fbd386dd91563a24" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.765156 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2563a8c77321339ba89e24bec49f1370326419ad7015bd8fbd386dd91563a24"} err="failed to get container status \"d2563a8c77321339ba89e24bec49f1370326419ad7015bd8fbd386dd91563a24\": rpc error: code = NotFound desc = could not find container \"d2563a8c77321339ba89e24bec49f1370326419ad7015bd8fbd386dd91563a24\": container with ID starting with d2563a8c77321339ba89e24bec49f1370326419ad7015bd8fbd386dd91563a24 not found: ID does not exist" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.765220 4810 scope.go:117] "RemoveContainer" containerID="ad22eaa19e88e697a80abeb04365ae185673de285c03735e4c2038238c47ef75" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.765695 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad22eaa19e88e697a80abeb04365ae185673de285c03735e4c2038238c47ef75"} err="failed to get container status \"ad22eaa19e88e697a80abeb04365ae185673de285c03735e4c2038238c47ef75\": rpc error: code = NotFound desc = could not find container \"ad22eaa19e88e697a80abeb04365ae185673de285c03735e4c2038238c47ef75\": container with ID starting with ad22eaa19e88e697a80abeb04365ae185673de285c03735e4c2038238c47ef75 not found: ID does not exist" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.765745 4810 scope.go:117] "RemoveContainer" containerID="d2563a8c77321339ba89e24bec49f1370326419ad7015bd8fbd386dd91563a24" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.766241 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2563a8c77321339ba89e24bec49f1370326419ad7015bd8fbd386dd91563a24"} err="failed to get container status \"d2563a8c77321339ba89e24bec49f1370326419ad7015bd8fbd386dd91563a24\": rpc error: code = NotFound desc = could not find container \"d2563a8c77321339ba89e24bec49f1370326419ad7015bd8fbd386dd91563a24\": container with ID starting with d2563a8c77321339ba89e24bec49f1370326419ad7015bd8fbd386dd91563a24 not found: ID does not exist" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.792463 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:57:46 crc kubenswrapper[4810]: W1201 14:57:46.795641 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbaf9222d_90d1_4719_9724_2fcc56f8a274.slice/crio-d67790785a106884d6c70b39cb8df398127638fd89d787aade696f4d01844832 WatchSource:0}: Error finding container d67790785a106884d6c70b39cb8df398127638fd89d787aade696f4d01844832: Status 404 returned error can't find the container with id d67790785a106884d6c70b39cb8df398127638fd89d787aade696f4d01844832 Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.860140 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fed2100e-43c2-4cef-923b-74ce03403318-config-data\") pod \"nova-metadata-0\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " pod="openstack/nova-metadata-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.860207 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fed2100e-43c2-4cef-923b-74ce03403318-logs\") pod \"nova-metadata-0\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " pod="openstack/nova-metadata-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.860301 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fed2100e-43c2-4cef-923b-74ce03403318-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " pod="openstack/nova-metadata-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.860399 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fed2100e-43c2-4cef-923b-74ce03403318-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " pod="openstack/nova-metadata-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.860463 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn44x\" (UniqueName: \"kubernetes.io/projected/fed2100e-43c2-4cef-923b-74ce03403318-kube-api-access-dn44x\") pod \"nova-metadata-0\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " pod="openstack/nova-metadata-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.861148 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fed2100e-43c2-4cef-923b-74ce03403318-logs\") pod \"nova-metadata-0\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " pod="openstack/nova-metadata-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.865992 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fed2100e-43c2-4cef-923b-74ce03403318-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " pod="openstack/nova-metadata-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.866254 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fed2100e-43c2-4cef-923b-74ce03403318-config-data\") pod \"nova-metadata-0\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " pod="openstack/nova-metadata-0" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.873060 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fed2100e-43c2-4cef-923b-74ce03403318-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " pod="openstack/nova-metadata-0" Dec 01 14:57:46 crc kubenswrapper[4810]: E1201 14:57:46.873933 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="90196ba74428e571cbc8ee8f9ebb13d1d28fa0a9224cc71182c6382ec51bfcb5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 14:57:46 crc kubenswrapper[4810]: E1201 14:57:46.876599 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="90196ba74428e571cbc8ee8f9ebb13d1d28fa0a9224cc71182c6382ec51bfcb5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 14:57:46 crc kubenswrapper[4810]: E1201 14:57:46.878747 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="90196ba74428e571cbc8ee8f9ebb13d1d28fa0a9224cc71182c6382ec51bfcb5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 14:57:46 crc kubenswrapper[4810]: E1201 14:57:46.878813 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="1a18952e-ee54-4510-b149-e9e5cb873af5" containerName="nova-scheduler-scheduler" Dec 01 14:57:46 crc kubenswrapper[4810]: I1201 14:57:46.883957 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn44x\" (UniqueName: \"kubernetes.io/projected/fed2100e-43c2-4cef-923b-74ce03403318-kube-api-access-dn44x\") pod \"nova-metadata-0\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " pod="openstack/nova-metadata-0" Dec 01 14:57:47 crc kubenswrapper[4810]: I1201 14:57:47.015829 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 14:57:47 crc kubenswrapper[4810]: I1201 14:57:47.483288 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:57:47 crc kubenswrapper[4810]: W1201 14:57:47.494293 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfed2100e_43c2_4cef_923b_74ce03403318.slice/crio-b8111cd0730afc40da67d4d58d1d2dd5235a317acf0be7b876615a1da2e05465 WatchSource:0}: Error finding container b8111cd0730afc40da67d4d58d1d2dd5235a317acf0be7b876615a1da2e05465: Status 404 returned error can't find the container with id b8111cd0730afc40da67d4d58d1d2dd5235a317acf0be7b876615a1da2e05465 Dec 01 14:57:47 crc kubenswrapper[4810]: I1201 14:57:47.653123 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"baf9222d-90d1-4719-9724-2fcc56f8a274","Type":"ContainerStarted","Data":"d67790785a106884d6c70b39cb8df398127638fd89d787aade696f4d01844832"} Dec 01 14:57:47 crc kubenswrapper[4810]: I1201 14:57:47.654525 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fed2100e-43c2-4cef-923b-74ce03403318","Type":"ContainerStarted","Data":"b8111cd0730afc40da67d4d58d1d2dd5235a317acf0be7b876615a1da2e05465"} Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.448122 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.509046 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6abe987-223f-4d84-ae4d-8c89282f4b18" path="/var/lib/kubelet/pods/a6abe987-223f-4d84-ae4d-8c89282f4b18/volumes" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.522913 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f54zq\" (UniqueName: \"kubernetes.io/projected/1a18952e-ee54-4510-b149-e9e5cb873af5-kube-api-access-f54zq\") pod \"1a18952e-ee54-4510-b149-e9e5cb873af5\" (UID: \"1a18952e-ee54-4510-b149-e9e5cb873af5\") " Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.523250 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a18952e-ee54-4510-b149-e9e5cb873af5-config-data\") pod \"1a18952e-ee54-4510-b149-e9e5cb873af5\" (UID: \"1a18952e-ee54-4510-b149-e9e5cb873af5\") " Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.523388 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a18952e-ee54-4510-b149-e9e5cb873af5-combined-ca-bundle\") pod \"1a18952e-ee54-4510-b149-e9e5cb873af5\" (UID: \"1a18952e-ee54-4510-b149-e9e5cb873af5\") " Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.527312 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a18952e-ee54-4510-b149-e9e5cb873af5-kube-api-access-f54zq" (OuterVolumeSpecName: "kube-api-access-f54zq") pod "1a18952e-ee54-4510-b149-e9e5cb873af5" (UID: "1a18952e-ee54-4510-b149-e9e5cb873af5"). InnerVolumeSpecName "kube-api-access-f54zq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.558758 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a18952e-ee54-4510-b149-e9e5cb873af5-config-data" (OuterVolumeSpecName: "config-data") pod "1a18952e-ee54-4510-b149-e9e5cb873af5" (UID: "1a18952e-ee54-4510-b149-e9e5cb873af5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.568699 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a18952e-ee54-4510-b149-e9e5cb873af5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a18952e-ee54-4510-b149-e9e5cb873af5" (UID: "1a18952e-ee54-4510-b149-e9e5cb873af5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.625952 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a18952e-ee54-4510-b149-e9e5cb873af5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.626431 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f54zq\" (UniqueName: \"kubernetes.io/projected/1a18952e-ee54-4510-b149-e9e5cb873af5-kube-api-access-f54zq\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.626525 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a18952e-ee54-4510-b149-e9e5cb873af5-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.664807 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fed2100e-43c2-4cef-923b-74ce03403318","Type":"ContainerStarted","Data":"6b1083d4a1c26e507fb0dd8bbb368816fc6e29c3ba96ceb93667d63917c2dce9"} Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.665816 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fed2100e-43c2-4cef-923b-74ce03403318","Type":"ContainerStarted","Data":"e4e3e6e7ce19455507ba99bbc505942abf9a93bf221f3bdbf08c61912cec851e"} Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.668158 4810 generic.go:334] "Generic (PLEG): container finished" podID="1a18952e-ee54-4510-b149-e9e5cb873af5" containerID="90196ba74428e571cbc8ee8f9ebb13d1d28fa0a9224cc71182c6382ec51bfcb5" exitCode=0 Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.668325 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1a18952e-ee54-4510-b149-e9e5cb873af5","Type":"ContainerDied","Data":"90196ba74428e571cbc8ee8f9ebb13d1d28fa0a9224cc71182c6382ec51bfcb5"} Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.668397 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1a18952e-ee54-4510-b149-e9e5cb873af5","Type":"ContainerDied","Data":"e2eeb1c7160e1a72ac6978e9c38e9af08db57280618b87f96c3f19f337a4ae45"} Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.668418 4810 scope.go:117] "RemoveContainer" containerID="90196ba74428e571cbc8ee8f9ebb13d1d28fa0a9224cc71182c6382ec51bfcb5" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.668317 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.675868 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"baf9222d-90d1-4719-9724-2fcc56f8a274","Type":"ContainerStarted","Data":"4a0462fbaab266d41877dedb114d3fba11083bbeacab1a97b09ed0a0df90c04c"} Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.675909 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"baf9222d-90d1-4719-9724-2fcc56f8a274","Type":"ContainerStarted","Data":"d22d5a4ae6e8a9a8cf55a941e9767828e5423b7f6cd141dcf5324ecc553557a6"} Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.689967 4810 scope.go:117] "RemoveContainer" containerID="90196ba74428e571cbc8ee8f9ebb13d1d28fa0a9224cc71182c6382ec51bfcb5" Dec 01 14:57:48 crc kubenswrapper[4810]: E1201 14:57:48.690384 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90196ba74428e571cbc8ee8f9ebb13d1d28fa0a9224cc71182c6382ec51bfcb5\": container with ID starting with 90196ba74428e571cbc8ee8f9ebb13d1d28fa0a9224cc71182c6382ec51bfcb5 not found: ID does not exist" containerID="90196ba74428e571cbc8ee8f9ebb13d1d28fa0a9224cc71182c6382ec51bfcb5" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.690430 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90196ba74428e571cbc8ee8f9ebb13d1d28fa0a9224cc71182c6382ec51bfcb5"} err="failed to get container status \"90196ba74428e571cbc8ee8f9ebb13d1d28fa0a9224cc71182c6382ec51bfcb5\": rpc error: code = NotFound desc = could not find container \"90196ba74428e571cbc8ee8f9ebb13d1d28fa0a9224cc71182c6382ec51bfcb5\": container with ID starting with 90196ba74428e571cbc8ee8f9ebb13d1d28fa0a9224cc71182c6382ec51bfcb5 not found: ID does not exist" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.710850 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.7108334530000002 podStartE2EDuration="2.710833453s" podCreationTimestamp="2025-12-01 14:57:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:57:48.70667289 +0000 UTC m=+1434.470182513" watchObservedRunningTime="2025-12-01 14:57:48.710833453 +0000 UTC m=+1434.474343056" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.737408 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.755325 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.774282 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 14:57:48 crc kubenswrapper[4810]: E1201 14:57:48.775959 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a18952e-ee54-4510-b149-e9e5cb873af5" containerName="nova-scheduler-scheduler" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.775987 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a18952e-ee54-4510-b149-e9e5cb873af5" containerName="nova-scheduler-scheduler" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.776271 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a18952e-ee54-4510-b149-e9e5cb873af5" containerName="nova-scheduler-scheduler" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.777101 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.779676 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.788905 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.932183 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96ed5fc8-e0bf-47c4-bc9a-1923178532e9-config-data\") pod \"nova-scheduler-0\" (UID: \"96ed5fc8-e0bf-47c4-bc9a-1923178532e9\") " pod="openstack/nova-scheduler-0" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.932539 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96ed5fc8-e0bf-47c4-bc9a-1923178532e9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"96ed5fc8-e0bf-47c4-bc9a-1923178532e9\") " pod="openstack/nova-scheduler-0" Dec 01 14:57:48 crc kubenswrapper[4810]: I1201 14:57:48.932658 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25v7h\" (UniqueName: \"kubernetes.io/projected/96ed5fc8-e0bf-47c4-bc9a-1923178532e9-kube-api-access-25v7h\") pod \"nova-scheduler-0\" (UID: \"96ed5fc8-e0bf-47c4-bc9a-1923178532e9\") " pod="openstack/nova-scheduler-0" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.033764 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96ed5fc8-e0bf-47c4-bc9a-1923178532e9-config-data\") pod \"nova-scheduler-0\" (UID: \"96ed5fc8-e0bf-47c4-bc9a-1923178532e9\") " pod="openstack/nova-scheduler-0" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.033878 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96ed5fc8-e0bf-47c4-bc9a-1923178532e9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"96ed5fc8-e0bf-47c4-bc9a-1923178532e9\") " pod="openstack/nova-scheduler-0" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.033924 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25v7h\" (UniqueName: \"kubernetes.io/projected/96ed5fc8-e0bf-47c4-bc9a-1923178532e9-kube-api-access-25v7h\") pod \"nova-scheduler-0\" (UID: \"96ed5fc8-e0bf-47c4-bc9a-1923178532e9\") " pod="openstack/nova-scheduler-0" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.038214 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96ed5fc8-e0bf-47c4-bc9a-1923178532e9-config-data\") pod \"nova-scheduler-0\" (UID: \"96ed5fc8-e0bf-47c4-bc9a-1923178532e9\") " pod="openstack/nova-scheduler-0" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.039739 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96ed5fc8-e0bf-47c4-bc9a-1923178532e9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"96ed5fc8-e0bf-47c4-bc9a-1923178532e9\") " pod="openstack/nova-scheduler-0" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.051409 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25v7h\" (UniqueName: \"kubernetes.io/projected/96ed5fc8-e0bf-47c4-bc9a-1923178532e9-kube-api-access-25v7h\") pod \"nova-scheduler-0\" (UID: \"96ed5fc8-e0bf-47c4-bc9a-1923178532e9\") " pod="openstack/nova-scheduler-0" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.132735 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.628854 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.699613 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"baf9222d-90d1-4719-9724-2fcc56f8a274","Type":"ContainerStarted","Data":"4b47b89a8a4d01ac0d94dce0f58e23ad08cfa8d443ca6fbc22c18b4c2ff21ffe"} Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.702810 4810 generic.go:334] "Generic (PLEG): container finished" podID="55df027d-d8ed-477f-85e8-9c5548846f2b" containerID="03dd105231934f6f599d549d1c6952103e38d93f40910a0c330ff1bc6b92dee3" exitCode=0 Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.703864 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.704373 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55df027d-d8ed-477f-85e8-9c5548846f2b","Type":"ContainerDied","Data":"03dd105231934f6f599d549d1c6952103e38d93f40910a0c330ff1bc6b92dee3"} Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.704397 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"55df027d-d8ed-477f-85e8-9c5548846f2b","Type":"ContainerDied","Data":"493f249b8651b93539f277c48d95aad7d598ab0cb025e2d891579188eeb70b12"} Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.704412 4810 scope.go:117] "RemoveContainer" containerID="03dd105231934f6f599d549d1c6952103e38d93f40910a0c330ff1bc6b92dee3" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.727766 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.732688 4810 scope.go:117] "RemoveContainer" containerID="2c0ceed3036077fcb735cc5df4ddd017139fa9df3cde3ab8a62b54c5fba97c37" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.745079 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55df027d-d8ed-477f-85e8-9c5548846f2b-logs\") pod \"55df027d-d8ed-477f-85e8-9c5548846f2b\" (UID: \"55df027d-d8ed-477f-85e8-9c5548846f2b\") " Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.745397 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55df027d-d8ed-477f-85e8-9c5548846f2b-combined-ca-bundle\") pod \"55df027d-d8ed-477f-85e8-9c5548846f2b\" (UID: \"55df027d-d8ed-477f-85e8-9c5548846f2b\") " Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.745506 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2ftl\" (UniqueName: \"kubernetes.io/projected/55df027d-d8ed-477f-85e8-9c5548846f2b-kube-api-access-g2ftl\") pod \"55df027d-d8ed-477f-85e8-9c5548846f2b\" (UID: \"55df027d-d8ed-477f-85e8-9c5548846f2b\") " Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.745713 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55df027d-d8ed-477f-85e8-9c5548846f2b-config-data\") pod \"55df027d-d8ed-477f-85e8-9c5548846f2b\" (UID: \"55df027d-d8ed-477f-85e8-9c5548846f2b\") " Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.746730 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55df027d-d8ed-477f-85e8-9c5548846f2b-logs" (OuterVolumeSpecName: "logs") pod "55df027d-d8ed-477f-85e8-9c5548846f2b" (UID: "55df027d-d8ed-477f-85e8-9c5548846f2b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.748007 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55df027d-d8ed-477f-85e8-9c5548846f2b-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.753615 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55df027d-d8ed-477f-85e8-9c5548846f2b-kube-api-access-g2ftl" (OuterVolumeSpecName: "kube-api-access-g2ftl") pod "55df027d-d8ed-477f-85e8-9c5548846f2b" (UID: "55df027d-d8ed-477f-85e8-9c5548846f2b"). InnerVolumeSpecName "kube-api-access-g2ftl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.779160 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55df027d-d8ed-477f-85e8-9c5548846f2b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55df027d-d8ed-477f-85e8-9c5548846f2b" (UID: "55df027d-d8ed-477f-85e8-9c5548846f2b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.779548 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55df027d-d8ed-477f-85e8-9c5548846f2b-config-data" (OuterVolumeSpecName: "config-data") pod "55df027d-d8ed-477f-85e8-9c5548846f2b" (UID: "55df027d-d8ed-477f-85e8-9c5548846f2b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.793829 4810 scope.go:117] "RemoveContainer" containerID="03dd105231934f6f599d549d1c6952103e38d93f40910a0c330ff1bc6b92dee3" Dec 01 14:57:49 crc kubenswrapper[4810]: E1201 14:57:49.794306 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03dd105231934f6f599d549d1c6952103e38d93f40910a0c330ff1bc6b92dee3\": container with ID starting with 03dd105231934f6f599d549d1c6952103e38d93f40910a0c330ff1bc6b92dee3 not found: ID does not exist" containerID="03dd105231934f6f599d549d1c6952103e38d93f40910a0c330ff1bc6b92dee3" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.794352 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03dd105231934f6f599d549d1c6952103e38d93f40910a0c330ff1bc6b92dee3"} err="failed to get container status \"03dd105231934f6f599d549d1c6952103e38d93f40910a0c330ff1bc6b92dee3\": rpc error: code = NotFound desc = could not find container \"03dd105231934f6f599d549d1c6952103e38d93f40910a0c330ff1bc6b92dee3\": container with ID starting with 03dd105231934f6f599d549d1c6952103e38d93f40910a0c330ff1bc6b92dee3 not found: ID does not exist" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.794383 4810 scope.go:117] "RemoveContainer" containerID="2c0ceed3036077fcb735cc5df4ddd017139fa9df3cde3ab8a62b54c5fba97c37" Dec 01 14:57:49 crc kubenswrapper[4810]: E1201 14:57:49.794787 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c0ceed3036077fcb735cc5df4ddd017139fa9df3cde3ab8a62b54c5fba97c37\": container with ID starting with 2c0ceed3036077fcb735cc5df4ddd017139fa9df3cde3ab8a62b54c5fba97c37 not found: ID does not exist" containerID="2c0ceed3036077fcb735cc5df4ddd017139fa9df3cde3ab8a62b54c5fba97c37" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.794903 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c0ceed3036077fcb735cc5df4ddd017139fa9df3cde3ab8a62b54c5fba97c37"} err="failed to get container status \"2c0ceed3036077fcb735cc5df4ddd017139fa9df3cde3ab8a62b54c5fba97c37\": rpc error: code = NotFound desc = could not find container \"2c0ceed3036077fcb735cc5df4ddd017139fa9df3cde3ab8a62b54c5fba97c37\": container with ID starting with 2c0ceed3036077fcb735cc5df4ddd017139fa9df3cde3ab8a62b54c5fba97c37 not found: ID does not exist" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.850074 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55df027d-d8ed-477f-85e8-9c5548846f2b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.850116 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2ftl\" (UniqueName: \"kubernetes.io/projected/55df027d-d8ed-477f-85e8-9c5548846f2b-kube-api-access-g2ftl\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:49 crc kubenswrapper[4810]: I1201 14:57:49.850131 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55df027d-d8ed-477f-85e8-9c5548846f2b-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.036034 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.052955 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.063987 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 14:57:50 crc kubenswrapper[4810]: E1201 14:57:50.064698 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55df027d-d8ed-477f-85e8-9c5548846f2b" containerName="nova-api-api" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.064801 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="55df027d-d8ed-477f-85e8-9c5548846f2b" containerName="nova-api-api" Dec 01 14:57:50 crc kubenswrapper[4810]: E1201 14:57:50.064917 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55df027d-d8ed-477f-85e8-9c5548846f2b" containerName="nova-api-log" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.065019 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="55df027d-d8ed-477f-85e8-9c5548846f2b" containerName="nova-api-log" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.065345 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="55df027d-d8ed-477f-85e8-9c5548846f2b" containerName="nova-api-log" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.065449 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="55df027d-d8ed-477f-85e8-9c5548846f2b" containerName="nova-api-api" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.066824 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.069083 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.074945 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.155500 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-logs\") pod \"nova-api-0\" (UID: \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\") " pod="openstack/nova-api-0" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.155597 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-config-data\") pod \"nova-api-0\" (UID: \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\") " pod="openstack/nova-api-0" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.155648 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8nmw\" (UniqueName: \"kubernetes.io/projected/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-kube-api-access-k8nmw\") pod \"nova-api-0\" (UID: \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\") " pod="openstack/nova-api-0" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.155732 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\") " pod="openstack/nova-api-0" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.257192 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-logs\") pod \"nova-api-0\" (UID: \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\") " pod="openstack/nova-api-0" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.257935 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-config-data\") pod \"nova-api-0\" (UID: \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\") " pod="openstack/nova-api-0" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.258086 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8nmw\" (UniqueName: \"kubernetes.io/projected/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-kube-api-access-k8nmw\") pod \"nova-api-0\" (UID: \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\") " pod="openstack/nova-api-0" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.258289 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\") " pod="openstack/nova-api-0" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.257768 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-logs\") pod \"nova-api-0\" (UID: \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\") " pod="openstack/nova-api-0" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.263071 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-config-data\") pod \"nova-api-0\" (UID: \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\") " pod="openstack/nova-api-0" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.263167 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\") " pod="openstack/nova-api-0" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.277689 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8nmw\" (UniqueName: \"kubernetes.io/projected/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-kube-api-access-k8nmw\") pod \"nova-api-0\" (UID: \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\") " pod="openstack/nova-api-0" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.433080 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.505546 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a18952e-ee54-4510-b149-e9e5cb873af5" path="/var/lib/kubelet/pods/1a18952e-ee54-4510-b149-e9e5cb873af5/volumes" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.506569 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55df027d-d8ed-477f-85e8-9c5548846f2b" path="/var/lib/kubelet/pods/55df027d-d8ed-477f-85e8-9c5548846f2b/volumes" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.725788 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"96ed5fc8-e0bf-47c4-bc9a-1923178532e9","Type":"ContainerStarted","Data":"8326409636bfc6c1f41facf074040861e2e7e94c5b8dacc1e80290776fc88a2d"} Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.725829 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"96ed5fc8-e0bf-47c4-bc9a-1923178532e9","Type":"ContainerStarted","Data":"0057b321702147c9cd4510d0d3fd872015d5ce6f08aaf4573f4e946c513d11c4"} Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.752018 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.752000722 podStartE2EDuration="2.752000722s" podCreationTimestamp="2025-12-01 14:57:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:57:50.741565628 +0000 UTC m=+1436.505075251" watchObservedRunningTime="2025-12-01 14:57:50.752000722 +0000 UTC m=+1436.515510325" Dec 01 14:57:50 crc kubenswrapper[4810]: I1201 14:57:50.912006 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:57:50 crc kubenswrapper[4810]: W1201 14:57:50.919208 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0a8fc95_1ea0_4ad4_baea_bb7d91526d27.slice/crio-79619367d732b5bd3614d6d37a21cc45e811f2c1f0cfaa379478b407ae5d8987 WatchSource:0}: Error finding container 79619367d732b5bd3614d6d37a21cc45e811f2c1f0cfaa379478b407ae5d8987: Status 404 returned error can't find the container with id 79619367d732b5bd3614d6d37a21cc45e811f2c1f0cfaa379478b407ae5d8987 Dec 01 14:57:51 crc kubenswrapper[4810]: I1201 14:57:51.741712 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27","Type":"ContainerStarted","Data":"5e3b5b05fd190fd39768c2e9e77e605aae085c26f5305e0e1bd5ab65c2bbcfe1"} Dec 01 14:57:51 crc kubenswrapper[4810]: I1201 14:57:51.742080 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27","Type":"ContainerStarted","Data":"40005d5f28de244bc20dce68453da08d96cd2c26ce5b4d9fbe03d8c1f9e64395"} Dec 01 14:57:51 crc kubenswrapper[4810]: I1201 14:57:51.742097 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27","Type":"ContainerStarted","Data":"79619367d732b5bd3614d6d37a21cc45e811f2c1f0cfaa379478b407ae5d8987"} Dec 01 14:57:51 crc kubenswrapper[4810]: I1201 14:57:51.744758 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"baf9222d-90d1-4719-9724-2fcc56f8a274","Type":"ContainerStarted","Data":"e106d084e32d937ef0fefe5ebd21beeef8d4cc8f364ff93d1222fb8ed26f6ab6"} Dec 01 14:57:51 crc kubenswrapper[4810]: I1201 14:57:51.744916 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 14:57:51 crc kubenswrapper[4810]: I1201 14:57:51.765333 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.765312404 podStartE2EDuration="1.765312404s" podCreationTimestamp="2025-12-01 14:57:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:57:51.757153862 +0000 UTC m=+1437.520663465" watchObservedRunningTime="2025-12-01 14:57:51.765312404 +0000 UTC m=+1437.528822017" Dec 01 14:57:51 crc kubenswrapper[4810]: I1201 14:57:51.783862 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.582169034 podStartE2EDuration="6.783843897s" podCreationTimestamp="2025-12-01 14:57:45 +0000 UTC" firstStartedPulling="2025-12-01 14:57:46.800212584 +0000 UTC m=+1432.563722187" lastFinishedPulling="2025-12-01 14:57:51.001887447 +0000 UTC m=+1436.765397050" observedRunningTime="2025-12-01 14:57:51.778936064 +0000 UTC m=+1437.542445677" watchObservedRunningTime="2025-12-01 14:57:51.783843897 +0000 UTC m=+1437.547353500" Dec 01 14:57:52 crc kubenswrapper[4810]: I1201 14:57:52.017108 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 14:57:52 crc kubenswrapper[4810]: I1201 14:57:52.017161 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 14:57:53 crc kubenswrapper[4810]: I1201 14:57:53.027850 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 01 14:57:54 crc kubenswrapper[4810]: I1201 14:57:54.133151 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 14:57:54 crc kubenswrapper[4810]: I1201 14:57:54.973193 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 01 14:57:57 crc kubenswrapper[4810]: I1201 14:57:57.016788 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 14:57:57 crc kubenswrapper[4810]: I1201 14:57:57.017329 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 14:57:58 crc kubenswrapper[4810]: I1201 14:57:58.029728 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fed2100e-43c2-4cef-923b-74ce03403318" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 14:57:58 crc kubenswrapper[4810]: I1201 14:57:58.029772 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fed2100e-43c2-4cef-923b-74ce03403318" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 14:57:59 crc kubenswrapper[4810]: I1201 14:57:59.133274 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 14:57:59 crc kubenswrapper[4810]: I1201 14:57:59.164342 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 14:57:59 crc kubenswrapper[4810]: I1201 14:57:59.845577 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 14:58:00 crc kubenswrapper[4810]: I1201 14:58:00.434348 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 14:58:00 crc kubenswrapper[4810]: I1201 14:58:00.434401 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 14:58:01 crc kubenswrapper[4810]: I1201 14:58:01.516705 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e0a8fc95-1ea0-4ad4-baea-bb7d91526d27" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 14:58:01 crc kubenswrapper[4810]: I1201 14:58:01.517567 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e0a8fc95-1ea0-4ad4-baea-bb7d91526d27" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 14:58:02 crc kubenswrapper[4810]: I1201 14:58:02.972099 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:58:02 crc kubenswrapper[4810]: I1201 14:58:02.972435 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.022153 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.023919 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.034115 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.679991 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.820761 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t97q7\" (UniqueName: \"kubernetes.io/projected/20bf5931-cfe6-4607-89dc-164eef1386a9-kube-api-access-t97q7\") pod \"20bf5931-cfe6-4607-89dc-164eef1386a9\" (UID: \"20bf5931-cfe6-4607-89dc-164eef1386a9\") " Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.820848 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20bf5931-cfe6-4607-89dc-164eef1386a9-config-data\") pod \"20bf5931-cfe6-4607-89dc-164eef1386a9\" (UID: \"20bf5931-cfe6-4607-89dc-164eef1386a9\") " Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.821106 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20bf5931-cfe6-4607-89dc-164eef1386a9-combined-ca-bundle\") pod \"20bf5931-cfe6-4607-89dc-164eef1386a9\" (UID: \"20bf5931-cfe6-4607-89dc-164eef1386a9\") " Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.833758 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20bf5931-cfe6-4607-89dc-164eef1386a9-kube-api-access-t97q7" (OuterVolumeSpecName: "kube-api-access-t97q7") pod "20bf5931-cfe6-4607-89dc-164eef1386a9" (UID: "20bf5931-cfe6-4607-89dc-164eef1386a9"). InnerVolumeSpecName "kube-api-access-t97q7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.858119 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20bf5931-cfe6-4607-89dc-164eef1386a9-config-data" (OuterVolumeSpecName: "config-data") pod "20bf5931-cfe6-4607-89dc-164eef1386a9" (UID: "20bf5931-cfe6-4607-89dc-164eef1386a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.858486 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20bf5931-cfe6-4607-89dc-164eef1386a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20bf5931-cfe6-4607-89dc-164eef1386a9" (UID: "20bf5931-cfe6-4607-89dc-164eef1386a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.888303 4810 generic.go:334] "Generic (PLEG): container finished" podID="20bf5931-cfe6-4607-89dc-164eef1386a9" containerID="f246c42d90ab5616b2e473b63be7d3941dc56405f046e81b96754d4fc5182551" exitCode=137 Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.888352 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.888439 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"20bf5931-cfe6-4607-89dc-164eef1386a9","Type":"ContainerDied","Data":"f246c42d90ab5616b2e473b63be7d3941dc56405f046e81b96754d4fc5182551"} Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.888465 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"20bf5931-cfe6-4607-89dc-164eef1386a9","Type":"ContainerDied","Data":"fc2a927f8ae77e44ae3a010953178caf95ee8be5e41c8346e9dc086340771dfe"} Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.888504 4810 scope.go:117] "RemoveContainer" containerID="f246c42d90ab5616b2e473b63be7d3941dc56405f046e81b96754d4fc5182551" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.893570 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.922826 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t97q7\" (UniqueName: \"kubernetes.io/projected/20bf5931-cfe6-4607-89dc-164eef1386a9-kube-api-access-t97q7\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.922857 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20bf5931-cfe6-4607-89dc-164eef1386a9-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.922866 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20bf5931-cfe6-4607-89dc-164eef1386a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.935287 4810 scope.go:117] "RemoveContainer" containerID="f246c42d90ab5616b2e473b63be7d3941dc56405f046e81b96754d4fc5182551" Dec 01 14:58:07 crc kubenswrapper[4810]: E1201 14:58:07.941971 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f246c42d90ab5616b2e473b63be7d3941dc56405f046e81b96754d4fc5182551\": container with ID starting with f246c42d90ab5616b2e473b63be7d3941dc56405f046e81b96754d4fc5182551 not found: ID does not exist" containerID="f246c42d90ab5616b2e473b63be7d3941dc56405f046e81b96754d4fc5182551" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.942018 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f246c42d90ab5616b2e473b63be7d3941dc56405f046e81b96754d4fc5182551"} err="failed to get container status \"f246c42d90ab5616b2e473b63be7d3941dc56405f046e81b96754d4fc5182551\": rpc error: code = NotFound desc = could not find container \"f246c42d90ab5616b2e473b63be7d3941dc56405f046e81b96754d4fc5182551\": container with ID starting with f246c42d90ab5616b2e473b63be7d3941dc56405f046e81b96754d4fc5182551 not found: ID does not exist" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.953771 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.976186 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.985065 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 14:58:07 crc kubenswrapper[4810]: E1201 14:58:07.985525 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20bf5931-cfe6-4607-89dc-164eef1386a9" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.985547 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="20bf5931-cfe6-4607-89dc-164eef1386a9" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.985766 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="20bf5931-cfe6-4607-89dc-164eef1386a9" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.986543 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.989529 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.989564 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.989529 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 01 14:58:07 crc kubenswrapper[4810]: I1201 14:58:07.994118 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.125552 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.126822 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.126952 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.127056 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8fc8\" (UniqueName: \"kubernetes.io/projected/0899579c-adaa-41fb-a36d-31ba657c6c1a-kube-api-access-z8fc8\") pod \"nova-cell1-novncproxy-0\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.127194 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.228557 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.228617 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.228659 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.228684 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8fc8\" (UniqueName: \"kubernetes.io/projected/0899579c-adaa-41fb-a36d-31ba657c6c1a-kube-api-access-z8fc8\") pod \"nova-cell1-novncproxy-0\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.228758 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.238186 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.239970 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.240316 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.240444 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.251867 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8fc8\" (UniqueName: \"kubernetes.io/projected/0899579c-adaa-41fb-a36d-31ba657c6c1a-kube-api-access-z8fc8\") pod \"nova-cell1-novncproxy-0\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.323072 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.504500 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20bf5931-cfe6-4607-89dc-164eef1386a9" path="/var/lib/kubelet/pods/20bf5931-cfe6-4607-89dc-164eef1386a9/volumes" Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.810930 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 14:58:08 crc kubenswrapper[4810]: I1201 14:58:08.899330 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0899579c-adaa-41fb-a36d-31ba657c6c1a","Type":"ContainerStarted","Data":"2fca946a71c8696f06f213361f8621fab5f339501bd9ae641ecaca38a90be485"} Dec 01 14:58:09 crc kubenswrapper[4810]: I1201 14:58:09.921557 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0899579c-adaa-41fb-a36d-31ba657c6c1a","Type":"ContainerStarted","Data":"7f92191f00ad28697cca6ad61e70b41f045bb53297c7b28a21e7c29a04def64d"} Dec 01 14:58:09 crc kubenswrapper[4810]: I1201 14:58:09.947574 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.947548062 podStartE2EDuration="2.947548062s" podCreationTimestamp="2025-12-01 14:58:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:58:09.939492303 +0000 UTC m=+1455.703001936" watchObservedRunningTime="2025-12-01 14:58:09.947548062 +0000 UTC m=+1455.711057675" Dec 01 14:58:10 crc kubenswrapper[4810]: I1201 14:58:10.438925 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 14:58:10 crc kubenswrapper[4810]: I1201 14:58:10.439785 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 14:58:10 crc kubenswrapper[4810]: I1201 14:58:10.442075 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 14:58:10 crc kubenswrapper[4810]: I1201 14:58:10.443207 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 14:58:10 crc kubenswrapper[4810]: I1201 14:58:10.930642 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 14:58:10 crc kubenswrapper[4810]: I1201 14:58:10.935026 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.124941 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf"] Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.131646 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.137083 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf"] Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.286284 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-dns-svc\") pod \"dnsmasq-dns-5d8fc4ccc9-tqkdf\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.286384 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-ovsdbserver-nb\") pod \"dnsmasq-dns-5d8fc4ccc9-tqkdf\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.286441 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-dns-swift-storage-0\") pod \"dnsmasq-dns-5d8fc4ccc9-tqkdf\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.286478 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b8fj\" (UniqueName: \"kubernetes.io/projected/b88404c2-d4bf-4257-986a-1b78178fca88-kube-api-access-4b8fj\") pod \"dnsmasq-dns-5d8fc4ccc9-tqkdf\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.286508 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-ovsdbserver-sb\") pod \"dnsmasq-dns-5d8fc4ccc9-tqkdf\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.286575 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-config\") pod \"dnsmasq-dns-5d8fc4ccc9-tqkdf\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.388743 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-ovsdbserver-nb\") pod \"dnsmasq-dns-5d8fc4ccc9-tqkdf\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.388805 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-dns-swift-storage-0\") pod \"dnsmasq-dns-5d8fc4ccc9-tqkdf\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.388839 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b8fj\" (UniqueName: \"kubernetes.io/projected/b88404c2-d4bf-4257-986a-1b78178fca88-kube-api-access-4b8fj\") pod \"dnsmasq-dns-5d8fc4ccc9-tqkdf\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.388869 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-ovsdbserver-sb\") pod \"dnsmasq-dns-5d8fc4ccc9-tqkdf\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.388911 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-config\") pod \"dnsmasq-dns-5d8fc4ccc9-tqkdf\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.388951 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-dns-svc\") pod \"dnsmasq-dns-5d8fc4ccc9-tqkdf\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.389857 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-dns-svc\") pod \"dnsmasq-dns-5d8fc4ccc9-tqkdf\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.390380 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-ovsdbserver-nb\") pod \"dnsmasq-dns-5d8fc4ccc9-tqkdf\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.390942 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-dns-swift-storage-0\") pod \"dnsmasq-dns-5d8fc4ccc9-tqkdf\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.391850 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-ovsdbserver-sb\") pod \"dnsmasq-dns-5d8fc4ccc9-tqkdf\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.393071 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-config\") pod \"dnsmasq-dns-5d8fc4ccc9-tqkdf\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.413870 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b8fj\" (UniqueName: \"kubernetes.io/projected/b88404c2-d4bf-4257-986a-1b78178fca88-kube-api-access-4b8fj\") pod \"dnsmasq-dns-5d8fc4ccc9-tqkdf\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.457451 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:11 crc kubenswrapper[4810]: I1201 14:58:11.938917 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf"] Dec 01 14:58:12 crc kubenswrapper[4810]: I1201 14:58:12.957150 4810 generic.go:334] "Generic (PLEG): container finished" podID="b88404c2-d4bf-4257-986a-1b78178fca88" containerID="d8dada242e998791e9f2bfdb36e94b01560f068e320ef4e7eec1b062c6a1f3ff" exitCode=0 Dec 01 14:58:12 crc kubenswrapper[4810]: I1201 14:58:12.957251 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" event={"ID":"b88404c2-d4bf-4257-986a-1b78178fca88","Type":"ContainerDied","Data":"d8dada242e998791e9f2bfdb36e94b01560f068e320ef4e7eec1b062c6a1f3ff"} Dec 01 14:58:12 crc kubenswrapper[4810]: I1201 14:58:12.957642 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" event={"ID":"b88404c2-d4bf-4257-986a-1b78178fca88","Type":"ContainerStarted","Data":"ee8a52a8581a97da6378bf0cfba49cdaf6dd1156e9099284a43ed8cf18689572"} Dec 01 14:58:13 crc kubenswrapper[4810]: I1201 14:58:13.010320 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:58:13 crc kubenswrapper[4810]: I1201 14:58:13.010616 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerName="ceilometer-central-agent" containerID="cri-o://d22d5a4ae6e8a9a8cf55a941e9767828e5423b7f6cd141dcf5324ecc553557a6" gracePeriod=30 Dec 01 14:58:13 crc kubenswrapper[4810]: I1201 14:58:13.011307 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerName="proxy-httpd" containerID="cri-o://e106d084e32d937ef0fefe5ebd21beeef8d4cc8f364ff93d1222fb8ed26f6ab6" gracePeriod=30 Dec 01 14:58:13 crc kubenswrapper[4810]: I1201 14:58:13.011362 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerName="ceilometer-notification-agent" containerID="cri-o://4a0462fbaab266d41877dedb114d3fba11083bbeacab1a97b09ed0a0df90c04c" gracePeriod=30 Dec 01 14:58:13 crc kubenswrapper[4810]: I1201 14:58:13.011427 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerName="sg-core" containerID="cri-o://4b47b89a8a4d01ac0d94dce0f58e23ad08cfa8d443ca6fbc22c18b4c2ff21ffe" gracePeriod=30 Dec 01 14:58:13 crc kubenswrapper[4810]: I1201 14:58:13.018895 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.189:3000/\": EOF" Dec 01 14:58:13 crc kubenswrapper[4810]: I1201 14:58:13.323979 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:13 crc kubenswrapper[4810]: I1201 14:58:13.968598 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" event={"ID":"b88404c2-d4bf-4257-986a-1b78178fca88","Type":"ContainerStarted","Data":"4d129b7c0c3d768fea14f715dcd8a374cb51cf6191bbfa597636e8ce11c8d85e"} Dec 01 14:58:13 crc kubenswrapper[4810]: I1201 14:58:13.968844 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:13 crc kubenswrapper[4810]: I1201 14:58:13.971404 4810 generic.go:334] "Generic (PLEG): container finished" podID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerID="e106d084e32d937ef0fefe5ebd21beeef8d4cc8f364ff93d1222fb8ed26f6ab6" exitCode=0 Dec 01 14:58:13 crc kubenswrapper[4810]: I1201 14:58:13.971433 4810 generic.go:334] "Generic (PLEG): container finished" podID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerID="4b47b89a8a4d01ac0d94dce0f58e23ad08cfa8d443ca6fbc22c18b4c2ff21ffe" exitCode=2 Dec 01 14:58:13 crc kubenswrapper[4810]: I1201 14:58:13.971442 4810 generic.go:334] "Generic (PLEG): container finished" podID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerID="d22d5a4ae6e8a9a8cf55a941e9767828e5423b7f6cd141dcf5324ecc553557a6" exitCode=0 Dec 01 14:58:13 crc kubenswrapper[4810]: I1201 14:58:13.971483 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"baf9222d-90d1-4719-9724-2fcc56f8a274","Type":"ContainerDied","Data":"e106d084e32d937ef0fefe5ebd21beeef8d4cc8f364ff93d1222fb8ed26f6ab6"} Dec 01 14:58:13 crc kubenswrapper[4810]: I1201 14:58:13.971528 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"baf9222d-90d1-4719-9724-2fcc56f8a274","Type":"ContainerDied","Data":"4b47b89a8a4d01ac0d94dce0f58e23ad08cfa8d443ca6fbc22c18b4c2ff21ffe"} Dec 01 14:58:13 crc kubenswrapper[4810]: I1201 14:58:13.971547 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"baf9222d-90d1-4719-9724-2fcc56f8a274","Type":"ContainerDied","Data":"d22d5a4ae6e8a9a8cf55a941e9767828e5423b7f6cd141dcf5324ecc553557a6"} Dec 01 14:58:14 crc kubenswrapper[4810]: I1201 14:58:14.007547 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" podStartSLOduration=3.007528692 podStartE2EDuration="3.007528692s" podCreationTimestamp="2025-12-01 14:58:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:58:14.003823941 +0000 UTC m=+1459.767333544" watchObservedRunningTime="2025-12-01 14:58:14.007528692 +0000 UTC m=+1459.771038305" Dec 01 14:58:14 crc kubenswrapper[4810]: I1201 14:58:14.550207 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:58:14 crc kubenswrapper[4810]: I1201 14:58:14.550854 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e0a8fc95-1ea0-4ad4-baea-bb7d91526d27" containerName="nova-api-log" containerID="cri-o://40005d5f28de244bc20dce68453da08d96cd2c26ce5b4d9fbe03d8c1f9e64395" gracePeriod=30 Dec 01 14:58:14 crc kubenswrapper[4810]: I1201 14:58:14.550905 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e0a8fc95-1ea0-4ad4-baea-bb7d91526d27" containerName="nova-api-api" containerID="cri-o://5e3b5b05fd190fd39768c2e9e77e605aae085c26f5305e0e1bd5ab65c2bbcfe1" gracePeriod=30 Dec 01 14:58:14 crc kubenswrapper[4810]: I1201 14:58:14.987824 4810 generic.go:334] "Generic (PLEG): container finished" podID="e0a8fc95-1ea0-4ad4-baea-bb7d91526d27" containerID="40005d5f28de244bc20dce68453da08d96cd2c26ce5b4d9fbe03d8c1f9e64395" exitCode=143 Dec 01 14:58:14 crc kubenswrapper[4810]: I1201 14:58:14.987884 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27","Type":"ContainerDied","Data":"40005d5f28de244bc20dce68453da08d96cd2c26ce5b4d9fbe03d8c1f9e64395"} Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.407596 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.563549 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcsjj\" (UniqueName: \"kubernetes.io/projected/baf9222d-90d1-4719-9724-2fcc56f8a274-kube-api-access-lcsjj\") pod \"baf9222d-90d1-4719-9724-2fcc56f8a274\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.563594 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-ceilometer-tls-certs\") pod \"baf9222d-90d1-4719-9724-2fcc56f8a274\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.563708 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/baf9222d-90d1-4719-9724-2fcc56f8a274-run-httpd\") pod \"baf9222d-90d1-4719-9724-2fcc56f8a274\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.563747 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-sg-core-conf-yaml\") pod \"baf9222d-90d1-4719-9724-2fcc56f8a274\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.563775 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-config-data\") pod \"baf9222d-90d1-4719-9724-2fcc56f8a274\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.563800 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-scripts\") pod \"baf9222d-90d1-4719-9724-2fcc56f8a274\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.563903 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/baf9222d-90d1-4719-9724-2fcc56f8a274-log-httpd\") pod \"baf9222d-90d1-4719-9724-2fcc56f8a274\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.563953 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-combined-ca-bundle\") pod \"baf9222d-90d1-4719-9724-2fcc56f8a274\" (UID: \"baf9222d-90d1-4719-9724-2fcc56f8a274\") " Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.564846 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/baf9222d-90d1-4719-9724-2fcc56f8a274-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "baf9222d-90d1-4719-9724-2fcc56f8a274" (UID: "baf9222d-90d1-4719-9724-2fcc56f8a274"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.565096 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/baf9222d-90d1-4719-9724-2fcc56f8a274-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "baf9222d-90d1-4719-9724-2fcc56f8a274" (UID: "baf9222d-90d1-4719-9724-2fcc56f8a274"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.565261 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/baf9222d-90d1-4719-9724-2fcc56f8a274-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.565285 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/baf9222d-90d1-4719-9724-2fcc56f8a274-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.570871 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-scripts" (OuterVolumeSpecName: "scripts") pod "baf9222d-90d1-4719-9724-2fcc56f8a274" (UID: "baf9222d-90d1-4719-9724-2fcc56f8a274"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.572915 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baf9222d-90d1-4719-9724-2fcc56f8a274-kube-api-access-lcsjj" (OuterVolumeSpecName: "kube-api-access-lcsjj") pod "baf9222d-90d1-4719-9724-2fcc56f8a274" (UID: "baf9222d-90d1-4719-9724-2fcc56f8a274"). InnerVolumeSpecName "kube-api-access-lcsjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.600224 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "baf9222d-90d1-4719-9724-2fcc56f8a274" (UID: "baf9222d-90d1-4719-9724-2fcc56f8a274"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.635364 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "baf9222d-90d1-4719-9724-2fcc56f8a274" (UID: "baf9222d-90d1-4719-9724-2fcc56f8a274"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.658718 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "baf9222d-90d1-4719-9724-2fcc56f8a274" (UID: "baf9222d-90d1-4719-9724-2fcc56f8a274"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.667655 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.667687 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcsjj\" (UniqueName: \"kubernetes.io/projected/baf9222d-90d1-4719-9724-2fcc56f8a274-kube-api-access-lcsjj\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.667700 4810 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.667709 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.667717 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.701417 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-config-data" (OuterVolumeSpecName: "config-data") pod "baf9222d-90d1-4719-9724-2fcc56f8a274" (UID: "baf9222d-90d1-4719-9724-2fcc56f8a274"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:15 crc kubenswrapper[4810]: I1201 14:58:15.769555 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf9222d-90d1-4719-9724-2fcc56f8a274-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.001286 4810 generic.go:334] "Generic (PLEG): container finished" podID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerID="4a0462fbaab266d41877dedb114d3fba11083bbeacab1a97b09ed0a0df90c04c" exitCode=0 Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.001352 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"baf9222d-90d1-4719-9724-2fcc56f8a274","Type":"ContainerDied","Data":"4a0462fbaab266d41877dedb114d3fba11083bbeacab1a97b09ed0a0df90c04c"} Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.001404 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"baf9222d-90d1-4719-9724-2fcc56f8a274","Type":"ContainerDied","Data":"d67790785a106884d6c70b39cb8df398127638fd89d787aade696f4d01844832"} Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.001424 4810 scope.go:117] "RemoveContainer" containerID="e106d084e32d937ef0fefe5ebd21beeef8d4cc8f364ff93d1222fb8ed26f6ab6" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.001672 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.031675 4810 scope.go:117] "RemoveContainer" containerID="4b47b89a8a4d01ac0d94dce0f58e23ad08cfa8d443ca6fbc22c18b4c2ff21ffe" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.061137 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.068187 4810 scope.go:117] "RemoveContainer" containerID="4a0462fbaab266d41877dedb114d3fba11083bbeacab1a97b09ed0a0df90c04c" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.087431 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.097102 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:58:16 crc kubenswrapper[4810]: E1201 14:58:16.097648 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerName="ceilometer-notification-agent" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.097714 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerName="ceilometer-notification-agent" Dec 01 14:58:16 crc kubenswrapper[4810]: E1201 14:58:16.097782 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerName="proxy-httpd" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.097852 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerName="proxy-httpd" Dec 01 14:58:16 crc kubenswrapper[4810]: E1201 14:58:16.097912 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerName="sg-core" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.097972 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerName="sg-core" Dec 01 14:58:16 crc kubenswrapper[4810]: E1201 14:58:16.098040 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerName="ceilometer-central-agent" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.098097 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerName="ceilometer-central-agent" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.098314 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerName="proxy-httpd" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.098388 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerName="ceilometer-notification-agent" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.098446 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerName="sg-core" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.098527 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="baf9222d-90d1-4719-9724-2fcc56f8a274" containerName="ceilometer-central-agent" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.100211 4810 scope.go:117] "RemoveContainer" containerID="d22d5a4ae6e8a9a8cf55a941e9767828e5423b7f6cd141dcf5324ecc553557a6" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.121945 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.125296 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.125533 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.125840 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.125910 4810 scope.go:117] "RemoveContainer" containerID="e106d084e32d937ef0fefe5ebd21beeef8d4cc8f364ff93d1222fb8ed26f6ab6" Dec 01 14:58:16 crc kubenswrapper[4810]: E1201 14:58:16.127832 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e106d084e32d937ef0fefe5ebd21beeef8d4cc8f364ff93d1222fb8ed26f6ab6\": container with ID starting with e106d084e32d937ef0fefe5ebd21beeef8d4cc8f364ff93d1222fb8ed26f6ab6 not found: ID does not exist" containerID="e106d084e32d937ef0fefe5ebd21beeef8d4cc8f364ff93d1222fb8ed26f6ab6" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.127869 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e106d084e32d937ef0fefe5ebd21beeef8d4cc8f364ff93d1222fb8ed26f6ab6"} err="failed to get container status \"e106d084e32d937ef0fefe5ebd21beeef8d4cc8f364ff93d1222fb8ed26f6ab6\": rpc error: code = NotFound desc = could not find container \"e106d084e32d937ef0fefe5ebd21beeef8d4cc8f364ff93d1222fb8ed26f6ab6\": container with ID starting with e106d084e32d937ef0fefe5ebd21beeef8d4cc8f364ff93d1222fb8ed26f6ab6 not found: ID does not exist" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.127895 4810 scope.go:117] "RemoveContainer" containerID="4b47b89a8a4d01ac0d94dce0f58e23ad08cfa8d443ca6fbc22c18b4c2ff21ffe" Dec 01 14:58:16 crc kubenswrapper[4810]: E1201 14:58:16.128168 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b47b89a8a4d01ac0d94dce0f58e23ad08cfa8d443ca6fbc22c18b4c2ff21ffe\": container with ID starting with 4b47b89a8a4d01ac0d94dce0f58e23ad08cfa8d443ca6fbc22c18b4c2ff21ffe not found: ID does not exist" containerID="4b47b89a8a4d01ac0d94dce0f58e23ad08cfa8d443ca6fbc22c18b4c2ff21ffe" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.128195 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b47b89a8a4d01ac0d94dce0f58e23ad08cfa8d443ca6fbc22c18b4c2ff21ffe"} err="failed to get container status \"4b47b89a8a4d01ac0d94dce0f58e23ad08cfa8d443ca6fbc22c18b4c2ff21ffe\": rpc error: code = NotFound desc = could not find container \"4b47b89a8a4d01ac0d94dce0f58e23ad08cfa8d443ca6fbc22c18b4c2ff21ffe\": container with ID starting with 4b47b89a8a4d01ac0d94dce0f58e23ad08cfa8d443ca6fbc22c18b4c2ff21ffe not found: ID does not exist" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.128214 4810 scope.go:117] "RemoveContainer" containerID="4a0462fbaab266d41877dedb114d3fba11083bbeacab1a97b09ed0a0df90c04c" Dec 01 14:58:16 crc kubenswrapper[4810]: E1201 14:58:16.128843 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a0462fbaab266d41877dedb114d3fba11083bbeacab1a97b09ed0a0df90c04c\": container with ID starting with 4a0462fbaab266d41877dedb114d3fba11083bbeacab1a97b09ed0a0df90c04c not found: ID does not exist" containerID="4a0462fbaab266d41877dedb114d3fba11083bbeacab1a97b09ed0a0df90c04c" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.128872 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a0462fbaab266d41877dedb114d3fba11083bbeacab1a97b09ed0a0df90c04c"} err="failed to get container status \"4a0462fbaab266d41877dedb114d3fba11083bbeacab1a97b09ed0a0df90c04c\": rpc error: code = NotFound desc = could not find container \"4a0462fbaab266d41877dedb114d3fba11083bbeacab1a97b09ed0a0df90c04c\": container with ID starting with 4a0462fbaab266d41877dedb114d3fba11083bbeacab1a97b09ed0a0df90c04c not found: ID does not exist" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.128891 4810 scope.go:117] "RemoveContainer" containerID="d22d5a4ae6e8a9a8cf55a941e9767828e5423b7f6cd141dcf5324ecc553557a6" Dec 01 14:58:16 crc kubenswrapper[4810]: E1201 14:58:16.129513 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d22d5a4ae6e8a9a8cf55a941e9767828e5423b7f6cd141dcf5324ecc553557a6\": container with ID starting with d22d5a4ae6e8a9a8cf55a941e9767828e5423b7f6cd141dcf5324ecc553557a6 not found: ID does not exist" containerID="d22d5a4ae6e8a9a8cf55a941e9767828e5423b7f6cd141dcf5324ecc553557a6" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.129543 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d22d5a4ae6e8a9a8cf55a941e9767828e5423b7f6cd141dcf5324ecc553557a6"} err="failed to get container status \"d22d5a4ae6e8a9a8cf55a941e9767828e5423b7f6cd141dcf5324ecc553557a6\": rpc error: code = NotFound desc = could not find container \"d22d5a4ae6e8a9a8cf55a941e9767828e5423b7f6cd141dcf5324ecc553557a6\": container with ID starting with d22d5a4ae6e8a9a8cf55a941e9767828e5423b7f6cd141dcf5324ecc553557a6 not found: ID does not exist" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.139357 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.276989 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.277043 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-config-data\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.277076 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-scripts\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.277110 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.277167 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.277228 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31046f73-59af-4c05-976d-65186bb31cec-run-httpd\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.277269 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxxk8\" (UniqueName: \"kubernetes.io/projected/31046f73-59af-4c05-976d-65186bb31cec-kube-api-access-mxxk8\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.277291 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31046f73-59af-4c05-976d-65186bb31cec-log-httpd\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.378995 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31046f73-59af-4c05-976d-65186bb31cec-run-httpd\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.379078 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxxk8\" (UniqueName: \"kubernetes.io/projected/31046f73-59af-4c05-976d-65186bb31cec-kube-api-access-mxxk8\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.379107 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31046f73-59af-4c05-976d-65186bb31cec-log-httpd\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.379181 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.379230 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-config-data\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.379270 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-scripts\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.379589 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31046f73-59af-4c05-976d-65186bb31cec-run-httpd\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.379637 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31046f73-59af-4c05-976d-65186bb31cec-log-httpd\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.380101 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.380264 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.384360 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.384488 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-scripts\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.385427 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.390882 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-config-data\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.391641 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.397009 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxxk8\" (UniqueName: \"kubernetes.io/projected/31046f73-59af-4c05-976d-65186bb31cec-kube-api-access-mxxk8\") pod \"ceilometer-0\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.444068 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.503825 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="baf9222d-90d1-4719-9724-2fcc56f8a274" path="/var/lib/kubelet/pods/baf9222d-90d1-4719-9724-2fcc56f8a274/volumes" Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.641370 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.965123 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 14:58:16 crc kubenswrapper[4810]: I1201 14:58:16.974910 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:58:17 crc kubenswrapper[4810]: I1201 14:58:17.010321 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31046f73-59af-4c05-976d-65186bb31cec","Type":"ContainerStarted","Data":"43b85130635d6b96e761d00728f40ebdd36f1363acb56976255e47235f6eeb05"} Dec 01 14:58:17 crc kubenswrapper[4810]: I1201 14:58:17.598950 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ftjmr"] Dec 01 14:58:17 crc kubenswrapper[4810]: I1201 14:58:17.601099 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ftjmr" Dec 01 14:58:17 crc kubenswrapper[4810]: I1201 14:58:17.620318 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ftjmr"] Dec 01 14:58:17 crc kubenswrapper[4810]: I1201 14:58:17.702865 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmcmm\" (UniqueName: \"kubernetes.io/projected/9dd22986-8614-4e8c-9a29-6639c5d14f96-kube-api-access-gmcmm\") pod \"redhat-operators-ftjmr\" (UID: \"9dd22986-8614-4e8c-9a29-6639c5d14f96\") " pod="openshift-marketplace/redhat-operators-ftjmr" Dec 01 14:58:17 crc kubenswrapper[4810]: I1201 14:58:17.702968 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dd22986-8614-4e8c-9a29-6639c5d14f96-catalog-content\") pod \"redhat-operators-ftjmr\" (UID: \"9dd22986-8614-4e8c-9a29-6639c5d14f96\") " pod="openshift-marketplace/redhat-operators-ftjmr" Dec 01 14:58:17 crc kubenswrapper[4810]: I1201 14:58:17.703090 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dd22986-8614-4e8c-9a29-6639c5d14f96-utilities\") pod \"redhat-operators-ftjmr\" (UID: \"9dd22986-8614-4e8c-9a29-6639c5d14f96\") " pod="openshift-marketplace/redhat-operators-ftjmr" Dec 01 14:58:17 crc kubenswrapper[4810]: I1201 14:58:17.805301 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dd22986-8614-4e8c-9a29-6639c5d14f96-utilities\") pod \"redhat-operators-ftjmr\" (UID: \"9dd22986-8614-4e8c-9a29-6639c5d14f96\") " pod="openshift-marketplace/redhat-operators-ftjmr" Dec 01 14:58:17 crc kubenswrapper[4810]: I1201 14:58:17.805399 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmcmm\" (UniqueName: \"kubernetes.io/projected/9dd22986-8614-4e8c-9a29-6639c5d14f96-kube-api-access-gmcmm\") pod \"redhat-operators-ftjmr\" (UID: \"9dd22986-8614-4e8c-9a29-6639c5d14f96\") " pod="openshift-marketplace/redhat-operators-ftjmr" Dec 01 14:58:17 crc kubenswrapper[4810]: I1201 14:58:17.805522 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dd22986-8614-4e8c-9a29-6639c5d14f96-catalog-content\") pod \"redhat-operators-ftjmr\" (UID: \"9dd22986-8614-4e8c-9a29-6639c5d14f96\") " pod="openshift-marketplace/redhat-operators-ftjmr" Dec 01 14:58:17 crc kubenswrapper[4810]: I1201 14:58:17.805744 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dd22986-8614-4e8c-9a29-6639c5d14f96-utilities\") pod \"redhat-operators-ftjmr\" (UID: \"9dd22986-8614-4e8c-9a29-6639c5d14f96\") " pod="openshift-marketplace/redhat-operators-ftjmr" Dec 01 14:58:17 crc kubenswrapper[4810]: I1201 14:58:17.806043 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dd22986-8614-4e8c-9a29-6639c5d14f96-catalog-content\") pod \"redhat-operators-ftjmr\" (UID: \"9dd22986-8614-4e8c-9a29-6639c5d14f96\") " pod="openshift-marketplace/redhat-operators-ftjmr" Dec 01 14:58:17 crc kubenswrapper[4810]: I1201 14:58:17.830579 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmcmm\" (UniqueName: \"kubernetes.io/projected/9dd22986-8614-4e8c-9a29-6639c5d14f96-kube-api-access-gmcmm\") pod \"redhat-operators-ftjmr\" (UID: \"9dd22986-8614-4e8c-9a29-6639c5d14f96\") " pod="openshift-marketplace/redhat-operators-ftjmr" Dec 01 14:58:17 crc kubenswrapper[4810]: I1201 14:58:17.972292 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ftjmr" Dec 01 14:58:18 crc kubenswrapper[4810]: I1201 14:58:18.323462 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:18 crc kubenswrapper[4810]: I1201 14:58:18.346858 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:18 crc kubenswrapper[4810]: W1201 14:58:18.472320 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9dd22986_8614_4e8c_9a29_6639c5d14f96.slice/crio-6bb5c7d0254a00e9e37968102879aa6fb1ac36054c843817774f2a1ae83c0f93 WatchSource:0}: Error finding container 6bb5c7d0254a00e9e37968102879aa6fb1ac36054c843817774f2a1ae83c0f93: Status 404 returned error can't find the container with id 6bb5c7d0254a00e9e37968102879aa6fb1ac36054c843817774f2a1ae83c0f93 Dec 01 14:58:18 crc kubenswrapper[4810]: I1201 14:58:18.472405 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ftjmr"] Dec 01 14:58:18 crc kubenswrapper[4810]: I1201 14:58:18.640865 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 14:58:18 crc kubenswrapper[4810]: I1201 14:58:18.724896 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-logs\") pod \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\" (UID: \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\") " Dec 01 14:58:18 crc kubenswrapper[4810]: I1201 14:58:18.725045 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8nmw\" (UniqueName: \"kubernetes.io/projected/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-kube-api-access-k8nmw\") pod \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\" (UID: \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\") " Dec 01 14:58:18 crc kubenswrapper[4810]: I1201 14:58:18.725112 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-config-data\") pod \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\" (UID: \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\") " Dec 01 14:58:18 crc kubenswrapper[4810]: I1201 14:58:18.725159 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-combined-ca-bundle\") pod \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\" (UID: \"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27\") " Dec 01 14:58:18 crc kubenswrapper[4810]: I1201 14:58:18.726579 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-logs" (OuterVolumeSpecName: "logs") pod "e0a8fc95-1ea0-4ad4-baea-bb7d91526d27" (UID: "e0a8fc95-1ea0-4ad4-baea-bb7d91526d27"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:58:18 crc kubenswrapper[4810]: I1201 14:58:18.735938 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-kube-api-access-k8nmw" (OuterVolumeSpecName: "kube-api-access-k8nmw") pod "e0a8fc95-1ea0-4ad4-baea-bb7d91526d27" (UID: "e0a8fc95-1ea0-4ad4-baea-bb7d91526d27"). InnerVolumeSpecName "kube-api-access-k8nmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:58:18 crc kubenswrapper[4810]: I1201 14:58:18.769770 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-config-data" (OuterVolumeSpecName: "config-data") pod "e0a8fc95-1ea0-4ad4-baea-bb7d91526d27" (UID: "e0a8fc95-1ea0-4ad4-baea-bb7d91526d27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:18 crc kubenswrapper[4810]: I1201 14:58:18.788512 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0a8fc95-1ea0-4ad4-baea-bb7d91526d27" (UID: "e0a8fc95-1ea0-4ad4-baea-bb7d91526d27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:18 crc kubenswrapper[4810]: I1201 14:58:18.827211 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:18 crc kubenswrapper[4810]: I1201 14:58:18.827248 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8nmw\" (UniqueName: \"kubernetes.io/projected/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-kube-api-access-k8nmw\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:18 crc kubenswrapper[4810]: I1201 14:58:18.827258 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:18 crc kubenswrapper[4810]: I1201 14:58:18.827267 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.029692 4810 generic.go:334] "Generic (PLEG): container finished" podID="9dd22986-8614-4e8c-9a29-6639c5d14f96" containerID="ee7c25279f755f3a8977dd66d11b1c3811f94c76bab20858c22f558be9294b19" exitCode=0 Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.029754 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftjmr" event={"ID":"9dd22986-8614-4e8c-9a29-6639c5d14f96","Type":"ContainerDied","Data":"ee7c25279f755f3a8977dd66d11b1c3811f94c76bab20858c22f558be9294b19"} Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.029815 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftjmr" event={"ID":"9dd22986-8614-4e8c-9a29-6639c5d14f96","Type":"ContainerStarted","Data":"6bb5c7d0254a00e9e37968102879aa6fb1ac36054c843817774f2a1ae83c0f93"} Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.032456 4810 generic.go:334] "Generic (PLEG): container finished" podID="e0a8fc95-1ea0-4ad4-baea-bb7d91526d27" containerID="5e3b5b05fd190fd39768c2e9e77e605aae085c26f5305e0e1bd5ab65c2bbcfe1" exitCode=0 Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.032502 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27","Type":"ContainerDied","Data":"5e3b5b05fd190fd39768c2e9e77e605aae085c26f5305e0e1bd5ab65c2bbcfe1"} Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.032571 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e0a8fc95-1ea0-4ad4-baea-bb7d91526d27","Type":"ContainerDied","Data":"79619367d732b5bd3614d6d37a21cc45e811f2c1f0cfaa379478b407ae5d8987"} Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.032571 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.032587 4810 scope.go:117] "RemoveContainer" containerID="5e3b5b05fd190fd39768c2e9e77e605aae085c26f5305e0e1bd5ab65c2bbcfe1" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.038527 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31046f73-59af-4c05-976d-65186bb31cec","Type":"ContainerStarted","Data":"08dc9cd0ed85a4bf549a111095b9209661ba67da14fe12de2f749ed800959244"} Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.067491 4810 scope.go:117] "RemoveContainer" containerID="40005d5f28de244bc20dce68453da08d96cd2c26ce5b4d9fbe03d8c1f9e64395" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.073854 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.082911 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.092869 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.093260 4810 scope.go:117] "RemoveContainer" containerID="5e3b5b05fd190fd39768c2e9e77e605aae085c26f5305e0e1bd5ab65c2bbcfe1" Dec 01 14:58:19 crc kubenswrapper[4810]: E1201 14:58:19.098173 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e3b5b05fd190fd39768c2e9e77e605aae085c26f5305e0e1bd5ab65c2bbcfe1\": container with ID starting with 5e3b5b05fd190fd39768c2e9e77e605aae085c26f5305e0e1bd5ab65c2bbcfe1 not found: ID does not exist" containerID="5e3b5b05fd190fd39768c2e9e77e605aae085c26f5305e0e1bd5ab65c2bbcfe1" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.098218 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e3b5b05fd190fd39768c2e9e77e605aae085c26f5305e0e1bd5ab65c2bbcfe1"} err="failed to get container status \"5e3b5b05fd190fd39768c2e9e77e605aae085c26f5305e0e1bd5ab65c2bbcfe1\": rpc error: code = NotFound desc = could not find container \"5e3b5b05fd190fd39768c2e9e77e605aae085c26f5305e0e1bd5ab65c2bbcfe1\": container with ID starting with 5e3b5b05fd190fd39768c2e9e77e605aae085c26f5305e0e1bd5ab65c2bbcfe1 not found: ID does not exist" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.098243 4810 scope.go:117] "RemoveContainer" containerID="40005d5f28de244bc20dce68453da08d96cd2c26ce5b4d9fbe03d8c1f9e64395" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.098330 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 14:58:19 crc kubenswrapper[4810]: E1201 14:58:19.098872 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0a8fc95-1ea0-4ad4-baea-bb7d91526d27" containerName="nova-api-api" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.098890 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0a8fc95-1ea0-4ad4-baea-bb7d91526d27" containerName="nova-api-api" Dec 01 14:58:19 crc kubenswrapper[4810]: E1201 14:58:19.098910 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0a8fc95-1ea0-4ad4-baea-bb7d91526d27" containerName="nova-api-log" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.098917 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0a8fc95-1ea0-4ad4-baea-bb7d91526d27" containerName="nova-api-log" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.099097 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0a8fc95-1ea0-4ad4-baea-bb7d91526d27" containerName="nova-api-api" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.099113 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0a8fc95-1ea0-4ad4-baea-bb7d91526d27" containerName="nova-api-log" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.100034 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: E1201 14:58:19.100416 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40005d5f28de244bc20dce68453da08d96cd2c26ce5b4d9fbe03d8c1f9e64395\": container with ID starting with 40005d5f28de244bc20dce68453da08d96cd2c26ce5b4d9fbe03d8c1f9e64395 not found: ID does not exist" containerID="40005d5f28de244bc20dce68453da08d96cd2c26ce5b4d9fbe03d8c1f9e64395" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.100584 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40005d5f28de244bc20dce68453da08d96cd2c26ce5b4d9fbe03d8c1f9e64395"} err="failed to get container status \"40005d5f28de244bc20dce68453da08d96cd2c26ce5b4d9fbe03d8c1f9e64395\": rpc error: code = NotFound desc = could not find container \"40005d5f28de244bc20dce68453da08d96cd2c26ce5b4d9fbe03d8c1f9e64395\": container with ID starting with 40005d5f28de244bc20dce68453da08d96cd2c26ce5b4d9fbe03d8c1f9e64395 not found: ID does not exist" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.103000 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.103217 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.108505 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.124753 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.233770 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-config-data\") pod \"nova-api-0\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.233926 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pf5vp\" (UniqueName: \"kubernetes.io/projected/e191dc16-6683-424f-8e41-42d91811ddc8-kube-api-access-pf5vp\") pod \"nova-api-0\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.233972 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.234053 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e191dc16-6683-424f-8e41-42d91811ddc8-logs\") pod \"nova-api-0\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.234207 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.234340 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-public-tls-certs\") pod \"nova-api-0\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.301309 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-mc5bd"] Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.302736 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mc5bd" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.304909 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.305591 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.316723 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-mc5bd"] Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.335792 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e191dc16-6683-424f-8e41-42d91811ddc8-logs\") pod \"nova-api-0\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.335857 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.335894 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-public-tls-certs\") pod \"nova-api-0\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.335957 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-config-data\") pod \"nova-api-0\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.335983 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pf5vp\" (UniqueName: \"kubernetes.io/projected/e191dc16-6683-424f-8e41-42d91811ddc8-kube-api-access-pf5vp\") pod \"nova-api-0\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.336008 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.337327 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e191dc16-6683-424f-8e41-42d91811ddc8-logs\") pod \"nova-api-0\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.340021 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.340379 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-public-tls-certs\") pod \"nova-api-0\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.341318 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.350080 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-config-data\") pod \"nova-api-0\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.369089 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pf5vp\" (UniqueName: \"kubernetes.io/projected/e191dc16-6683-424f-8e41-42d91811ddc8-kube-api-access-pf5vp\") pod \"nova-api-0\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.426343 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.437799 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636564e1-d025-4f10-bf66-22e848c21de0-config-data\") pod \"nova-cell1-cell-mapping-mc5bd\" (UID: \"636564e1-d025-4f10-bf66-22e848c21de0\") " pod="openstack/nova-cell1-cell-mapping-mc5bd" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.437891 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636564e1-d025-4f10-bf66-22e848c21de0-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mc5bd\" (UID: \"636564e1-d025-4f10-bf66-22e848c21de0\") " pod="openstack/nova-cell1-cell-mapping-mc5bd" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.437940 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff6hg\" (UniqueName: \"kubernetes.io/projected/636564e1-d025-4f10-bf66-22e848c21de0-kube-api-access-ff6hg\") pod \"nova-cell1-cell-mapping-mc5bd\" (UID: \"636564e1-d025-4f10-bf66-22e848c21de0\") " pod="openstack/nova-cell1-cell-mapping-mc5bd" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.438059 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/636564e1-d025-4f10-bf66-22e848c21de0-scripts\") pod \"nova-cell1-cell-mapping-mc5bd\" (UID: \"636564e1-d025-4f10-bf66-22e848c21de0\") " pod="openstack/nova-cell1-cell-mapping-mc5bd" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.540710 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636564e1-d025-4f10-bf66-22e848c21de0-config-data\") pod \"nova-cell1-cell-mapping-mc5bd\" (UID: \"636564e1-d025-4f10-bf66-22e848c21de0\") " pod="openstack/nova-cell1-cell-mapping-mc5bd" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.541133 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636564e1-d025-4f10-bf66-22e848c21de0-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mc5bd\" (UID: \"636564e1-d025-4f10-bf66-22e848c21de0\") " pod="openstack/nova-cell1-cell-mapping-mc5bd" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.541215 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff6hg\" (UniqueName: \"kubernetes.io/projected/636564e1-d025-4f10-bf66-22e848c21de0-kube-api-access-ff6hg\") pod \"nova-cell1-cell-mapping-mc5bd\" (UID: \"636564e1-d025-4f10-bf66-22e848c21de0\") " pod="openstack/nova-cell1-cell-mapping-mc5bd" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.541397 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/636564e1-d025-4f10-bf66-22e848c21de0-scripts\") pod \"nova-cell1-cell-mapping-mc5bd\" (UID: \"636564e1-d025-4f10-bf66-22e848c21de0\") " pod="openstack/nova-cell1-cell-mapping-mc5bd" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.551948 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636564e1-d025-4f10-bf66-22e848c21de0-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mc5bd\" (UID: \"636564e1-d025-4f10-bf66-22e848c21de0\") " pod="openstack/nova-cell1-cell-mapping-mc5bd" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.552175 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/636564e1-d025-4f10-bf66-22e848c21de0-scripts\") pod \"nova-cell1-cell-mapping-mc5bd\" (UID: \"636564e1-d025-4f10-bf66-22e848c21de0\") " pod="openstack/nova-cell1-cell-mapping-mc5bd" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.552442 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636564e1-d025-4f10-bf66-22e848c21de0-config-data\") pod \"nova-cell1-cell-mapping-mc5bd\" (UID: \"636564e1-d025-4f10-bf66-22e848c21de0\") " pod="openstack/nova-cell1-cell-mapping-mc5bd" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.572225 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff6hg\" (UniqueName: \"kubernetes.io/projected/636564e1-d025-4f10-bf66-22e848c21de0-kube-api-access-ff6hg\") pod \"nova-cell1-cell-mapping-mc5bd\" (UID: \"636564e1-d025-4f10-bf66-22e848c21de0\") " pod="openstack/nova-cell1-cell-mapping-mc5bd" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.635773 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mc5bd" Dec 01 14:58:19 crc kubenswrapper[4810]: I1201 14:58:19.873563 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:58:19 crc kubenswrapper[4810]: W1201 14:58:19.907709 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode191dc16_6683_424f_8e41_42d91811ddc8.slice/crio-872e79c3d7c40eea3a6bdcef0b65b385de166ff978801bbf4cf2a4d31ff5185b WatchSource:0}: Error finding container 872e79c3d7c40eea3a6bdcef0b65b385de166ff978801bbf4cf2a4d31ff5185b: Status 404 returned error can't find the container with id 872e79c3d7c40eea3a6bdcef0b65b385de166ff978801bbf4cf2a4d31ff5185b Dec 01 14:58:20 crc kubenswrapper[4810]: I1201 14:58:20.054713 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31046f73-59af-4c05-976d-65186bb31cec","Type":"ContainerStarted","Data":"d94da3e2e43536ab2044032575c579b392f9637451fe336030187a32c1883b3b"} Dec 01 14:58:20 crc kubenswrapper[4810]: I1201 14:58:20.056150 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e191dc16-6683-424f-8e41-42d91811ddc8","Type":"ContainerStarted","Data":"872e79c3d7c40eea3a6bdcef0b65b385de166ff978801bbf4cf2a4d31ff5185b"} Dec 01 14:58:20 crc kubenswrapper[4810]: I1201 14:58:20.136176 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-mc5bd"] Dec 01 14:58:20 crc kubenswrapper[4810]: I1201 14:58:20.504021 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0a8fc95-1ea0-4ad4-baea-bb7d91526d27" path="/var/lib/kubelet/pods/e0a8fc95-1ea0-4ad4-baea-bb7d91526d27/volumes" Dec 01 14:58:21 crc kubenswrapper[4810]: I1201 14:58:21.067403 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e191dc16-6683-424f-8e41-42d91811ddc8","Type":"ContainerStarted","Data":"3f4880ec489c8f519326fa474e2f7dd774bbd4b59b2fe45ceecbceec0dcf41a7"} Dec 01 14:58:21 crc kubenswrapper[4810]: I1201 14:58:21.067917 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e191dc16-6683-424f-8e41-42d91811ddc8","Type":"ContainerStarted","Data":"7339fbdb3ddbc851f36eff92f6ca5d4b7be7cd91f7ff4f1dfd48304c4ee6d694"} Dec 01 14:58:21 crc kubenswrapper[4810]: I1201 14:58:21.070624 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mc5bd" event={"ID":"636564e1-d025-4f10-bf66-22e848c21de0","Type":"ContainerStarted","Data":"9f4287eaf131c1548b675beb18785d80dde067cda2f647db8236c725db0e0507"} Dec 01 14:58:21 crc kubenswrapper[4810]: I1201 14:58:21.070667 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mc5bd" event={"ID":"636564e1-d025-4f10-bf66-22e848c21de0","Type":"ContainerStarted","Data":"8983a3e50059a6eed414f2a0f38d00c4231a9516a2cd587ae8ee19ca573bc7c8"} Dec 01 14:58:21 crc kubenswrapper[4810]: I1201 14:58:21.073142 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftjmr" event={"ID":"9dd22986-8614-4e8c-9a29-6639c5d14f96","Type":"ContainerStarted","Data":"471e9e622fd0b10824cbe25d1a4cd9adb3b998018442e85b565263caa507b8fe"} Dec 01 14:58:21 crc kubenswrapper[4810]: I1201 14:58:21.075228 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31046f73-59af-4c05-976d-65186bb31cec","Type":"ContainerStarted","Data":"9f977eefdbefbf88215823985ab51d0b45d5ff521f8f4908a3d8f4baf7b8ea03"} Dec 01 14:58:21 crc kubenswrapper[4810]: I1201 14:58:21.099165 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.099147211 podStartE2EDuration="2.099147211s" podCreationTimestamp="2025-12-01 14:58:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:58:21.084726658 +0000 UTC m=+1466.848236271" watchObservedRunningTime="2025-12-01 14:58:21.099147211 +0000 UTC m=+1466.862656814" Dec 01 14:58:21 crc kubenswrapper[4810]: I1201 14:58:21.121871 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-mc5bd" podStartSLOduration=2.121852648 podStartE2EDuration="2.121852648s" podCreationTimestamp="2025-12-01 14:58:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:58:21.116103272 +0000 UTC m=+1466.879612865" watchObservedRunningTime="2025-12-01 14:58:21.121852648 +0000 UTC m=+1466.885362251" Dec 01 14:58:21 crc kubenswrapper[4810]: I1201 14:58:21.459296 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:58:21 crc kubenswrapper[4810]: I1201 14:58:21.520179 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5594d9b959-wrzd5"] Dec 01 14:58:21 crc kubenswrapper[4810]: I1201 14:58:21.520493 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" podUID="af2b1a14-3eb7-41b7-90f7-08c4b93525ec" containerName="dnsmasq-dns" containerID="cri-o://7a2e28f7ee7d517895c84a1fdb6b11d1364cb74ed5a8484dd188be059f519de4" gracePeriod=10 Dec 01 14:58:21 crc kubenswrapper[4810]: I1201 14:58:21.961253 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" podUID="af2b1a14-3eb7-41b7-90f7-08c4b93525ec" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.184:5353: connect: connection refused" Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.087539 4810 generic.go:334] "Generic (PLEG): container finished" podID="af2b1a14-3eb7-41b7-90f7-08c4b93525ec" containerID="7a2e28f7ee7d517895c84a1fdb6b11d1364cb74ed5a8484dd188be059f519de4" exitCode=0 Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.087615 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" event={"ID":"af2b1a14-3eb7-41b7-90f7-08c4b93525ec","Type":"ContainerDied","Data":"7a2e28f7ee7d517895c84a1fdb6b11d1364cb74ed5a8484dd188be059f519de4"} Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.090889 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31046f73-59af-4c05-976d-65186bb31cec","Type":"ContainerStarted","Data":"a15084f9f3b84020b6304d0dab9e5fe86bc99bdcb1432e5ed7e798ae56fba3fd"} Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.091193 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="31046f73-59af-4c05-976d-65186bb31cec" containerName="ceilometer-central-agent" containerID="cri-o://08dc9cd0ed85a4bf549a111095b9209661ba67da14fe12de2f749ed800959244" gracePeriod=30 Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.091382 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.091959 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="31046f73-59af-4c05-976d-65186bb31cec" containerName="proxy-httpd" containerID="cri-o://a15084f9f3b84020b6304d0dab9e5fe86bc99bdcb1432e5ed7e798ae56fba3fd" gracePeriod=30 Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.092055 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="31046f73-59af-4c05-976d-65186bb31cec" containerName="sg-core" containerID="cri-o://9f977eefdbefbf88215823985ab51d0b45d5ff521f8f4908a3d8f4baf7b8ea03" gracePeriod=30 Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.092126 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="31046f73-59af-4c05-976d-65186bb31cec" containerName="ceilometer-notification-agent" containerID="cri-o://d94da3e2e43536ab2044032575c579b392f9637451fe336030187a32c1883b3b" gracePeriod=30 Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.645701 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.666584 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.896406949 podStartE2EDuration="6.666564898s" podCreationTimestamp="2025-12-01 14:58:16 +0000 UTC" firstStartedPulling="2025-12-01 14:58:16.964769328 +0000 UTC m=+1462.728278931" lastFinishedPulling="2025-12-01 14:58:21.734927277 +0000 UTC m=+1467.498436880" observedRunningTime="2025-12-01 14:58:22.122840494 +0000 UTC m=+1467.886350097" watchObservedRunningTime="2025-12-01 14:58:22.666564898 +0000 UTC m=+1468.430074501" Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.743628 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-dns-swift-storage-0\") pod \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.744502 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-config\") pod \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.744538 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-ovsdbserver-nb\") pod \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.744569 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-ovsdbserver-sb\") pod \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.744749 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2j48\" (UniqueName: \"kubernetes.io/projected/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-kube-api-access-d2j48\") pod \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.744824 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-dns-svc\") pod \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\" (UID: \"af2b1a14-3eb7-41b7-90f7-08c4b93525ec\") " Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.750585 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-kube-api-access-d2j48" (OuterVolumeSpecName: "kube-api-access-d2j48") pod "af2b1a14-3eb7-41b7-90f7-08c4b93525ec" (UID: "af2b1a14-3eb7-41b7-90f7-08c4b93525ec"). InnerVolumeSpecName "kube-api-access-d2j48". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.803920 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "af2b1a14-3eb7-41b7-90f7-08c4b93525ec" (UID: "af2b1a14-3eb7-41b7-90f7-08c4b93525ec"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.819215 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "af2b1a14-3eb7-41b7-90f7-08c4b93525ec" (UID: "af2b1a14-3eb7-41b7-90f7-08c4b93525ec"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.819381 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "af2b1a14-3eb7-41b7-90f7-08c4b93525ec" (UID: "af2b1a14-3eb7-41b7-90f7-08c4b93525ec"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.822377 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "af2b1a14-3eb7-41b7-90f7-08c4b93525ec" (UID: "af2b1a14-3eb7-41b7-90f7-08c4b93525ec"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.823689 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-config" (OuterVolumeSpecName: "config") pod "af2b1a14-3eb7-41b7-90f7-08c4b93525ec" (UID: "af2b1a14-3eb7-41b7-90f7-08c4b93525ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.847136 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.847387 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.847499 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.847588 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.847646 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2j48\" (UniqueName: \"kubernetes.io/projected/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-kube-api-access-d2j48\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:22 crc kubenswrapper[4810]: I1201 14:58:22.847887 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af2b1a14-3eb7-41b7-90f7-08c4b93525ec-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:23 crc kubenswrapper[4810]: I1201 14:58:23.105703 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" event={"ID":"af2b1a14-3eb7-41b7-90f7-08c4b93525ec","Type":"ContainerDied","Data":"b16040d54efd5cc544057757aad240a38c99cf6dbaf529b13945b0a84863072d"} Dec 01 14:58:23 crc kubenswrapper[4810]: I1201 14:58:23.105775 4810 scope.go:117] "RemoveContainer" containerID="7a2e28f7ee7d517895c84a1fdb6b11d1364cb74ed5a8484dd188be059f519de4" Dec 01 14:58:23 crc kubenswrapper[4810]: I1201 14:58:23.105951 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5594d9b959-wrzd5" Dec 01 14:58:23 crc kubenswrapper[4810]: I1201 14:58:23.118645 4810 generic.go:334] "Generic (PLEG): container finished" podID="9dd22986-8614-4e8c-9a29-6639c5d14f96" containerID="471e9e622fd0b10824cbe25d1a4cd9adb3b998018442e85b565263caa507b8fe" exitCode=0 Dec 01 14:58:23 crc kubenswrapper[4810]: I1201 14:58:23.118736 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftjmr" event={"ID":"9dd22986-8614-4e8c-9a29-6639c5d14f96","Type":"ContainerDied","Data":"471e9e622fd0b10824cbe25d1a4cd9adb3b998018442e85b565263caa507b8fe"} Dec 01 14:58:23 crc kubenswrapper[4810]: I1201 14:58:23.123888 4810 generic.go:334] "Generic (PLEG): container finished" podID="31046f73-59af-4c05-976d-65186bb31cec" containerID="9f977eefdbefbf88215823985ab51d0b45d5ff521f8f4908a3d8f4baf7b8ea03" exitCode=2 Dec 01 14:58:23 crc kubenswrapper[4810]: I1201 14:58:23.123921 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31046f73-59af-4c05-976d-65186bb31cec","Type":"ContainerDied","Data":"9f977eefdbefbf88215823985ab51d0b45d5ff521f8f4908a3d8f4baf7b8ea03"} Dec 01 14:58:23 crc kubenswrapper[4810]: I1201 14:58:23.156522 4810 scope.go:117] "RemoveContainer" containerID="b21f2a7ba3de1521ea49e1c9aac6a01c8cd63406e9ead0864ac59de0a527ef05" Dec 01 14:58:23 crc kubenswrapper[4810]: I1201 14:58:23.183512 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5594d9b959-wrzd5"] Dec 01 14:58:23 crc kubenswrapper[4810]: I1201 14:58:23.194989 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5594d9b959-wrzd5"] Dec 01 14:58:24 crc kubenswrapper[4810]: I1201 14:58:24.136799 4810 generic.go:334] "Generic (PLEG): container finished" podID="31046f73-59af-4c05-976d-65186bb31cec" containerID="d94da3e2e43536ab2044032575c579b392f9637451fe336030187a32c1883b3b" exitCode=0 Dec 01 14:58:24 crc kubenswrapper[4810]: I1201 14:58:24.136867 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31046f73-59af-4c05-976d-65186bb31cec","Type":"ContainerDied","Data":"d94da3e2e43536ab2044032575c579b392f9637451fe336030187a32c1883b3b"} Dec 01 14:58:24 crc kubenswrapper[4810]: I1201 14:58:24.502197 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af2b1a14-3eb7-41b7-90f7-08c4b93525ec" path="/var/lib/kubelet/pods/af2b1a14-3eb7-41b7-90f7-08c4b93525ec/volumes" Dec 01 14:58:25 crc kubenswrapper[4810]: I1201 14:58:25.161228 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftjmr" event={"ID":"9dd22986-8614-4e8c-9a29-6639c5d14f96","Type":"ContainerStarted","Data":"01f8fd7636f404649e87ff35b78759fd5d3fd98a1e4b1d9e5048c798cc7729fb"} Dec 01 14:58:25 crc kubenswrapper[4810]: I1201 14:58:25.173509 4810 generic.go:334] "Generic (PLEG): container finished" podID="31046f73-59af-4c05-976d-65186bb31cec" containerID="08dc9cd0ed85a4bf549a111095b9209661ba67da14fe12de2f749ed800959244" exitCode=0 Dec 01 14:58:25 crc kubenswrapper[4810]: I1201 14:58:25.173561 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31046f73-59af-4c05-976d-65186bb31cec","Type":"ContainerDied","Data":"08dc9cd0ed85a4bf549a111095b9209661ba67da14fe12de2f749ed800959244"} Dec 01 14:58:25 crc kubenswrapper[4810]: I1201 14:58:25.194621 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ftjmr" podStartSLOduration=3.239314402 podStartE2EDuration="8.194597185s" podCreationTimestamp="2025-12-01 14:58:17 +0000 UTC" firstStartedPulling="2025-12-01 14:58:19.032017876 +0000 UTC m=+1464.795527469" lastFinishedPulling="2025-12-01 14:58:23.987300649 +0000 UTC m=+1469.750810252" observedRunningTime="2025-12-01 14:58:25.182823215 +0000 UTC m=+1470.946332838" watchObservedRunningTime="2025-12-01 14:58:25.194597185 +0000 UTC m=+1470.958106788" Dec 01 14:58:26 crc kubenswrapper[4810]: I1201 14:58:26.183019 4810 generic.go:334] "Generic (PLEG): container finished" podID="636564e1-d025-4f10-bf66-22e848c21de0" containerID="9f4287eaf131c1548b675beb18785d80dde067cda2f647db8236c725db0e0507" exitCode=0 Dec 01 14:58:26 crc kubenswrapper[4810]: I1201 14:58:26.183069 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mc5bd" event={"ID":"636564e1-d025-4f10-bf66-22e848c21de0","Type":"ContainerDied","Data":"9f4287eaf131c1548b675beb18785d80dde067cda2f647db8236c725db0e0507"} Dec 01 14:58:27 crc kubenswrapper[4810]: I1201 14:58:27.511264 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mc5bd" Dec 01 14:58:27 crc kubenswrapper[4810]: I1201 14:58:27.648306 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636564e1-d025-4f10-bf66-22e848c21de0-combined-ca-bundle\") pod \"636564e1-d025-4f10-bf66-22e848c21de0\" (UID: \"636564e1-d025-4f10-bf66-22e848c21de0\") " Dec 01 14:58:27 crc kubenswrapper[4810]: I1201 14:58:27.648385 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636564e1-d025-4f10-bf66-22e848c21de0-config-data\") pod \"636564e1-d025-4f10-bf66-22e848c21de0\" (UID: \"636564e1-d025-4f10-bf66-22e848c21de0\") " Dec 01 14:58:27 crc kubenswrapper[4810]: I1201 14:58:27.648606 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ff6hg\" (UniqueName: \"kubernetes.io/projected/636564e1-d025-4f10-bf66-22e848c21de0-kube-api-access-ff6hg\") pod \"636564e1-d025-4f10-bf66-22e848c21de0\" (UID: \"636564e1-d025-4f10-bf66-22e848c21de0\") " Dec 01 14:58:27 crc kubenswrapper[4810]: I1201 14:58:27.648635 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/636564e1-d025-4f10-bf66-22e848c21de0-scripts\") pod \"636564e1-d025-4f10-bf66-22e848c21de0\" (UID: \"636564e1-d025-4f10-bf66-22e848c21de0\") " Dec 01 14:58:27 crc kubenswrapper[4810]: I1201 14:58:27.654608 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/636564e1-d025-4f10-bf66-22e848c21de0-kube-api-access-ff6hg" (OuterVolumeSpecName: "kube-api-access-ff6hg") pod "636564e1-d025-4f10-bf66-22e848c21de0" (UID: "636564e1-d025-4f10-bf66-22e848c21de0"). InnerVolumeSpecName "kube-api-access-ff6hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:58:27 crc kubenswrapper[4810]: I1201 14:58:27.655013 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/636564e1-d025-4f10-bf66-22e848c21de0-scripts" (OuterVolumeSpecName: "scripts") pod "636564e1-d025-4f10-bf66-22e848c21de0" (UID: "636564e1-d025-4f10-bf66-22e848c21de0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:27 crc kubenswrapper[4810]: I1201 14:58:27.688332 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/636564e1-d025-4f10-bf66-22e848c21de0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "636564e1-d025-4f10-bf66-22e848c21de0" (UID: "636564e1-d025-4f10-bf66-22e848c21de0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:27 crc kubenswrapper[4810]: I1201 14:58:27.701732 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/636564e1-d025-4f10-bf66-22e848c21de0-config-data" (OuterVolumeSpecName: "config-data") pod "636564e1-d025-4f10-bf66-22e848c21de0" (UID: "636564e1-d025-4f10-bf66-22e848c21de0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:27 crc kubenswrapper[4810]: I1201 14:58:27.752010 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636564e1-d025-4f10-bf66-22e848c21de0-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:27 crc kubenswrapper[4810]: I1201 14:58:27.752067 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ff6hg\" (UniqueName: \"kubernetes.io/projected/636564e1-d025-4f10-bf66-22e848c21de0-kube-api-access-ff6hg\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:27 crc kubenswrapper[4810]: I1201 14:58:27.752082 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/636564e1-d025-4f10-bf66-22e848c21de0-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:27 crc kubenswrapper[4810]: I1201 14:58:27.752093 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636564e1-d025-4f10-bf66-22e848c21de0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:27 crc kubenswrapper[4810]: I1201 14:58:27.972646 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ftjmr" Dec 01 14:58:27 crc kubenswrapper[4810]: I1201 14:58:27.972962 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ftjmr" Dec 01 14:58:28 crc kubenswrapper[4810]: I1201 14:58:28.200246 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mc5bd" event={"ID":"636564e1-d025-4f10-bf66-22e848c21de0","Type":"ContainerDied","Data":"8983a3e50059a6eed414f2a0f38d00c4231a9516a2cd587ae8ee19ca573bc7c8"} Dec 01 14:58:28 crc kubenswrapper[4810]: I1201 14:58:28.200280 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8983a3e50059a6eed414f2a0f38d00c4231a9516a2cd587ae8ee19ca573bc7c8" Dec 01 14:58:28 crc kubenswrapper[4810]: I1201 14:58:28.200302 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mc5bd" Dec 01 14:58:28 crc kubenswrapper[4810]: I1201 14:58:28.394393 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 14:58:28 crc kubenswrapper[4810]: I1201 14:58:28.394683 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="96ed5fc8-e0bf-47c4-bc9a-1923178532e9" containerName="nova-scheduler-scheduler" containerID="cri-o://8326409636bfc6c1f41facf074040861e2e7e94c5b8dacc1e80290776fc88a2d" gracePeriod=30 Dec 01 14:58:28 crc kubenswrapper[4810]: I1201 14:58:28.412310 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:58:28 crc kubenswrapper[4810]: I1201 14:58:28.412634 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e191dc16-6683-424f-8e41-42d91811ddc8" containerName="nova-api-log" containerID="cri-o://7339fbdb3ddbc851f36eff92f6ca5d4b7be7cd91f7ff4f1dfd48304c4ee6d694" gracePeriod=30 Dec 01 14:58:28 crc kubenswrapper[4810]: I1201 14:58:28.412660 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e191dc16-6683-424f-8e41-42d91811ddc8" containerName="nova-api-api" containerID="cri-o://3f4880ec489c8f519326fa474e2f7dd774bbd4b59b2fe45ceecbceec0dcf41a7" gracePeriod=30 Dec 01 14:58:28 crc kubenswrapper[4810]: I1201 14:58:28.432488 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:58:28 crc kubenswrapper[4810]: I1201 14:58:28.432772 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fed2100e-43c2-4cef-923b-74ce03403318" containerName="nova-metadata-log" containerID="cri-o://e4e3e6e7ce19455507ba99bbc505942abf9a93bf221f3bdbf08c61912cec851e" gracePeriod=30 Dec 01 14:58:28 crc kubenswrapper[4810]: I1201 14:58:28.432825 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fed2100e-43c2-4cef-923b-74ce03403318" containerName="nova-metadata-metadata" containerID="cri-o://6b1083d4a1c26e507fb0dd8bbb368816fc6e29c3ba96ceb93667d63917c2dce9" gracePeriod=30 Dec 01 14:58:28 crc kubenswrapper[4810]: I1201 14:58:28.971997 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.015414 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ftjmr" podUID="9dd22986-8614-4e8c-9a29-6639c5d14f96" containerName="registry-server" probeResult="failure" output=< Dec 01 14:58:29 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Dec 01 14:58:29 crc kubenswrapper[4810]: > Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.073744 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-combined-ca-bundle\") pod \"e191dc16-6683-424f-8e41-42d91811ddc8\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.073799 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-public-tls-certs\") pod \"e191dc16-6683-424f-8e41-42d91811ddc8\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.073944 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e191dc16-6683-424f-8e41-42d91811ddc8-logs\") pod \"e191dc16-6683-424f-8e41-42d91811ddc8\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.074029 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-internal-tls-certs\") pod \"e191dc16-6683-424f-8e41-42d91811ddc8\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.074072 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-config-data\") pod \"e191dc16-6683-424f-8e41-42d91811ddc8\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.074092 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pf5vp\" (UniqueName: \"kubernetes.io/projected/e191dc16-6683-424f-8e41-42d91811ddc8-kube-api-access-pf5vp\") pod \"e191dc16-6683-424f-8e41-42d91811ddc8\" (UID: \"e191dc16-6683-424f-8e41-42d91811ddc8\") " Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.074513 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e191dc16-6683-424f-8e41-42d91811ddc8-logs" (OuterVolumeSpecName: "logs") pod "e191dc16-6683-424f-8e41-42d91811ddc8" (UID: "e191dc16-6683-424f-8e41-42d91811ddc8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.079022 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e191dc16-6683-424f-8e41-42d91811ddc8-kube-api-access-pf5vp" (OuterVolumeSpecName: "kube-api-access-pf5vp") pod "e191dc16-6683-424f-8e41-42d91811ddc8" (UID: "e191dc16-6683-424f-8e41-42d91811ddc8"). InnerVolumeSpecName "kube-api-access-pf5vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.102118 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e191dc16-6683-424f-8e41-42d91811ddc8" (UID: "e191dc16-6683-424f-8e41-42d91811ddc8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.103784 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-config-data" (OuterVolumeSpecName: "config-data") pod "e191dc16-6683-424f-8e41-42d91811ddc8" (UID: "e191dc16-6683-424f-8e41-42d91811ddc8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.122594 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e191dc16-6683-424f-8e41-42d91811ddc8" (UID: "e191dc16-6683-424f-8e41-42d91811ddc8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.122949 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e191dc16-6683-424f-8e41-42d91811ddc8" (UID: "e191dc16-6683-424f-8e41-42d91811ddc8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:29 crc kubenswrapper[4810]: E1201 14:58:29.135252 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8326409636bfc6c1f41facf074040861e2e7e94c5b8dacc1e80290776fc88a2d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 14:58:29 crc kubenswrapper[4810]: E1201 14:58:29.136581 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8326409636bfc6c1f41facf074040861e2e7e94c5b8dacc1e80290776fc88a2d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 14:58:29 crc kubenswrapper[4810]: E1201 14:58:29.138665 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8326409636bfc6c1f41facf074040861e2e7e94c5b8dacc1e80290776fc88a2d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 14:58:29 crc kubenswrapper[4810]: E1201 14:58:29.138702 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="96ed5fc8-e0bf-47c4-bc9a-1923178532e9" containerName="nova-scheduler-scheduler" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.175771 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.175804 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.175814 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pf5vp\" (UniqueName: \"kubernetes.io/projected/e191dc16-6683-424f-8e41-42d91811ddc8-kube-api-access-pf5vp\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.175824 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.175832 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e191dc16-6683-424f-8e41-42d91811ddc8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.175840 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e191dc16-6683-424f-8e41-42d91811ddc8-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.222019 4810 generic.go:334] "Generic (PLEG): container finished" podID="e191dc16-6683-424f-8e41-42d91811ddc8" containerID="3f4880ec489c8f519326fa474e2f7dd774bbd4b59b2fe45ceecbceec0dcf41a7" exitCode=0 Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.222068 4810 generic.go:334] "Generic (PLEG): container finished" podID="e191dc16-6683-424f-8e41-42d91811ddc8" containerID="7339fbdb3ddbc851f36eff92f6ca5d4b7be7cd91f7ff4f1dfd48304c4ee6d694" exitCode=143 Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.222225 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.222262 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e191dc16-6683-424f-8e41-42d91811ddc8","Type":"ContainerDied","Data":"3f4880ec489c8f519326fa474e2f7dd774bbd4b59b2fe45ceecbceec0dcf41a7"} Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.222305 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e191dc16-6683-424f-8e41-42d91811ddc8","Type":"ContainerDied","Data":"7339fbdb3ddbc851f36eff92f6ca5d4b7be7cd91f7ff4f1dfd48304c4ee6d694"} Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.222330 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e191dc16-6683-424f-8e41-42d91811ddc8","Type":"ContainerDied","Data":"872e79c3d7c40eea3a6bdcef0b65b385de166ff978801bbf4cf2a4d31ff5185b"} Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.222346 4810 scope.go:117] "RemoveContainer" containerID="3f4880ec489c8f519326fa474e2f7dd774bbd4b59b2fe45ceecbceec0dcf41a7" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.225005 4810 generic.go:334] "Generic (PLEG): container finished" podID="fed2100e-43c2-4cef-923b-74ce03403318" containerID="e4e3e6e7ce19455507ba99bbc505942abf9a93bf221f3bdbf08c61912cec851e" exitCode=143 Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.225043 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fed2100e-43c2-4cef-923b-74ce03403318","Type":"ContainerDied","Data":"e4e3e6e7ce19455507ba99bbc505942abf9a93bf221f3bdbf08c61912cec851e"} Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.256654 4810 scope.go:117] "RemoveContainer" containerID="7339fbdb3ddbc851f36eff92f6ca5d4b7be7cd91f7ff4f1dfd48304c4ee6d694" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.277611 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.280409 4810 scope.go:117] "RemoveContainer" containerID="3f4880ec489c8f519326fa474e2f7dd774bbd4b59b2fe45ceecbceec0dcf41a7" Dec 01 14:58:29 crc kubenswrapper[4810]: E1201 14:58:29.280894 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f4880ec489c8f519326fa474e2f7dd774bbd4b59b2fe45ceecbceec0dcf41a7\": container with ID starting with 3f4880ec489c8f519326fa474e2f7dd774bbd4b59b2fe45ceecbceec0dcf41a7 not found: ID does not exist" containerID="3f4880ec489c8f519326fa474e2f7dd774bbd4b59b2fe45ceecbceec0dcf41a7" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.280948 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f4880ec489c8f519326fa474e2f7dd774bbd4b59b2fe45ceecbceec0dcf41a7"} err="failed to get container status \"3f4880ec489c8f519326fa474e2f7dd774bbd4b59b2fe45ceecbceec0dcf41a7\": rpc error: code = NotFound desc = could not find container \"3f4880ec489c8f519326fa474e2f7dd774bbd4b59b2fe45ceecbceec0dcf41a7\": container with ID starting with 3f4880ec489c8f519326fa474e2f7dd774bbd4b59b2fe45ceecbceec0dcf41a7 not found: ID does not exist" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.280984 4810 scope.go:117] "RemoveContainer" containerID="7339fbdb3ddbc851f36eff92f6ca5d4b7be7cd91f7ff4f1dfd48304c4ee6d694" Dec 01 14:58:29 crc kubenswrapper[4810]: E1201 14:58:29.281306 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7339fbdb3ddbc851f36eff92f6ca5d4b7be7cd91f7ff4f1dfd48304c4ee6d694\": container with ID starting with 7339fbdb3ddbc851f36eff92f6ca5d4b7be7cd91f7ff4f1dfd48304c4ee6d694 not found: ID does not exist" containerID="7339fbdb3ddbc851f36eff92f6ca5d4b7be7cd91f7ff4f1dfd48304c4ee6d694" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.281322 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7339fbdb3ddbc851f36eff92f6ca5d4b7be7cd91f7ff4f1dfd48304c4ee6d694"} err="failed to get container status \"7339fbdb3ddbc851f36eff92f6ca5d4b7be7cd91f7ff4f1dfd48304c4ee6d694\": rpc error: code = NotFound desc = could not find container \"7339fbdb3ddbc851f36eff92f6ca5d4b7be7cd91f7ff4f1dfd48304c4ee6d694\": container with ID starting with 7339fbdb3ddbc851f36eff92f6ca5d4b7be7cd91f7ff4f1dfd48304c4ee6d694 not found: ID does not exist" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.281345 4810 scope.go:117] "RemoveContainer" containerID="3f4880ec489c8f519326fa474e2f7dd774bbd4b59b2fe45ceecbceec0dcf41a7" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.281636 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f4880ec489c8f519326fa474e2f7dd774bbd4b59b2fe45ceecbceec0dcf41a7"} err="failed to get container status \"3f4880ec489c8f519326fa474e2f7dd774bbd4b59b2fe45ceecbceec0dcf41a7\": rpc error: code = NotFound desc = could not find container \"3f4880ec489c8f519326fa474e2f7dd774bbd4b59b2fe45ceecbceec0dcf41a7\": container with ID starting with 3f4880ec489c8f519326fa474e2f7dd774bbd4b59b2fe45ceecbceec0dcf41a7 not found: ID does not exist" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.281663 4810 scope.go:117] "RemoveContainer" containerID="7339fbdb3ddbc851f36eff92f6ca5d4b7be7cd91f7ff4f1dfd48304c4ee6d694" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.281956 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7339fbdb3ddbc851f36eff92f6ca5d4b7be7cd91f7ff4f1dfd48304c4ee6d694"} err="failed to get container status \"7339fbdb3ddbc851f36eff92f6ca5d4b7be7cd91f7ff4f1dfd48304c4ee6d694\": rpc error: code = NotFound desc = could not find container \"7339fbdb3ddbc851f36eff92f6ca5d4b7be7cd91f7ff4f1dfd48304c4ee6d694\": container with ID starting with 7339fbdb3ddbc851f36eff92f6ca5d4b7be7cd91f7ff4f1dfd48304c4ee6d694 not found: ID does not exist" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.289555 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.300130 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 14:58:29 crc kubenswrapper[4810]: E1201 14:58:29.300567 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e191dc16-6683-424f-8e41-42d91811ddc8" containerName="nova-api-log" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.300599 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e191dc16-6683-424f-8e41-42d91811ddc8" containerName="nova-api-log" Dec 01 14:58:29 crc kubenswrapper[4810]: E1201 14:58:29.300622 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af2b1a14-3eb7-41b7-90f7-08c4b93525ec" containerName="init" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.300630 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="af2b1a14-3eb7-41b7-90f7-08c4b93525ec" containerName="init" Dec 01 14:58:29 crc kubenswrapper[4810]: E1201 14:58:29.300662 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e191dc16-6683-424f-8e41-42d91811ddc8" containerName="nova-api-api" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.300672 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e191dc16-6683-424f-8e41-42d91811ddc8" containerName="nova-api-api" Dec 01 14:58:29 crc kubenswrapper[4810]: E1201 14:58:29.300687 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af2b1a14-3eb7-41b7-90f7-08c4b93525ec" containerName="dnsmasq-dns" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.300694 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="af2b1a14-3eb7-41b7-90f7-08c4b93525ec" containerName="dnsmasq-dns" Dec 01 14:58:29 crc kubenswrapper[4810]: E1201 14:58:29.300704 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="636564e1-d025-4f10-bf66-22e848c21de0" containerName="nova-manage" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.300712 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="636564e1-d025-4f10-bf66-22e848c21de0" containerName="nova-manage" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.300920 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e191dc16-6683-424f-8e41-42d91811ddc8" containerName="nova-api-log" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.300944 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="636564e1-d025-4f10-bf66-22e848c21de0" containerName="nova-manage" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.300961 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e191dc16-6683-424f-8e41-42d91811ddc8" containerName="nova-api-api" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.300981 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="af2b1a14-3eb7-41b7-90f7-08c4b93525ec" containerName="dnsmasq-dns" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.302238 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.303974 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.305330 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.305498 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.307413 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.380400 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-666rf\" (UniqueName: \"kubernetes.io/projected/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-kube-api-access-666rf\") pod \"nova-api-0\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.380444 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-public-tls-certs\") pod \"nova-api-0\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.380606 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.380634 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-config-data\") pod \"nova-api-0\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.380650 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-logs\") pod \"nova-api-0\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.380664 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.482726 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.483329 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-config-data\") pod \"nova-api-0\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.483381 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-logs\") pod \"nova-api-0\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.483403 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.483485 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-666rf\" (UniqueName: \"kubernetes.io/projected/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-kube-api-access-666rf\") pod \"nova-api-0\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.483525 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-public-tls-certs\") pod \"nova-api-0\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.484419 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-logs\") pod \"nova-api-0\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.489080 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-public-tls-certs\") pod \"nova-api-0\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.489172 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-config-data\") pod \"nova-api-0\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.490257 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.491015 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.501126 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-666rf\" (UniqueName: \"kubernetes.io/projected/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-kube-api-access-666rf\") pod \"nova-api-0\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " pod="openstack/nova-api-0" Dec 01 14:58:29 crc kubenswrapper[4810]: I1201 14:58:29.666508 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 14:58:30 crc kubenswrapper[4810]: W1201 14:58:30.111408 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9acd8c29_93f5_4b0f_8b32_d8f3d576767b.slice/crio-848a6230705ec969b06b50a8f363749de8fa626d9de709eafed5297752f8496d WatchSource:0}: Error finding container 848a6230705ec969b06b50a8f363749de8fa626d9de709eafed5297752f8496d: Status 404 returned error can't find the container with id 848a6230705ec969b06b50a8f363749de8fa626d9de709eafed5297752f8496d Dec 01 14:58:30 crc kubenswrapper[4810]: I1201 14:58:30.113732 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:58:30 crc kubenswrapper[4810]: I1201 14:58:30.237336 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9acd8c29-93f5-4b0f-8b32-d8f3d576767b","Type":"ContainerStarted","Data":"848a6230705ec969b06b50a8f363749de8fa626d9de709eafed5297752f8496d"} Dec 01 14:58:30 crc kubenswrapper[4810]: I1201 14:58:30.503457 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e191dc16-6683-424f-8e41-42d91811ddc8" path="/var/lib/kubelet/pods/e191dc16-6683-424f-8e41-42d91811ddc8/volumes" Dec 01 14:58:31 crc kubenswrapper[4810]: I1201 14:58:31.250073 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9acd8c29-93f5-4b0f-8b32-d8f3d576767b","Type":"ContainerStarted","Data":"1f6241fff3fa641130500b0c5b06297d792485793ad36fdc4dd06993a79f221d"} Dec 01 14:58:31 crc kubenswrapper[4810]: I1201 14:58:31.250125 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9acd8c29-93f5-4b0f-8b32-d8f3d576767b","Type":"ContainerStarted","Data":"6c951f3e299a57aa9f4c575d8d057fd04b80e714ae7050a14a4ca0758c9a0549"} Dec 01 14:58:31 crc kubenswrapper[4810]: I1201 14:58:31.292274 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.292258459 podStartE2EDuration="2.292258459s" podCreationTimestamp="2025-12-01 14:58:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:58:31.280830377 +0000 UTC m=+1477.044340020" watchObservedRunningTime="2025-12-01 14:58:31.292258459 +0000 UTC m=+1477.055768062" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.000297 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.130701 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fed2100e-43c2-4cef-923b-74ce03403318-config-data\") pod \"fed2100e-43c2-4cef-923b-74ce03403318\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.130823 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fed2100e-43c2-4cef-923b-74ce03403318-nova-metadata-tls-certs\") pod \"fed2100e-43c2-4cef-923b-74ce03403318\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.130933 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fed2100e-43c2-4cef-923b-74ce03403318-logs\") pod \"fed2100e-43c2-4cef-923b-74ce03403318\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.130985 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fed2100e-43c2-4cef-923b-74ce03403318-combined-ca-bundle\") pod \"fed2100e-43c2-4cef-923b-74ce03403318\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.131128 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dn44x\" (UniqueName: \"kubernetes.io/projected/fed2100e-43c2-4cef-923b-74ce03403318-kube-api-access-dn44x\") pod \"fed2100e-43c2-4cef-923b-74ce03403318\" (UID: \"fed2100e-43c2-4cef-923b-74ce03403318\") " Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.131604 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fed2100e-43c2-4cef-923b-74ce03403318-logs" (OuterVolumeSpecName: "logs") pod "fed2100e-43c2-4cef-923b-74ce03403318" (UID: "fed2100e-43c2-4cef-923b-74ce03403318"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.131996 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fed2100e-43c2-4cef-923b-74ce03403318-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.137793 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fed2100e-43c2-4cef-923b-74ce03403318-kube-api-access-dn44x" (OuterVolumeSpecName: "kube-api-access-dn44x") pod "fed2100e-43c2-4cef-923b-74ce03403318" (UID: "fed2100e-43c2-4cef-923b-74ce03403318"). InnerVolumeSpecName "kube-api-access-dn44x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.158491 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fed2100e-43c2-4cef-923b-74ce03403318-config-data" (OuterVolumeSpecName: "config-data") pod "fed2100e-43c2-4cef-923b-74ce03403318" (UID: "fed2100e-43c2-4cef-923b-74ce03403318"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.159200 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fed2100e-43c2-4cef-923b-74ce03403318-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fed2100e-43c2-4cef-923b-74ce03403318" (UID: "fed2100e-43c2-4cef-923b-74ce03403318"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.191816 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fed2100e-43c2-4cef-923b-74ce03403318-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "fed2100e-43c2-4cef-923b-74ce03403318" (UID: "fed2100e-43c2-4cef-923b-74ce03403318"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.234090 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fed2100e-43c2-4cef-923b-74ce03403318-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.234502 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dn44x\" (UniqueName: \"kubernetes.io/projected/fed2100e-43c2-4cef-923b-74ce03403318-kube-api-access-dn44x\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.234617 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fed2100e-43c2-4cef-923b-74ce03403318-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.234741 4810 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fed2100e-43c2-4cef-923b-74ce03403318-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.261512 4810 generic.go:334] "Generic (PLEG): container finished" podID="fed2100e-43c2-4cef-923b-74ce03403318" containerID="6b1083d4a1c26e507fb0dd8bbb368816fc6e29c3ba96ceb93667d63917c2dce9" exitCode=0 Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.261745 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fed2100e-43c2-4cef-923b-74ce03403318","Type":"ContainerDied","Data":"6b1083d4a1c26e507fb0dd8bbb368816fc6e29c3ba96ceb93667d63917c2dce9"} Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.262274 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fed2100e-43c2-4cef-923b-74ce03403318","Type":"ContainerDied","Data":"b8111cd0730afc40da67d4d58d1d2dd5235a317acf0be7b876615a1da2e05465"} Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.262301 4810 scope.go:117] "RemoveContainer" containerID="6b1083d4a1c26e507fb0dd8bbb368816fc6e29c3ba96ceb93667d63917c2dce9" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.261770 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.293439 4810 scope.go:117] "RemoveContainer" containerID="e4e3e6e7ce19455507ba99bbc505942abf9a93bf221f3bdbf08c61912cec851e" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.304074 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.318747 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.323177 4810 scope.go:117] "RemoveContainer" containerID="6b1083d4a1c26e507fb0dd8bbb368816fc6e29c3ba96ceb93667d63917c2dce9" Dec 01 14:58:32 crc kubenswrapper[4810]: E1201 14:58:32.323627 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b1083d4a1c26e507fb0dd8bbb368816fc6e29c3ba96ceb93667d63917c2dce9\": container with ID starting with 6b1083d4a1c26e507fb0dd8bbb368816fc6e29c3ba96ceb93667d63917c2dce9 not found: ID does not exist" containerID="6b1083d4a1c26e507fb0dd8bbb368816fc6e29c3ba96ceb93667d63917c2dce9" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.323655 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b1083d4a1c26e507fb0dd8bbb368816fc6e29c3ba96ceb93667d63917c2dce9"} err="failed to get container status \"6b1083d4a1c26e507fb0dd8bbb368816fc6e29c3ba96ceb93667d63917c2dce9\": rpc error: code = NotFound desc = could not find container \"6b1083d4a1c26e507fb0dd8bbb368816fc6e29c3ba96ceb93667d63917c2dce9\": container with ID starting with 6b1083d4a1c26e507fb0dd8bbb368816fc6e29c3ba96ceb93667d63917c2dce9 not found: ID does not exist" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.323677 4810 scope.go:117] "RemoveContainer" containerID="e4e3e6e7ce19455507ba99bbc505942abf9a93bf221f3bdbf08c61912cec851e" Dec 01 14:58:32 crc kubenswrapper[4810]: E1201 14:58:32.323964 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4e3e6e7ce19455507ba99bbc505942abf9a93bf221f3bdbf08c61912cec851e\": container with ID starting with e4e3e6e7ce19455507ba99bbc505942abf9a93bf221f3bdbf08c61912cec851e not found: ID does not exist" containerID="e4e3e6e7ce19455507ba99bbc505942abf9a93bf221f3bdbf08c61912cec851e" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.323983 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4e3e6e7ce19455507ba99bbc505942abf9a93bf221f3bdbf08c61912cec851e"} err="failed to get container status \"e4e3e6e7ce19455507ba99bbc505942abf9a93bf221f3bdbf08c61912cec851e\": rpc error: code = NotFound desc = could not find container \"e4e3e6e7ce19455507ba99bbc505942abf9a93bf221f3bdbf08c61912cec851e\": container with ID starting with e4e3e6e7ce19455507ba99bbc505942abf9a93bf221f3bdbf08c61912cec851e not found: ID does not exist" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.331295 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:58:32 crc kubenswrapper[4810]: E1201 14:58:32.331797 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fed2100e-43c2-4cef-923b-74ce03403318" containerName="nova-metadata-metadata" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.331809 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fed2100e-43c2-4cef-923b-74ce03403318" containerName="nova-metadata-metadata" Dec 01 14:58:32 crc kubenswrapper[4810]: E1201 14:58:32.332087 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fed2100e-43c2-4cef-923b-74ce03403318" containerName="nova-metadata-log" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.332098 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fed2100e-43c2-4cef-923b-74ce03403318" containerName="nova-metadata-log" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.332549 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fed2100e-43c2-4cef-923b-74ce03403318" containerName="nova-metadata-metadata" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.332573 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fed2100e-43c2-4cef-923b-74ce03403318" containerName="nova-metadata-log" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.333574 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.335961 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.336268 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.362369 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.437495 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.437570 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.437704 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-config-data\") pod \"nova-metadata-0\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.437830 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzlvt\" (UniqueName: \"kubernetes.io/projected/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-kube-api-access-hzlvt\") pod \"nova-metadata-0\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.437932 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-logs\") pod \"nova-metadata-0\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.501287 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fed2100e-43c2-4cef-923b-74ce03403318" path="/var/lib/kubelet/pods/fed2100e-43c2-4cef-923b-74ce03403318/volumes" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.540616 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-logs\") pod \"nova-metadata-0\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.540723 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.540749 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.540791 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-config-data\") pod \"nova-metadata-0\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.540828 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzlvt\" (UniqueName: \"kubernetes.io/projected/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-kube-api-access-hzlvt\") pod \"nova-metadata-0\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.541162 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-logs\") pod \"nova-metadata-0\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.544560 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.545174 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-config-data\") pod \"nova-metadata-0\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.545996 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.555385 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzlvt\" (UniqueName: \"kubernetes.io/projected/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-kube-api-access-hzlvt\") pod \"nova-metadata-0\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.660938 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.972898 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.973226 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.973271 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.991250 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5eb750c936738b66d65230eb7997d67c73bd78542b11814d460ae2c7e8d2d0eb"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 14:58:32 crc kubenswrapper[4810]: I1201 14:58:32.991397 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://5eb750c936738b66d65230eb7997d67c73bd78542b11814d460ae2c7e8d2d0eb" gracePeriod=600 Dec 01 14:58:33 crc kubenswrapper[4810]: I1201 14:58:33.112196 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:58:33 crc kubenswrapper[4810]: I1201 14:58:33.276185 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="5eb750c936738b66d65230eb7997d67c73bd78542b11814d460ae2c7e8d2d0eb" exitCode=0 Dec 01 14:58:33 crc kubenswrapper[4810]: I1201 14:58:33.276232 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"5eb750c936738b66d65230eb7997d67c73bd78542b11814d460ae2c7e8d2d0eb"} Dec 01 14:58:33 crc kubenswrapper[4810]: I1201 14:58:33.276577 4810 scope.go:117] "RemoveContainer" containerID="8e1ade99d7746e14166be3755e17c0d1329bcaa748d059179439c1c0ba6fc843" Dec 01 14:58:33 crc kubenswrapper[4810]: I1201 14:58:33.281350 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6","Type":"ContainerStarted","Data":"3e6e63432073a7b17d3fb456f38359b231178a6af7b3213df4e0f311b915c146"} Dec 01 14:58:33 crc kubenswrapper[4810]: I1201 14:58:33.582397 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 14:58:33 crc kubenswrapper[4810]: I1201 14:58:33.659171 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25v7h\" (UniqueName: \"kubernetes.io/projected/96ed5fc8-e0bf-47c4-bc9a-1923178532e9-kube-api-access-25v7h\") pod \"96ed5fc8-e0bf-47c4-bc9a-1923178532e9\" (UID: \"96ed5fc8-e0bf-47c4-bc9a-1923178532e9\") " Dec 01 14:58:33 crc kubenswrapper[4810]: I1201 14:58:33.659237 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96ed5fc8-e0bf-47c4-bc9a-1923178532e9-config-data\") pod \"96ed5fc8-e0bf-47c4-bc9a-1923178532e9\" (UID: \"96ed5fc8-e0bf-47c4-bc9a-1923178532e9\") " Dec 01 14:58:33 crc kubenswrapper[4810]: I1201 14:58:33.659413 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96ed5fc8-e0bf-47c4-bc9a-1923178532e9-combined-ca-bundle\") pod \"96ed5fc8-e0bf-47c4-bc9a-1923178532e9\" (UID: \"96ed5fc8-e0bf-47c4-bc9a-1923178532e9\") " Dec 01 14:58:33 crc kubenswrapper[4810]: I1201 14:58:33.664018 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96ed5fc8-e0bf-47c4-bc9a-1923178532e9-kube-api-access-25v7h" (OuterVolumeSpecName: "kube-api-access-25v7h") pod "96ed5fc8-e0bf-47c4-bc9a-1923178532e9" (UID: "96ed5fc8-e0bf-47c4-bc9a-1923178532e9"). InnerVolumeSpecName "kube-api-access-25v7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:58:33 crc kubenswrapper[4810]: I1201 14:58:33.691648 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96ed5fc8-e0bf-47c4-bc9a-1923178532e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "96ed5fc8-e0bf-47c4-bc9a-1923178532e9" (UID: "96ed5fc8-e0bf-47c4-bc9a-1923178532e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:33 crc kubenswrapper[4810]: I1201 14:58:33.694724 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96ed5fc8-e0bf-47c4-bc9a-1923178532e9-config-data" (OuterVolumeSpecName: "config-data") pod "96ed5fc8-e0bf-47c4-bc9a-1923178532e9" (UID: "96ed5fc8-e0bf-47c4-bc9a-1923178532e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:33 crc kubenswrapper[4810]: I1201 14:58:33.761317 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25v7h\" (UniqueName: \"kubernetes.io/projected/96ed5fc8-e0bf-47c4-bc9a-1923178532e9-kube-api-access-25v7h\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:33 crc kubenswrapper[4810]: I1201 14:58:33.761369 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96ed5fc8-e0bf-47c4-bc9a-1923178532e9-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:33 crc kubenswrapper[4810]: I1201 14:58:33.761381 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96ed5fc8-e0bf-47c4-bc9a-1923178532e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.290567 4810 generic.go:334] "Generic (PLEG): container finished" podID="96ed5fc8-e0bf-47c4-bc9a-1923178532e9" containerID="8326409636bfc6c1f41facf074040861e2e7e94c5b8dacc1e80290776fc88a2d" exitCode=0 Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.290626 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.290619 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"96ed5fc8-e0bf-47c4-bc9a-1923178532e9","Type":"ContainerDied","Data":"8326409636bfc6c1f41facf074040861e2e7e94c5b8dacc1e80290776fc88a2d"} Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.291089 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"96ed5fc8-e0bf-47c4-bc9a-1923178532e9","Type":"ContainerDied","Data":"0057b321702147c9cd4510d0d3fd872015d5ce6f08aaf4573f4e946c513d11c4"} Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.291109 4810 scope.go:117] "RemoveContainer" containerID="8326409636bfc6c1f41facf074040861e2e7e94c5b8dacc1e80290776fc88a2d" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.294140 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef"} Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.298450 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6","Type":"ContainerStarted","Data":"08362c9d20e6c5ffc7ef2c58a68a05a6da9d37da0f16242f1435dc93d6db4209"} Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.298546 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6","Type":"ContainerStarted","Data":"97cf7e0548b92abb2ae87397174ade605222b8ac75fdc9ae22ab5abf9a072ec8"} Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.314562 4810 scope.go:117] "RemoveContainer" containerID="8326409636bfc6c1f41facf074040861e2e7e94c5b8dacc1e80290776fc88a2d" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.317776 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.31775499 podStartE2EDuration="2.31775499s" podCreationTimestamp="2025-12-01 14:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:58:34.31588106 +0000 UTC m=+1480.079390663" watchObservedRunningTime="2025-12-01 14:58:34.31775499 +0000 UTC m=+1480.081264593" Dec 01 14:58:34 crc kubenswrapper[4810]: E1201 14:58:34.318396 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8326409636bfc6c1f41facf074040861e2e7e94c5b8dacc1e80290776fc88a2d\": container with ID starting with 8326409636bfc6c1f41facf074040861e2e7e94c5b8dacc1e80290776fc88a2d not found: ID does not exist" containerID="8326409636bfc6c1f41facf074040861e2e7e94c5b8dacc1e80290776fc88a2d" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.318513 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8326409636bfc6c1f41facf074040861e2e7e94c5b8dacc1e80290776fc88a2d"} err="failed to get container status \"8326409636bfc6c1f41facf074040861e2e7e94c5b8dacc1e80290776fc88a2d\": rpc error: code = NotFound desc = could not find container \"8326409636bfc6c1f41facf074040861e2e7e94c5b8dacc1e80290776fc88a2d\": container with ID starting with 8326409636bfc6c1f41facf074040861e2e7e94c5b8dacc1e80290776fc88a2d not found: ID does not exist" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.371369 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.381035 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.400084 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 14:58:34 crc kubenswrapper[4810]: E1201 14:58:34.400548 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96ed5fc8-e0bf-47c4-bc9a-1923178532e9" containerName="nova-scheduler-scheduler" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.400568 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="96ed5fc8-e0bf-47c4-bc9a-1923178532e9" containerName="nova-scheduler-scheduler" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.400821 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="96ed5fc8-e0bf-47c4-bc9a-1923178532e9" containerName="nova-scheduler-scheduler" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.401453 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.403935 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.408588 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.471816 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzr6m\" (UniqueName: \"kubernetes.io/projected/444ec27a-77d0-4b01-b59e-c6590c46a0c2-kube-api-access-fzr6m\") pod \"nova-scheduler-0\" (UID: \"444ec27a-77d0-4b01-b59e-c6590c46a0c2\") " pod="openstack/nova-scheduler-0" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.471943 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/444ec27a-77d0-4b01-b59e-c6590c46a0c2-config-data\") pod \"nova-scheduler-0\" (UID: \"444ec27a-77d0-4b01-b59e-c6590c46a0c2\") " pod="openstack/nova-scheduler-0" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.471991 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/444ec27a-77d0-4b01-b59e-c6590c46a0c2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"444ec27a-77d0-4b01-b59e-c6590c46a0c2\") " pod="openstack/nova-scheduler-0" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.501729 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96ed5fc8-e0bf-47c4-bc9a-1923178532e9" path="/var/lib/kubelet/pods/96ed5fc8-e0bf-47c4-bc9a-1923178532e9/volumes" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.573659 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/444ec27a-77d0-4b01-b59e-c6590c46a0c2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"444ec27a-77d0-4b01-b59e-c6590c46a0c2\") " pod="openstack/nova-scheduler-0" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.574111 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzr6m\" (UniqueName: \"kubernetes.io/projected/444ec27a-77d0-4b01-b59e-c6590c46a0c2-kube-api-access-fzr6m\") pod \"nova-scheduler-0\" (UID: \"444ec27a-77d0-4b01-b59e-c6590c46a0c2\") " pod="openstack/nova-scheduler-0" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.574337 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/444ec27a-77d0-4b01-b59e-c6590c46a0c2-config-data\") pod \"nova-scheduler-0\" (UID: \"444ec27a-77d0-4b01-b59e-c6590c46a0c2\") " pod="openstack/nova-scheduler-0" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.580736 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/444ec27a-77d0-4b01-b59e-c6590c46a0c2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"444ec27a-77d0-4b01-b59e-c6590c46a0c2\") " pod="openstack/nova-scheduler-0" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.585176 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/444ec27a-77d0-4b01-b59e-c6590c46a0c2-config-data\") pod \"nova-scheduler-0\" (UID: \"444ec27a-77d0-4b01-b59e-c6590c46a0c2\") " pod="openstack/nova-scheduler-0" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.603119 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzr6m\" (UniqueName: \"kubernetes.io/projected/444ec27a-77d0-4b01-b59e-c6590c46a0c2-kube-api-access-fzr6m\") pod \"nova-scheduler-0\" (UID: \"444ec27a-77d0-4b01-b59e-c6590c46a0c2\") " pod="openstack/nova-scheduler-0" Dec 01 14:58:34 crc kubenswrapper[4810]: I1201 14:58:34.719707 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 14:58:35 crc kubenswrapper[4810]: I1201 14:58:35.200602 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 14:58:35 crc kubenswrapper[4810]: I1201 14:58:35.308242 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"444ec27a-77d0-4b01-b59e-c6590c46a0c2","Type":"ContainerStarted","Data":"a8047f5795ca6b6c992b59d4f4a759af0edb049e712a524429d10a5d1fdbc17f"} Dec 01 14:58:36 crc kubenswrapper[4810]: I1201 14:58:36.319988 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"444ec27a-77d0-4b01-b59e-c6590c46a0c2","Type":"ContainerStarted","Data":"ffc643854f69e1e2f23a40254dbae2944125c570cca2c9e027760a19987124a6"} Dec 01 14:58:36 crc kubenswrapper[4810]: I1201 14:58:36.344808 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.344787665 podStartE2EDuration="2.344787665s" podCreationTimestamp="2025-12-01 14:58:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:58:36.342237145 +0000 UTC m=+1482.105746738" watchObservedRunningTime="2025-12-01 14:58:36.344787665 +0000 UTC m=+1482.108297268" Dec 01 14:58:37 crc kubenswrapper[4810]: I1201 14:58:37.661617 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 14:58:37 crc kubenswrapper[4810]: I1201 14:58:37.662061 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 14:58:38 crc kubenswrapper[4810]: I1201 14:58:38.013575 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ftjmr" Dec 01 14:58:38 crc kubenswrapper[4810]: I1201 14:58:38.062817 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ftjmr" Dec 01 14:58:38 crc kubenswrapper[4810]: I1201 14:58:38.251258 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ftjmr"] Dec 01 14:58:39 crc kubenswrapper[4810]: I1201 14:58:39.345508 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ftjmr" podUID="9dd22986-8614-4e8c-9a29-6639c5d14f96" containerName="registry-server" containerID="cri-o://01f8fd7636f404649e87ff35b78759fd5d3fd98a1e4b1d9e5048c798cc7729fb" gracePeriod=2 Dec 01 14:58:39 crc kubenswrapper[4810]: I1201 14:58:39.667762 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 14:58:39 crc kubenswrapper[4810]: I1201 14:58:39.668130 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 14:58:39 crc kubenswrapper[4810]: I1201 14:58:39.720792 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 14:58:39 crc kubenswrapper[4810]: I1201 14:58:39.802965 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ftjmr" Dec 01 14:58:39 crc kubenswrapper[4810]: I1201 14:58:39.928673 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dd22986-8614-4e8c-9a29-6639c5d14f96-utilities\") pod \"9dd22986-8614-4e8c-9a29-6639c5d14f96\" (UID: \"9dd22986-8614-4e8c-9a29-6639c5d14f96\") " Dec 01 14:58:39 crc kubenswrapper[4810]: I1201 14:58:39.928776 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmcmm\" (UniqueName: \"kubernetes.io/projected/9dd22986-8614-4e8c-9a29-6639c5d14f96-kube-api-access-gmcmm\") pod \"9dd22986-8614-4e8c-9a29-6639c5d14f96\" (UID: \"9dd22986-8614-4e8c-9a29-6639c5d14f96\") " Dec 01 14:58:39 crc kubenswrapper[4810]: I1201 14:58:39.928826 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dd22986-8614-4e8c-9a29-6639c5d14f96-catalog-content\") pod \"9dd22986-8614-4e8c-9a29-6639c5d14f96\" (UID: \"9dd22986-8614-4e8c-9a29-6639c5d14f96\") " Dec 01 14:58:39 crc kubenswrapper[4810]: I1201 14:58:39.929591 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dd22986-8614-4e8c-9a29-6639c5d14f96-utilities" (OuterVolumeSpecName: "utilities") pod "9dd22986-8614-4e8c-9a29-6639c5d14f96" (UID: "9dd22986-8614-4e8c-9a29-6639c5d14f96"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:58:39 crc kubenswrapper[4810]: I1201 14:58:39.936071 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dd22986-8614-4e8c-9a29-6639c5d14f96-kube-api-access-gmcmm" (OuterVolumeSpecName: "kube-api-access-gmcmm") pod "9dd22986-8614-4e8c-9a29-6639c5d14f96" (UID: "9dd22986-8614-4e8c-9a29-6639c5d14f96"). InnerVolumeSpecName "kube-api-access-gmcmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.031147 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dd22986-8614-4e8c-9a29-6639c5d14f96-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.031181 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmcmm\" (UniqueName: \"kubernetes.io/projected/9dd22986-8614-4e8c-9a29-6639c5d14f96-kube-api-access-gmcmm\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.046732 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dd22986-8614-4e8c-9a29-6639c5d14f96-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9dd22986-8614-4e8c-9a29-6639c5d14f96" (UID: "9dd22986-8614-4e8c-9a29-6639c5d14f96"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.132418 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dd22986-8614-4e8c-9a29-6639c5d14f96-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.356320 4810 generic.go:334] "Generic (PLEG): container finished" podID="9dd22986-8614-4e8c-9a29-6639c5d14f96" containerID="01f8fd7636f404649e87ff35b78759fd5d3fd98a1e4b1d9e5048c798cc7729fb" exitCode=0 Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.356362 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftjmr" event={"ID":"9dd22986-8614-4e8c-9a29-6639c5d14f96","Type":"ContainerDied","Data":"01f8fd7636f404649e87ff35b78759fd5d3fd98a1e4b1d9e5048c798cc7729fb"} Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.356388 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftjmr" event={"ID":"9dd22986-8614-4e8c-9a29-6639c5d14f96","Type":"ContainerDied","Data":"6bb5c7d0254a00e9e37968102879aa6fb1ac36054c843817774f2a1ae83c0f93"} Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.356408 4810 scope.go:117] "RemoveContainer" containerID="01f8fd7636f404649e87ff35b78759fd5d3fd98a1e4b1d9e5048c798cc7729fb" Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.356436 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ftjmr" Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.381328 4810 scope.go:117] "RemoveContainer" containerID="471e9e622fd0b10824cbe25d1a4cd9adb3b998018442e85b565263caa507b8fe" Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.397220 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ftjmr"] Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.407332 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ftjmr"] Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.420641 4810 scope.go:117] "RemoveContainer" containerID="ee7c25279f755f3a8977dd66d11b1c3811f94c76bab20858c22f558be9294b19" Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.451422 4810 scope.go:117] "RemoveContainer" containerID="01f8fd7636f404649e87ff35b78759fd5d3fd98a1e4b1d9e5048c798cc7729fb" Dec 01 14:58:40 crc kubenswrapper[4810]: E1201 14:58:40.452027 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01f8fd7636f404649e87ff35b78759fd5d3fd98a1e4b1d9e5048c798cc7729fb\": container with ID starting with 01f8fd7636f404649e87ff35b78759fd5d3fd98a1e4b1d9e5048c798cc7729fb not found: ID does not exist" containerID="01f8fd7636f404649e87ff35b78759fd5d3fd98a1e4b1d9e5048c798cc7729fb" Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.452192 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01f8fd7636f404649e87ff35b78759fd5d3fd98a1e4b1d9e5048c798cc7729fb"} err="failed to get container status \"01f8fd7636f404649e87ff35b78759fd5d3fd98a1e4b1d9e5048c798cc7729fb\": rpc error: code = NotFound desc = could not find container \"01f8fd7636f404649e87ff35b78759fd5d3fd98a1e4b1d9e5048c798cc7729fb\": container with ID starting with 01f8fd7636f404649e87ff35b78759fd5d3fd98a1e4b1d9e5048c798cc7729fb not found: ID does not exist" Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.452272 4810 scope.go:117] "RemoveContainer" containerID="471e9e622fd0b10824cbe25d1a4cd9adb3b998018442e85b565263caa507b8fe" Dec 01 14:58:40 crc kubenswrapper[4810]: E1201 14:58:40.452766 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"471e9e622fd0b10824cbe25d1a4cd9adb3b998018442e85b565263caa507b8fe\": container with ID starting with 471e9e622fd0b10824cbe25d1a4cd9adb3b998018442e85b565263caa507b8fe not found: ID does not exist" containerID="471e9e622fd0b10824cbe25d1a4cd9adb3b998018442e85b565263caa507b8fe" Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.452824 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"471e9e622fd0b10824cbe25d1a4cd9adb3b998018442e85b565263caa507b8fe"} err="failed to get container status \"471e9e622fd0b10824cbe25d1a4cd9adb3b998018442e85b565263caa507b8fe\": rpc error: code = NotFound desc = could not find container \"471e9e622fd0b10824cbe25d1a4cd9adb3b998018442e85b565263caa507b8fe\": container with ID starting with 471e9e622fd0b10824cbe25d1a4cd9adb3b998018442e85b565263caa507b8fe not found: ID does not exist" Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.452859 4810 scope.go:117] "RemoveContainer" containerID="ee7c25279f755f3a8977dd66d11b1c3811f94c76bab20858c22f558be9294b19" Dec 01 14:58:40 crc kubenswrapper[4810]: E1201 14:58:40.453290 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee7c25279f755f3a8977dd66d11b1c3811f94c76bab20858c22f558be9294b19\": container with ID starting with ee7c25279f755f3a8977dd66d11b1c3811f94c76bab20858c22f558be9294b19 not found: ID does not exist" containerID="ee7c25279f755f3a8977dd66d11b1c3811f94c76bab20858c22f558be9294b19" Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.453324 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee7c25279f755f3a8977dd66d11b1c3811f94c76bab20858c22f558be9294b19"} err="failed to get container status \"ee7c25279f755f3a8977dd66d11b1c3811f94c76bab20858c22f558be9294b19\": rpc error: code = NotFound desc = could not find container \"ee7c25279f755f3a8977dd66d11b1c3811f94c76bab20858c22f558be9294b19\": container with ID starting with ee7c25279f755f3a8977dd66d11b1c3811f94c76bab20858c22f558be9294b19 not found: ID does not exist" Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.505051 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dd22986-8614-4e8c-9a29-6639c5d14f96" path="/var/lib/kubelet/pods/9dd22986-8614-4e8c-9a29-6639c5d14f96/volumes" Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.684787 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9acd8c29-93f5-4b0f-8b32-d8f3d576767b" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 14:58:40 crc kubenswrapper[4810]: I1201 14:58:40.684842 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9acd8c29-93f5-4b0f-8b32-d8f3d576767b" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 14:58:42 crc kubenswrapper[4810]: I1201 14:58:42.661348 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 14:58:42 crc kubenswrapper[4810]: I1201 14:58:42.661945 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 14:58:43 crc kubenswrapper[4810]: I1201 14:58:43.680772 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 14:58:43 crc kubenswrapper[4810]: I1201 14:58:43.680881 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 14:58:44 crc kubenswrapper[4810]: I1201 14:58:44.720199 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 14:58:44 crc kubenswrapper[4810]: I1201 14:58:44.749864 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 14:58:45 crc kubenswrapper[4810]: I1201 14:58:45.476483 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 14:58:46 crc kubenswrapper[4810]: I1201 14:58:46.454732 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="31046f73-59af-4c05-976d-65186bb31cec" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 01 14:58:49 crc kubenswrapper[4810]: I1201 14:58:49.674350 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 14:58:49 crc kubenswrapper[4810]: I1201 14:58:49.674739 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 14:58:49 crc kubenswrapper[4810]: I1201 14:58:49.675035 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 14:58:49 crc kubenswrapper[4810]: I1201 14:58:49.675079 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 14:58:49 crc kubenswrapper[4810]: I1201 14:58:49.680952 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 14:58:49 crc kubenswrapper[4810]: I1201 14:58:49.682484 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.488238 4810 generic.go:334] "Generic (PLEG): container finished" podID="31046f73-59af-4c05-976d-65186bb31cec" containerID="a15084f9f3b84020b6304d0dab9e5fe86bc99bdcb1432e5ed7e798ae56fba3fd" exitCode=137 Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.488348 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31046f73-59af-4c05-976d-65186bb31cec","Type":"ContainerDied","Data":"a15084f9f3b84020b6304d0dab9e5fe86bc99bdcb1432e5ed7e798ae56fba3fd"} Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.489050 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31046f73-59af-4c05-976d-65186bb31cec","Type":"ContainerDied","Data":"43b85130635d6b96e761d00728f40ebdd36f1363acb56976255e47235f6eeb05"} Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.489068 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43b85130635d6b96e761d00728f40ebdd36f1363acb56976255e47235f6eeb05" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.521291 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.631994 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-sg-core-conf-yaml\") pod \"31046f73-59af-4c05-976d-65186bb31cec\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.632151 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-config-data\") pod \"31046f73-59af-4c05-976d-65186bb31cec\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.632210 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-scripts\") pod \"31046f73-59af-4c05-976d-65186bb31cec\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.632304 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-combined-ca-bundle\") pod \"31046f73-59af-4c05-976d-65186bb31cec\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.632334 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxxk8\" (UniqueName: \"kubernetes.io/projected/31046f73-59af-4c05-976d-65186bb31cec-kube-api-access-mxxk8\") pod \"31046f73-59af-4c05-976d-65186bb31cec\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.632356 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31046f73-59af-4c05-976d-65186bb31cec-run-httpd\") pod \"31046f73-59af-4c05-976d-65186bb31cec\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.632375 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31046f73-59af-4c05-976d-65186bb31cec-log-httpd\") pod \"31046f73-59af-4c05-976d-65186bb31cec\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.632396 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-ceilometer-tls-certs\") pod \"31046f73-59af-4c05-976d-65186bb31cec\" (UID: \"31046f73-59af-4c05-976d-65186bb31cec\") " Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.633291 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31046f73-59af-4c05-976d-65186bb31cec-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "31046f73-59af-4c05-976d-65186bb31cec" (UID: "31046f73-59af-4c05-976d-65186bb31cec"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.633301 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31046f73-59af-4c05-976d-65186bb31cec-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "31046f73-59af-4c05-976d-65186bb31cec" (UID: "31046f73-59af-4c05-976d-65186bb31cec"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.637608 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-scripts" (OuterVolumeSpecName: "scripts") pod "31046f73-59af-4c05-976d-65186bb31cec" (UID: "31046f73-59af-4c05-976d-65186bb31cec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.637929 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31046f73-59af-4c05-976d-65186bb31cec-kube-api-access-mxxk8" (OuterVolumeSpecName: "kube-api-access-mxxk8") pod "31046f73-59af-4c05-976d-65186bb31cec" (UID: "31046f73-59af-4c05-976d-65186bb31cec"). InnerVolumeSpecName "kube-api-access-mxxk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.664017 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "31046f73-59af-4c05-976d-65186bb31cec" (UID: "31046f73-59af-4c05-976d-65186bb31cec"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.673859 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.674309 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.680327 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.715958 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "31046f73-59af-4c05-976d-65186bb31cec" (UID: "31046f73-59af-4c05-976d-65186bb31cec"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.717374 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31046f73-59af-4c05-976d-65186bb31cec" (UID: "31046f73-59af-4c05-976d-65186bb31cec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.736052 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.736090 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.736099 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.736108 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxxk8\" (UniqueName: \"kubernetes.io/projected/31046f73-59af-4c05-976d-65186bb31cec-kube-api-access-mxxk8\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.736117 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31046f73-59af-4c05-976d-65186bb31cec-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.736124 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31046f73-59af-4c05-976d-65186bb31cec-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.736141 4810 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.752972 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-config-data" (OuterVolumeSpecName: "config-data") pod "31046f73-59af-4c05-976d-65186bb31cec" (UID: "31046f73-59af-4c05-976d-65186bb31cec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:58:52 crc kubenswrapper[4810]: I1201 14:58:52.837959 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31046f73-59af-4c05-976d-65186bb31cec-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.497083 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.505294 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.577823 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.591905 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.599406 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:58:53 crc kubenswrapper[4810]: E1201 14:58:53.599878 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd22986-8614-4e8c-9a29-6639c5d14f96" containerName="registry-server" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.599896 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd22986-8614-4e8c-9a29-6639c5d14f96" containerName="registry-server" Dec 01 14:58:53 crc kubenswrapper[4810]: E1201 14:58:53.599909 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31046f73-59af-4c05-976d-65186bb31cec" containerName="sg-core" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.599916 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="31046f73-59af-4c05-976d-65186bb31cec" containerName="sg-core" Dec 01 14:58:53 crc kubenswrapper[4810]: E1201 14:58:53.599928 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31046f73-59af-4c05-976d-65186bb31cec" containerName="ceilometer-notification-agent" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.599947 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="31046f73-59af-4c05-976d-65186bb31cec" containerName="ceilometer-notification-agent" Dec 01 14:58:53 crc kubenswrapper[4810]: E1201 14:58:53.599968 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd22986-8614-4e8c-9a29-6639c5d14f96" containerName="extract-utilities" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.599974 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd22986-8614-4e8c-9a29-6639c5d14f96" containerName="extract-utilities" Dec 01 14:58:53 crc kubenswrapper[4810]: E1201 14:58:53.599985 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31046f73-59af-4c05-976d-65186bb31cec" containerName="proxy-httpd" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.599990 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="31046f73-59af-4c05-976d-65186bb31cec" containerName="proxy-httpd" Dec 01 14:58:53 crc kubenswrapper[4810]: E1201 14:58:53.600007 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd22986-8614-4e8c-9a29-6639c5d14f96" containerName="extract-content" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.600017 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd22986-8614-4e8c-9a29-6639c5d14f96" containerName="extract-content" Dec 01 14:58:53 crc kubenswrapper[4810]: E1201 14:58:53.600048 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31046f73-59af-4c05-976d-65186bb31cec" containerName="ceilometer-central-agent" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.600057 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="31046f73-59af-4c05-976d-65186bb31cec" containerName="ceilometer-central-agent" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.600228 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="31046f73-59af-4c05-976d-65186bb31cec" containerName="sg-core" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.600241 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="31046f73-59af-4c05-976d-65186bb31cec" containerName="ceilometer-central-agent" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.600256 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="31046f73-59af-4c05-976d-65186bb31cec" containerName="ceilometer-notification-agent" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.600267 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="31046f73-59af-4c05-976d-65186bb31cec" containerName="proxy-httpd" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.600273 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dd22986-8614-4e8c-9a29-6639c5d14f96" containerName="registry-server" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.601838 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.605828 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.606953 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.607221 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.623555 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.652420 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-run-httpd\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.652573 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.652720 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-log-httpd\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.652780 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-config-data\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.652842 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.652939 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-scripts\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.652978 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrlvv\" (UniqueName: \"kubernetes.io/projected/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-kube-api-access-zrlvv\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.653031 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.754219 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.754287 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-scripts\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.754315 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrlvv\" (UniqueName: \"kubernetes.io/projected/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-kube-api-access-zrlvv\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.754331 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.754371 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-run-httpd\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.754416 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.754448 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-log-httpd\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.754464 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-config-data\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.755287 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-run-httpd\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.755611 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-log-httpd\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.759420 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.759585 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-scripts\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.760128 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-config-data\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.760204 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.761227 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.775284 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrlvv\" (UniqueName: \"kubernetes.io/projected/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-kube-api-access-zrlvv\") pod \"ceilometer-0\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " pod="openstack/ceilometer-0" Dec 01 14:58:53 crc kubenswrapper[4810]: I1201 14:58:53.931407 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:58:54 crc kubenswrapper[4810]: I1201 14:58:54.420991 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:58:54 crc kubenswrapper[4810]: W1201 14:58:54.425738 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc32cbe3a_7d7b_44b9_a3a1_945dfaf765b3.slice/crio-9516a9dad515dab02cb19edfecb587a3cf75167d1413868d1ee522450e715d7b WatchSource:0}: Error finding container 9516a9dad515dab02cb19edfecb587a3cf75167d1413868d1ee522450e715d7b: Status 404 returned error can't find the container with id 9516a9dad515dab02cb19edfecb587a3cf75167d1413868d1ee522450e715d7b Dec 01 14:58:54 crc kubenswrapper[4810]: I1201 14:58:54.509600 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31046f73-59af-4c05-976d-65186bb31cec" path="/var/lib/kubelet/pods/31046f73-59af-4c05-976d-65186bb31cec/volumes" Dec 01 14:58:54 crc kubenswrapper[4810]: I1201 14:58:54.513530 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3","Type":"ContainerStarted","Data":"9516a9dad515dab02cb19edfecb587a3cf75167d1413868d1ee522450e715d7b"} Dec 01 14:58:56 crc kubenswrapper[4810]: I1201 14:58:56.539917 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3","Type":"ContainerStarted","Data":"b1b79938d10b71fee29a6515a87d1fb415850230bcaa5ce800b1e5ba6bb61661"} Dec 01 14:58:57 crc kubenswrapper[4810]: I1201 14:58:57.553545 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3","Type":"ContainerStarted","Data":"3cb0e1b4c05926fcfa43bb4c227b62243d1a9c4234879cdb02f9c40af684935e"} Dec 01 14:58:58 crc kubenswrapper[4810]: I1201 14:58:58.577049 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3","Type":"ContainerStarted","Data":"90f166a6125c50147c3cb78e63b7459cc98d8a4218a7a6ddd44ce0fb9b8a4fc2"} Dec 01 14:58:59 crc kubenswrapper[4810]: I1201 14:58:59.588549 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3","Type":"ContainerStarted","Data":"ac4ab87e1191fa810eaa9e1774fc31d9ed1e7d6475958ea936337c2a47b7b3e0"} Dec 01 14:58:59 crc kubenswrapper[4810]: I1201 14:58:59.589852 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 14:58:59 crc kubenswrapper[4810]: I1201 14:58:59.619883 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.872118846 podStartE2EDuration="6.619863666s" podCreationTimestamp="2025-12-01 14:58:53 +0000 UTC" firstStartedPulling="2025-12-01 14:58:54.429485011 +0000 UTC m=+1500.192994614" lastFinishedPulling="2025-12-01 14:58:59.177229831 +0000 UTC m=+1504.940739434" observedRunningTime="2025-12-01 14:58:59.618248332 +0000 UTC m=+1505.381757945" watchObservedRunningTime="2025-12-01 14:58:59.619863666 +0000 UTC m=+1505.383373269" Dec 01 14:59:23 crc kubenswrapper[4810]: I1201 14:59:23.938763 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 14:59:43 crc kubenswrapper[4810]: I1201 14:59:43.844412 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 01 14:59:43 crc kubenswrapper[4810]: I1201 14:59:43.845199 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="c35fc22d-59ab-4396-87c5-fadfc17ecc53" containerName="openstackclient" containerID="cri-o://a303d0364f2f0ad764ac7d54b67330e695ac583a1627abd27d1f6447c9de1ee2" gracePeriod=2 Dec 01 14:59:43 crc kubenswrapper[4810]: I1201 14:59:43.862985 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 01 14:59:43 crc kubenswrapper[4810]: I1201 14:59:43.998272 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.137568 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance324e-account-delete-6s4dm"] Dec 01 14:59:44 crc kubenswrapper[4810]: E1201 14:59:44.139805 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c35fc22d-59ab-4396-87c5-fadfc17ecc53" containerName="openstackclient" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.139848 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c35fc22d-59ab-4396-87c5-fadfc17ecc53" containerName="openstackclient" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.140144 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c35fc22d-59ab-4396-87c5-fadfc17ecc53" containerName="openstackclient" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.141063 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance324e-account-delete-6s4dm" Dec 01 14:59:44 crc kubenswrapper[4810]: E1201 14:59:44.157051 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 01 14:59:44 crc kubenswrapper[4810]: E1201 14:59:44.157105 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-config-data podName:3b81b6a9-0be6-4ce1-81b4-043edb1c221f nodeName:}" failed. No retries permitted until 2025-12-01 14:59:44.65708971 +0000 UTC m=+1550.420599313 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-config-data") pod "rabbitmq-server-0" (UID: "3b81b6a9-0be6-4ce1-81b4-043edb1c221f") : configmap "rabbitmq-config-data" not found Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.174299 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance324e-account-delete-6s4dm"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.189643 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-x82dv"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.254545 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-6lqnm"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.254847 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-6lqnm" podUID="fd2094a7-9a7b-417b-b4ea-b1da8b69cac6" containerName="openstack-network-exporter" containerID="cri-o://d78defca92531f2b0db0c17c5224178008057f1cf21db48f31f088bda1fc039b" gracePeriod=30 Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.260675 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd8zj\" (UniqueName: \"kubernetes.io/projected/91932480-76aa-4751-9251-3ad82d849882-kube-api-access-jd8zj\") pod \"glance324e-account-delete-6s4dm\" (UID: \"91932480-76aa-4751-9251-3ad82d849882\") " pod="openstack/glance324e-account-delete-6s4dm" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.260789 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91932480-76aa-4751-9251-3ad82d849882-operator-scripts\") pod \"glance324e-account-delete-6s4dm\" (UID: \"91932480-76aa-4751-9251-3ad82d849882\") " pod="openstack/glance324e-account-delete-6s4dm" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.273629 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-hfmpn"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.335305 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.335704 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="604a8b79-fcf3-40cf-82e6-8af2cf667b0f" containerName="ovn-northd" containerID="cri-o://87439695570ef9b3e9da98681ae2d1dec7fb73d967ccb8912a2e40ca13529cf7" gracePeriod=30 Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.335904 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="604a8b79-fcf3-40cf-82e6-8af2cf667b0f" containerName="openstack-network-exporter" containerID="cri-o://4403ed72fa83e2083ad29edab75aeefc505950c9b561718246753461398f9c33" gracePeriod=30 Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.375024 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91932480-76aa-4751-9251-3ad82d849882-operator-scripts\") pod \"glance324e-account-delete-6s4dm\" (UID: \"91932480-76aa-4751-9251-3ad82d849882\") " pod="openstack/glance324e-account-delete-6s4dm" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.375390 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd8zj\" (UniqueName: \"kubernetes.io/projected/91932480-76aa-4751-9251-3ad82d849882-kube-api-access-jd8zj\") pod \"glance324e-account-delete-6s4dm\" (UID: \"91932480-76aa-4751-9251-3ad82d849882\") " pod="openstack/glance324e-account-delete-6s4dm" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.376742 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91932480-76aa-4751-9251-3ad82d849882-operator-scripts\") pod \"glance324e-account-delete-6s4dm\" (UID: \"91932480-76aa-4751-9251-3ad82d849882\") " pod="openstack/glance324e-account-delete-6s4dm" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.431079 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-h4brs"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.484755 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-h4brs"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.535040 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd8zj\" (UniqueName: \"kubernetes.io/projected/91932480-76aa-4751-9251-3ad82d849882-kube-api-access-jd8zj\") pod \"glance324e-account-delete-6s4dm\" (UID: \"91932480-76aa-4751-9251-3ad82d849882\") " pod="openstack/glance324e-account-delete-6s4dm" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.556242 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="986d954e-89dd-4878-bbdd-b93e6a0ca155" path="/var/lib/kubelet/pods/986d954e-89dd-4878-bbdd-b93e6a0ca155/volumes" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.557098 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement8657-account-delete-hx45b"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.580053 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.580609 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement8657-account-delete-hx45b" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.596407 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gww9m\" (UniqueName: \"kubernetes.io/projected/2a3a8b5d-01e4-4396-8a9b-abd85dc55572-kube-api-access-gww9m\") pod \"placement8657-account-delete-hx45b\" (UID: \"2a3a8b5d-01e4-4396-8a9b-abd85dc55572\") " pod="openstack/placement8657-account-delete-hx45b" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.596638 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a3a8b5d-01e4-4396-8a9b-abd85dc55572-operator-scripts\") pod \"placement8657-account-delete-hx45b\" (UID: \"2a3a8b5d-01e4-4396-8a9b-abd85dc55572\") " pod="openstack/placement8657-account-delete-hx45b" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.596761 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-l4gdq"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.622585 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement8657-account-delete-hx45b"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.639356 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-l4gdq"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.656996 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.657305 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" podUID="b88404c2-d4bf-4257-986a-1b78178fca88" containerName="dnsmasq-dns" containerID="cri-o://4d129b7c0c3d768fea14f715dcd8a374cb51cf6191bbfa597636e8ce11c8d85e" gracePeriod=10 Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.668312 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-zwn5q"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.678487 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-zwn5q"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.689273 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbicanff71-account-delete-2fnl8"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.693595 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanff71-account-delete-2fnl8" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.698785 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gww9m\" (UniqueName: \"kubernetes.io/projected/2a3a8b5d-01e4-4396-8a9b-abd85dc55572-kube-api-access-gww9m\") pod \"placement8657-account-delete-hx45b\" (UID: \"2a3a8b5d-01e4-4396-8a9b-abd85dc55572\") " pod="openstack/placement8657-account-delete-hx45b" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.698891 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a3a8b5d-01e4-4396-8a9b-abd85dc55572-operator-scripts\") pod \"placement8657-account-delete-hx45b\" (UID: \"2a3a8b5d-01e4-4396-8a9b-abd85dc55572\") " pod="openstack/placement8657-account-delete-hx45b" Dec 01 14:59:44 crc kubenswrapper[4810]: E1201 14:59:44.700339 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 01 14:59:44 crc kubenswrapper[4810]: E1201 14:59:44.700386 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-config-data podName:3b81b6a9-0be6-4ce1-81b4-043edb1c221f nodeName:}" failed. No retries permitted until 2025-12-01 14:59:45.700370881 +0000 UTC m=+1551.463880484 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-config-data") pod "rabbitmq-server-0" (UID: "3b81b6a9-0be6-4ce1-81b4-043edb1c221f") : configmap "rabbitmq-config-data" not found Dec 01 14:59:44 crc kubenswrapper[4810]: E1201 14:59:44.700753 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 01 14:59:44 crc kubenswrapper[4810]: E1201 14:59:44.700799 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-config-data podName:56bbf32f-35b5-4f92-bfc0-9266b0ecccac nodeName:}" failed. No retries permitted until 2025-12-01 14:59:45.200782393 +0000 UTC m=+1550.964291996 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-config-data") pod "rabbitmq-cell1-server-0" (UID: "56bbf32f-35b5-4f92-bfc0-9266b0ecccac") : configmap "rabbitmq-cell1-config-data" not found Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.700960 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a3a8b5d-01e4-4396-8a9b-abd85dc55572-operator-scripts\") pod \"placement8657-account-delete-hx45b\" (UID: \"2a3a8b5d-01e4-4396-8a9b-abd85dc55572\") " pod="openstack/placement8657-account-delete-hx45b" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.715272 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicanff71-account-delete-2fnl8"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.732534 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-vlw7c"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.766898 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance324e-account-delete-6s4dm" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.767922 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gww9m\" (UniqueName: \"kubernetes.io/projected/2a3a8b5d-01e4-4396-8a9b-abd85dc55572-kube-api-access-gww9m\") pod \"placement8657-account-delete-hx45b\" (UID: \"2a3a8b5d-01e4-4396-8a9b-abd85dc55572\") " pod="openstack/placement8657-account-delete-hx45b" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.801131 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8cjq\" (UniqueName: \"kubernetes.io/projected/1ef9fa82-f302-47e9-bed3-b02f34a2340f-kube-api-access-l8cjq\") pod \"barbicanff71-account-delete-2fnl8\" (UID: \"1ef9fa82-f302-47e9-bed3-b02f34a2340f\") " pod="openstack/barbicanff71-account-delete-2fnl8" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.802963 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts\") pod \"barbicanff71-account-delete-2fnl8\" (UID: \"1ef9fa82-f302-47e9-bed3-b02f34a2340f\") " pod="openstack/barbicanff71-account-delete-2fnl8" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.837124 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-mslgc"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.860870 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-vlw7c"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.892730 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-mslgc"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.904361 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8cjq\" (UniqueName: \"kubernetes.io/projected/1ef9fa82-f302-47e9-bed3-b02f34a2340f-kube-api-access-l8cjq\") pod \"barbicanff71-account-delete-2fnl8\" (UID: \"1ef9fa82-f302-47e9-bed3-b02f34a2340f\") " pod="openstack/barbicanff71-account-delete-2fnl8" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.904540 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts\") pod \"barbicanff71-account-delete-2fnl8\" (UID: \"1ef9fa82-f302-47e9-bed3-b02f34a2340f\") " pod="openstack/barbicanff71-account-delete-2fnl8" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.905235 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts\") pod \"barbicanff71-account-delete-2fnl8\" (UID: \"1ef9fa82-f302-47e9-bed3-b02f34a2340f\") " pod="openstack/barbicanff71-account-delete-2fnl8" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.928157 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8cjq\" (UniqueName: \"kubernetes.io/projected/1ef9fa82-f302-47e9-bed3-b02f34a2340f-kube-api-access-l8cjq\") pod \"barbicanff71-account-delete-2fnl8\" (UID: \"1ef9fa82-f302-47e9-bed3-b02f34a2340f\") " pod="openstack/barbicanff71-account-delete-2fnl8" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.956621 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder17c6-account-delete-qktk6"] Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.958052 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder17c6-account-delete-qktk6" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.964006 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement8657-account-delete-hx45b" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.985947 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder17c6-account-delete-qktk6"] Dec 01 14:59:44 crc kubenswrapper[4810]: E1201 14:59:44.991901 4810 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-hfmpn" message=< Dec 01 14:59:44 crc kubenswrapper[4810]: Exiting ovn-controller (1) [ OK ] Dec 01 14:59:44 crc kubenswrapper[4810]: > Dec 01 14:59:44 crc kubenswrapper[4810]: E1201 14:59:44.991948 4810 kuberuntime_container.go:691] "PreStop hook failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " pod="openstack/ovn-controller-hfmpn" podUID="88635a0d-cb9d-43c1-9086-5623c9a780f8" containerName="ovn-controller" containerID="cri-o://1c3b60c932bcc8ce61ca6d0f28d631b382aa2b237627432ebf90c92ecb1485d0" Dec 01 14:59:44 crc kubenswrapper[4810]: I1201 14:59:44.991983 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-hfmpn" podUID="88635a0d-cb9d-43c1-9086-5623c9a780f8" containerName="ovn-controller" containerID="cri-o://1c3b60c932bcc8ce61ca6d0f28d631b382aa2b237627432ebf90c92ecb1485d0" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.007993 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6c5f549c7d-52fkz"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.008452 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-6c5f549c7d-52fkz" podUID="31b83b20-83c4-4e06-8eb0-1e1d83d4b160" containerName="placement-log" containerID="cri-o://ba0d20517bfab8cdb0b867e26e4238d3696e54b13598a77e2b77184f47eb47d0" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.008599 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-6c5f549c7d-52fkz" podUID="31b83b20-83c4-4e06-8eb0-1e1d83d4b160" containerName="placement-api" containerID="cri-o://c81abf769f22e0110d5df574ceaf9e47a00a5723bf172e5cebaebd7bde8ecee2" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.035764 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.036289 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="account-server" containerID="cri-o://cc265ca64611ebb2ffafcd64f73b893f0707c1a67a0e439ab6b7c8ef0dec3227" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.036485 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="container-updater" containerID="cri-o://0920962c564ec41c800c30de23663892437885ded40e3db2acbc5ff4420710d7" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.036528 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-expirer" containerID="cri-o://10f0178b4b19110c943a68bcae88fac92f4d9354eaf37cbb5fce2607db759cb3" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.036666 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="container-auditor" containerID="cri-o://7c0fca921a04e4ac4e65934e2a8f95df9f334f4836f26bf901cf250082061924" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.036790 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="container-replicator" containerID="cri-o://7c19582265bd684bc9384473ffe961c1afdbc4a8cbb8d269b1882eb4dab174dc" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.036856 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-replicator" containerID="cri-o://4d0480c8c52ca952777f48c3418a4710c30076e9fef37046f82aa4cf60de8531" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.036819 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-server" containerID="cri-o://ed49f712640c3ec14f04807a032c7cf926007968ca0dcd16ace0f1dc97a84d7a" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.036848 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-auditor" containerID="cri-o://18f2b5779be96d5d0ed49bece9ede2da2e81ab667fe59b1dcf9a253cd41860c6" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.037132 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="container-server" containerID="cri-o://184d4e1ea01b4832c5828fa2fe97577fb38a0e14ab4fe7d23a5339c79d094ef7" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.037339 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="account-replicator" containerID="cri-o://7b27d0e70deed824a6d31268fa11d712beb18bff269cd2c61fb15665851547b6" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.037421 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="swift-recon-cron" containerID="cri-o://24cce379393b557c9eef88db330dd60281782cf5c7918b19088e9f8504b09390" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.037401 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="account-reaper" containerID="cri-o://7a9772a6d14d26dd0282c93cad34c0f2f79a0710fc08b93b93ed46c435efb103" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.036837 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-updater" containerID="cri-o://3eeb13544ba7950f14cd2ed7a1ec0e2b77f5688fbb09bf134178e7165abccdf9" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.037644 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="rsync" containerID="cri-o://bfb0384344df9db35a4ff5ee4331e251ef264c0c11ab485df0d97f60541193de" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.037206 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="account-auditor" containerID="cri-o://8124aab34c8595c54770f6fe4ffe9d15bf793940c0b2cdffcb037b35b9a1e5ce" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.104245 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.104916 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" containerName="openstack-network-exporter" containerID="cri-o://6fcdb0e74e2281103d173be66fd77a226460e47c85adf65b5481538a371e0a73" gracePeriod=300 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.113512 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts\") pod \"cinder17c6-account-delete-qktk6\" (UID: \"bf15ccdf-1e4e-44ab-8ccb-819502935a7a\") " pod="openstack/cinder17c6-account-delete-qktk6" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.113742 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7274\" (UniqueName: \"kubernetes.io/projected/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-kube-api-access-n7274\") pod \"cinder17c6-account-delete-qktk6\" (UID: \"bf15ccdf-1e4e-44ab-8ccb-819502935a7a\") " pod="openstack/cinder17c6-account-delete-qktk6" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.148041 4810 generic.go:334] "Generic (PLEG): container finished" podID="88635a0d-cb9d-43c1-9086-5623c9a780f8" containerID="1c3b60c932bcc8ce61ca6d0f28d631b382aa2b237627432ebf90c92ecb1485d0" exitCode=0 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.148454 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hfmpn" event={"ID":"88635a0d-cb9d-43c1-9086-5623c9a780f8","Type":"ContainerDied","Data":"1c3b60c932bcc8ce61ca6d0f28d631b382aa2b237627432ebf90c92ecb1485d0"} Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.149459 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutronc91e-account-delete-t585q"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.150753 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutronc91e-account-delete-t585q" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.153767 4810 generic.go:334] "Generic (PLEG): container finished" podID="604a8b79-fcf3-40cf-82e6-8af2cf667b0f" containerID="4403ed72fa83e2083ad29edab75aeefc505950c9b561718246753461398f9c33" exitCode=2 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.154206 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"604a8b79-fcf3-40cf-82e6-8af2cf667b0f","Type":"ContainerDied","Data":"4403ed72fa83e2083ad29edab75aeefc505950c9b561718246753461398f9c33"} Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.177116 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanff71-account-delete-2fnl8" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.179653 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutronc91e-account-delete-t585q"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.189855 4810 generic.go:334] "Generic (PLEG): container finished" podID="b88404c2-d4bf-4257-986a-1b78178fca88" containerID="4d129b7c0c3d768fea14f715dcd8a374cb51cf6191bbfa597636e8ce11c8d85e" exitCode=0 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.189984 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" event={"ID":"b88404c2-d4bf-4257-986a-1b78178fca88","Type":"ContainerDied","Data":"4d129b7c0c3d768fea14f715dcd8a374cb51cf6191bbfa597636e8ce11c8d85e"} Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.218092 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts\") pod \"cinder17c6-account-delete-qktk6\" (UID: \"bf15ccdf-1e4e-44ab-8ccb-819502935a7a\") " pod="openstack/cinder17c6-account-delete-qktk6" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.218866 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7274\" (UniqueName: \"kubernetes.io/projected/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-kube-api-access-n7274\") pod \"cinder17c6-account-delete-qktk6\" (UID: \"bf15ccdf-1e4e-44ab-8ccb-819502935a7a\") " pod="openstack/cinder17c6-account-delete-qktk6" Dec 01 14:59:45 crc kubenswrapper[4810]: E1201 14:59:45.219294 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 01 14:59:45 crc kubenswrapper[4810]: E1201 14:59:45.219347 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-config-data podName:56bbf32f-35b5-4f92-bfc0-9266b0ecccac nodeName:}" failed. No retries permitted until 2025-12-01 14:59:46.219330262 +0000 UTC m=+1551.982839865 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-config-data") pod "rabbitmq-cell1-server-0" (UID: "56bbf32f-35b5-4f92-bfc0-9266b0ecccac") : configmap "rabbitmq-cell1-config-data" not found Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.220415 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts\") pod \"cinder17c6-account-delete-qktk6\" (UID: \"bf15ccdf-1e4e-44ab-8ccb-819502935a7a\") " pod="openstack/cinder17c6-account-delete-qktk6" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.233504 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-6lqnm_fd2094a7-9a7b-417b-b4ea-b1da8b69cac6/openstack-network-exporter/0.log" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.233560 4810 generic.go:334] "Generic (PLEG): container finished" podID="fd2094a7-9a7b-417b-b4ea-b1da8b69cac6" containerID="d78defca92531f2b0db0c17c5224178008057f1cf21db48f31f088bda1fc039b" exitCode=2 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.233628 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-6lqnm" event={"ID":"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6","Type":"ContainerDied","Data":"d78defca92531f2b0db0c17c5224178008057f1cf21db48f31f088bda1fc039b"} Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.265967 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.266560 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="3798bfc4-fb37-48d1-8146-42d177cf7861" containerName="openstack-network-exporter" containerID="cri-o://d74f9aed9f56dc44f786f3d28286286934f841d05512a1633b5fe3b4cc217204" gracePeriod=300 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.279254 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7274\" (UniqueName: \"kubernetes.io/projected/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-kube-api-access-n7274\") pod \"cinder17c6-account-delete-qktk6\" (UID: \"bf15ccdf-1e4e-44ab-8ccb-819502935a7a\") " pod="openstack/cinder17c6-account-delete-qktk6" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.280818 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.281084 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" containerName="glance-log" containerID="cri-o://630e5f798fca1eedf921c4930a1d5c0f33b14ea41b717a6558b2a6c06e4903f1" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.281237 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" containerName="glance-httpd" containerID="cri-o://aba75285c97143bda5c120fd54711afa0a88b59eeac5b761fa88578cc1f0f3b2" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.303820 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell1c683-account-delete-46h9z"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.305220 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell1c683-account-delete-46h9z" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.320858 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qgd7\" (UniqueName: \"kubernetes.io/projected/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-kube-api-access-6qgd7\") pod \"neutronc91e-account-delete-t585q\" (UID: \"f7b06553-ec05-4c6f-82c0-ef0b4dd3005a\") " pod="openstack/neutronc91e-account-delete-t585q" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.320900 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts\") pod \"neutronc91e-account-delete-t585q\" (UID: \"f7b06553-ec05-4c6f-82c0-ef0b4dd3005a\") " pod="openstack/neutronc91e-account-delete-t585q" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.333534 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell1c683-account-delete-46h9z"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.367846 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.368359 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bef8b0c7-66cd-4930-a78b-bf73804a73d2" containerName="cinder-scheduler" containerID="cri-o://d01cd11ccf42943023a23e361f67460aa3a8387758c28bbf2c4f4a6dba26d397" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.376455 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bef8b0c7-66cd-4930-a78b-bf73804a73d2" containerName="probe" containerID="cri-o://c3e8f637d25ffa49a1106b343b7c3861c122e5ccdeb4f52b034a012b1bf12322" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.409133 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder17c6-account-delete-qktk6" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.432727 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qgd7\" (UniqueName: \"kubernetes.io/projected/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-kube-api-access-6qgd7\") pod \"neutronc91e-account-delete-t585q\" (UID: \"f7b06553-ec05-4c6f-82c0-ef0b4dd3005a\") " pod="openstack/neutronc91e-account-delete-t585q" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.432796 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts\") pod \"neutronc91e-account-delete-t585q\" (UID: \"f7b06553-ec05-4c6f-82c0-ef0b4dd3005a\") " pod="openstack/neutronc91e-account-delete-t585q" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.432893 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/068d65be-c5cf-4198-a7ce-0aedeb5e9985-operator-scripts\") pod \"novacell1c683-account-delete-46h9z\" (UID: \"068d65be-c5cf-4198-a7ce-0aedeb5e9985\") " pod="openstack/novacell1c683-account-delete-46h9z" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.432969 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9rn6\" (UniqueName: \"kubernetes.io/projected/068d65be-c5cf-4198-a7ce-0aedeb5e9985-kube-api-access-f9rn6\") pod \"novacell1c683-account-delete-46h9z\" (UID: \"068d65be-c5cf-4198-a7ce-0aedeb5e9985\") " pod="openstack/novacell1c683-account-delete-46h9z" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.434083 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts\") pod \"neutronc91e-account-delete-t585q\" (UID: \"f7b06553-ec05-4c6f-82c0-ef0b4dd3005a\") " pod="openstack/neutronc91e-account-delete-t585q" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.473674 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-b6vj7"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.492070 4810 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/neutron-5477779d95-wg2vp" secret="" err="secret \"neutron-neutron-dockercfg-22xgp\" not found" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.511783 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" containerName="ovsdbserver-sb" containerID="cri-o://2bb8d93447e17f91799220c99cae83b3109ca92b16e6bbce7a1bf9da7f081d60" gracePeriod=300 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.517854 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qgd7\" (UniqueName: \"kubernetes.io/projected/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-kube-api-access-6qgd7\") pod \"neutronc91e-account-delete-t585q\" (UID: \"f7b06553-ec05-4c6f-82c0-ef0b4dd3005a\") " pod="openstack/neutronc91e-account-delete-t585q" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.535583 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-b6vj7"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.535368 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutronc91e-account-delete-t585q" Dec 01 14:59:45 crc kubenswrapper[4810]: E1201 14:59:45.587054 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2bb8d93447e17f91799220c99cae83b3109ca92b16e6bbce7a1bf9da7f081d60" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.599517 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9rn6\" (UniqueName: \"kubernetes.io/projected/068d65be-c5cf-4198-a7ce-0aedeb5e9985-kube-api-access-f9rn6\") pod \"novacell1c683-account-delete-46h9z\" (UID: \"068d65be-c5cf-4198-a7ce-0aedeb5e9985\") " pod="openstack/novacell1c683-account-delete-46h9z" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.599949 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/068d65be-c5cf-4198-a7ce-0aedeb5e9985-operator-scripts\") pod \"novacell1c683-account-delete-46h9z\" (UID: \"068d65be-c5cf-4198-a7ce-0aedeb5e9985\") " pod="openstack/novacell1c683-account-delete-46h9z" Dec 01 14:59:45 crc kubenswrapper[4810]: E1201 14:59:45.601287 4810 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Dec 01 14:59:45 crc kubenswrapper[4810]: E1201 14:59:45.601342 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/068d65be-c5cf-4198-a7ce-0aedeb5e9985-operator-scripts podName:068d65be-c5cf-4198-a7ce-0aedeb5e9985 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:46.101325269 +0000 UTC m=+1551.864834872 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/068d65be-c5cf-4198-a7ce-0aedeb5e9985-operator-scripts") pod "novacell1c683-account-delete-46h9z" (UID: "068d65be-c5cf-4198-a7ce-0aedeb5e9985") : configmap "openstack-cell1-scripts" not found Dec 01 14:59:45 crc kubenswrapper[4810]: E1201 14:59:45.601405 4810 secret.go:188] Couldn't get secret openstack/neutron-config: secret "neutron-config" not found Dec 01 14:59:45 crc kubenswrapper[4810]: E1201 14:59:45.601429 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-config podName:ba5f8a69-22cf-4196-bc89-ffa60eda9303 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:46.101418191 +0000 UTC m=+1551.864927794 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-config") pod "neutron-5477779d95-wg2vp" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303") : secret "neutron-config" not found Dec 01 14:59:45 crc kubenswrapper[4810]: E1201 14:59:45.602036 4810 secret.go:188] Couldn't get secret openstack/neutron-httpd-config: secret "neutron-httpd-config" not found Dec 01 14:59:45 crc kubenswrapper[4810]: E1201 14:59:45.602065 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-httpd-config podName:ba5f8a69-22cf-4196-bc89-ffa60eda9303 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:46.102053118 +0000 UTC m=+1551.865562721 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "httpd-config" (UniqueName: "kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-httpd-config") pod "neutron-5477779d95-wg2vp" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303") : secret "neutron-httpd-config" not found Dec 01 14:59:45 crc kubenswrapper[4810]: E1201 14:59:45.614759 4810 projected.go:194] Error preparing data for projected volume kube-api-access-f9rn6 for pod openstack/novacell1c683-account-delete-46h9z: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 01 14:59:45 crc kubenswrapper[4810]: E1201 14:59:45.615002 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/068d65be-c5cf-4198-a7ce-0aedeb5e9985-kube-api-access-f9rn6 podName:068d65be-c5cf-4198-a7ce-0aedeb5e9985 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:46.11498384 +0000 UTC m=+1551.878493443 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-f9rn6" (UniqueName: "kubernetes.io/projected/068d65be-c5cf-4198-a7ce-0aedeb5e9985-kube-api-access-f9rn6") pod "novacell1c683-account-delete-46h9z" (UID: "068d65be-c5cf-4198-a7ce-0aedeb5e9985") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 01 14:59:45 crc kubenswrapper[4810]: E1201 14:59:45.624649 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2bb8d93447e17f91799220c99cae83b3109ca92b16e6bbce7a1bf9da7f081d60" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.626432 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="3798bfc4-fb37-48d1-8146-42d177cf7861" containerName="ovsdbserver-nb" containerID="cri-o://afd5408dd106d9875d318c4f58d0916d20a0514bcf3f71e7b57c27a91050bc80" gracePeriod=300 Dec 01 14:59:45 crc kubenswrapper[4810]: E1201 14:59:45.647869 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2bb8d93447e17f91799220c99cae83b3109ca92b16e6bbce7a1bf9da7f081d60" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 01 14:59:45 crc kubenswrapper[4810]: E1201 14:59:45.647931 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovsdbserver-sb-0" podUID="081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" containerName="ovsdbserver-sb" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.684331 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.684654 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9dd705f6-0d8f-49df-aff6-813792212999" containerName="glance-log" containerID="cri-o://1077d64ab29e8e4a5e08c7d81078a1254c035e9de419585bf1944ac1e43a909c" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.685452 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9dd705f6-0d8f-49df-aff6-813792212999" containerName="glance-httpd" containerID="cri-o://ca8c16cdb56a677709c0ac603398c85e43e5a2f9e1cb3036f753b35de804a1f8" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.695544 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell096e4-account-delete-qqlg4"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.700187 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell096e4-account-delete-qqlg4" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.704954 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell096e4-account-delete-qqlg4"] Dec 01 14:59:45 crc kubenswrapper[4810]: E1201 14:59:45.704999 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 01 14:59:45 crc kubenswrapper[4810]: E1201 14:59:45.705077 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-config-data podName:3b81b6a9-0be6-4ce1-81b4-043edb1c221f nodeName:}" failed. No retries permitted until 2025-12-01 14:59:47.705055609 +0000 UTC m=+1553.468565212 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-config-data") pod "rabbitmq-server-0" (UID: "3b81b6a9-0be6-4ce1-81b4-043edb1c221f") : configmap "rabbitmq-config-data" not found Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.715266 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.728182 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.728448 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e79c5698-ab2d-4292-bda9-fab8dbbb2abb" containerName="cinder-api-log" containerID="cri-o://ebb0282754daba83e3b7cf8ef6784b85c2feea8bbea0bae1fde73a5239ce6c9c" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.728631 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e79c5698-ab2d-4292-bda9-fab8dbbb2abb" containerName="cinder-api" containerID="cri-o://085bfd508911509acb2f401ff11dc8f197ed30234eac469830bf853fef1a201a" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.736761 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-6c8f8b5cf7-65pxh"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.736990 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" podUID="7478f42b-4be2-430b-9854-5930dccaee43" containerName="proxy-httpd" containerID="cri-o://65c24e7fbab051767fd3a4645fd6dfed9e0a34a7bc48306bde5ba5dbb76b4aad" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.737078 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" podUID="7478f42b-4be2-430b-9854-5930dccaee43" containerName="proxy-server" containerID="cri-o://1cc6810ebfaa94211940478b494df0cc163de04d06f714491ab4d71d2b9f032d" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.760062 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-84864d5bd5-wmq86"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.760413 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-84864d5bd5-wmq86" podUID="129ae854-1430-465f-a177-0c459bd8b4c8" containerName="barbican-worker-log" containerID="cri-o://b8a19753910ff67350d9da738f60efda1e12ea5411279e28d972ad0cb28c8f79" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.760889 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-84864d5bd5-wmq86" podUID="129ae854-1430-465f-a177-0c459bd8b4c8" containerName="barbican-worker" containerID="cri-o://8c00c76d0791b2c2bc8a97328993ef7e1e8aa6997c54b9aa8cb7c3385b227dd7" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.778167 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-mc5bd"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.798954 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-mc5bd"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.819089 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-78w6c"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.827790 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc8s4\" (UniqueName: \"kubernetes.io/projected/4f7b5a25-eed0-4122-a036-093ac7adca22-kube-api-access-bc8s4\") pod \"novacell096e4-account-delete-qqlg4\" (UID: \"4f7b5a25-eed0-4122-a036-093ac7adca22\") " pod="openstack/novacell096e4-account-delete-qqlg4" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.827880 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts\") pod \"novacell096e4-account-delete-qqlg4\" (UID: \"4f7b5a25-eed0-4122-a036-093ac7adca22\") " pod="openstack/novacell096e4-account-delete-qqlg4" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.833743 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-78w6c"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.840497 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapic0fd-account-delete-w7ttc"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.842284 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic0fd-account-delete-w7ttc" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.851966 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapic0fd-account-delete-w7ttc"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.862249 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-6d8fb448c4-2hg9g"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.862503 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" podUID="20964a12-27b5-46f7-a9b9-e68b8c7dfce2" containerName="barbican-keystone-listener-log" containerID="cri-o://5a5b7334cdb5668aa597f83c43e1f06755c33b97ba720234d356c46686f469f6" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.862734 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" podUID="20964a12-27b5-46f7-a9b9-e68b8c7dfce2" containerName="barbican-keystone-listener" containerID="cri-o://a40c19049624e61426f700a570719d59ec7a3b73107e4fbc2d3c42482e3793b0" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.874446 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="3b81b6a9-0be6-4ce1-81b4-043edb1c221f" containerName="rabbitmq" containerID="cri-o://74e6d90ac1ab1306dd26c5e8c6cd4cd8b45f47eb0ab68865c9c1575f423fce76" gracePeriod=604800 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.893076 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5477779d95-wg2vp"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.910868 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5b4ff7d6cb-m2g6c"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.911118 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" podUID="bdd90cbb-4242-4227-9de4-09f99e60a516" containerName="barbican-api-log" containerID="cri-o://0af6bd36655b389f1090223a090ab3a8d47cace5e570943004f63de4c038edb3" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.911663 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" podUID="bdd90cbb-4242-4227-9de4-09f99e60a516" containerName="barbican-api" containerID="cri-o://4a8d673779dd5bbbb9060129ba8fe9ee012663d8194ce9496b1684874e3903dc" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.931904 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc8s4\" (UniqueName: \"kubernetes.io/projected/4f7b5a25-eed0-4122-a036-093ac7adca22-kube-api-access-bc8s4\") pod \"novacell096e4-account-delete-qqlg4\" (UID: \"4f7b5a25-eed0-4122-a036-093ac7adca22\") " pod="openstack/novacell096e4-account-delete-qqlg4" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.932149 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts\") pod \"novacell096e4-account-delete-qqlg4\" (UID: \"4f7b5a25-eed0-4122-a036-093ac7adca22\") " pod="openstack/novacell096e4-account-delete-qqlg4" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.932995 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts\") pod \"novacell096e4-account-delete-qqlg4\" (UID: \"4f7b5a25-eed0-4122-a036-093ac7adca22\") " pod="openstack/novacell096e4-account-delete-qqlg4" Dec 01 14:59:45 crc kubenswrapper[4810]: E1201 14:59:45.950190 4810 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Dec 01 14:59:45 crc kubenswrapper[4810]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 01 14:59:45 crc kubenswrapper[4810]: + source /usr/local/bin/container-scripts/functions Dec 01 14:59:45 crc kubenswrapper[4810]: ++ OVNBridge=br-int Dec 01 14:59:45 crc kubenswrapper[4810]: ++ OVNRemote=tcp:localhost:6642 Dec 01 14:59:45 crc kubenswrapper[4810]: ++ OVNEncapType=geneve Dec 01 14:59:45 crc kubenswrapper[4810]: ++ OVNAvailabilityZones= Dec 01 14:59:45 crc kubenswrapper[4810]: ++ EnableChassisAsGateway=true Dec 01 14:59:45 crc kubenswrapper[4810]: ++ PhysicalNetworks= Dec 01 14:59:45 crc kubenswrapper[4810]: ++ OVNHostName= Dec 01 14:59:45 crc kubenswrapper[4810]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 01 14:59:45 crc kubenswrapper[4810]: ++ ovs_dir=/var/lib/openvswitch Dec 01 14:59:45 crc kubenswrapper[4810]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 01 14:59:45 crc kubenswrapper[4810]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 01 14:59:45 crc kubenswrapper[4810]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 01 14:59:45 crc kubenswrapper[4810]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 01 14:59:45 crc kubenswrapper[4810]: + sleep 0.5 Dec 01 14:59:45 crc kubenswrapper[4810]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 01 14:59:45 crc kubenswrapper[4810]: + sleep 0.5 Dec 01 14:59:45 crc kubenswrapper[4810]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 01 14:59:45 crc kubenswrapper[4810]: + sleep 0.5 Dec 01 14:59:45 crc kubenswrapper[4810]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 01 14:59:45 crc kubenswrapper[4810]: + cleanup_ovsdb_server_semaphore Dec 01 14:59:45 crc kubenswrapper[4810]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 01 14:59:45 crc kubenswrapper[4810]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 01 14:59:45 crc kubenswrapper[4810]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-x82dv" message=< Dec 01 14:59:45 crc kubenswrapper[4810]: Exiting ovsdb-server (5) [ OK ] Dec 01 14:59:45 crc kubenswrapper[4810]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 01 14:59:45 crc kubenswrapper[4810]: + source /usr/local/bin/container-scripts/functions Dec 01 14:59:45 crc kubenswrapper[4810]: ++ OVNBridge=br-int Dec 01 14:59:45 crc kubenswrapper[4810]: ++ OVNRemote=tcp:localhost:6642 Dec 01 14:59:45 crc kubenswrapper[4810]: ++ OVNEncapType=geneve Dec 01 14:59:45 crc kubenswrapper[4810]: ++ OVNAvailabilityZones= Dec 01 14:59:45 crc kubenswrapper[4810]: ++ EnableChassisAsGateway=true Dec 01 14:59:45 crc kubenswrapper[4810]: ++ PhysicalNetworks= Dec 01 14:59:45 crc kubenswrapper[4810]: ++ OVNHostName= Dec 01 14:59:45 crc kubenswrapper[4810]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 01 14:59:45 crc kubenswrapper[4810]: ++ ovs_dir=/var/lib/openvswitch Dec 01 14:59:45 crc kubenswrapper[4810]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 01 14:59:45 crc kubenswrapper[4810]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 01 14:59:45 crc kubenswrapper[4810]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 01 14:59:45 crc kubenswrapper[4810]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 01 14:59:45 crc kubenswrapper[4810]: + sleep 0.5 Dec 01 14:59:45 crc kubenswrapper[4810]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 01 14:59:45 crc kubenswrapper[4810]: + sleep 0.5 Dec 01 14:59:45 crc kubenswrapper[4810]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 01 14:59:45 crc kubenswrapper[4810]: + sleep 0.5 Dec 01 14:59:45 crc kubenswrapper[4810]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 01 14:59:45 crc kubenswrapper[4810]: + cleanup_ovsdb_server_semaphore Dec 01 14:59:45 crc kubenswrapper[4810]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 01 14:59:45 crc kubenswrapper[4810]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 01 14:59:45 crc kubenswrapper[4810]: > Dec 01 14:59:45 crc kubenswrapper[4810]: E1201 14:59:45.950242 4810 kuberuntime_container.go:691] "PreStop hook failed" err=< Dec 01 14:59:45 crc kubenswrapper[4810]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 01 14:59:45 crc kubenswrapper[4810]: + source /usr/local/bin/container-scripts/functions Dec 01 14:59:45 crc kubenswrapper[4810]: ++ OVNBridge=br-int Dec 01 14:59:45 crc kubenswrapper[4810]: ++ OVNRemote=tcp:localhost:6642 Dec 01 14:59:45 crc kubenswrapper[4810]: ++ OVNEncapType=geneve Dec 01 14:59:45 crc kubenswrapper[4810]: ++ OVNAvailabilityZones= Dec 01 14:59:45 crc kubenswrapper[4810]: ++ EnableChassisAsGateway=true Dec 01 14:59:45 crc kubenswrapper[4810]: ++ PhysicalNetworks= Dec 01 14:59:45 crc kubenswrapper[4810]: ++ OVNHostName= Dec 01 14:59:45 crc kubenswrapper[4810]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 01 14:59:45 crc kubenswrapper[4810]: ++ ovs_dir=/var/lib/openvswitch Dec 01 14:59:45 crc kubenswrapper[4810]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 01 14:59:45 crc kubenswrapper[4810]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 01 14:59:45 crc kubenswrapper[4810]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 01 14:59:45 crc kubenswrapper[4810]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 01 14:59:45 crc kubenswrapper[4810]: + sleep 0.5 Dec 01 14:59:45 crc kubenswrapper[4810]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 01 14:59:45 crc kubenswrapper[4810]: + sleep 0.5 Dec 01 14:59:45 crc kubenswrapper[4810]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 01 14:59:45 crc kubenswrapper[4810]: + sleep 0.5 Dec 01 14:59:45 crc kubenswrapper[4810]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 01 14:59:45 crc kubenswrapper[4810]: + cleanup_ovsdb_server_semaphore Dec 01 14:59:45 crc kubenswrapper[4810]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 01 14:59:45 crc kubenswrapper[4810]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 01 14:59:45 crc kubenswrapper[4810]: > pod="openstack/ovn-controller-ovs-x82dv" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovsdb-server" containerID="cri-o://0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.950308 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-x82dv" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovsdb-server" containerID="cri-o://0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" gracePeriod=29 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.957701 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.957961 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" containerName="nova-metadata-log" containerID="cri-o://97cf7e0548b92abb2ae87397174ade605222b8ac75fdc9ae22ab5abf9a072ec8" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.958488 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" containerName="nova-metadata-metadata" containerID="cri-o://08362c9d20e6c5ffc7ef2c58a68a05a6da9d37da0f16242f1435dc93d6db4209" gracePeriod=30 Dec 01 14:59:45 crc kubenswrapper[4810]: I1201 14:59:45.981710 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc8s4\" (UniqueName: \"kubernetes.io/projected/4f7b5a25-eed0-4122-a036-093ac7adca22-kube-api-access-bc8s4\") pod \"novacell096e4-account-delete-qqlg4\" (UID: \"4f7b5a25-eed0-4122-a036-093ac7adca22\") " pod="openstack/novacell096e4-account-delete-qqlg4" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.005866 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.011305 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.035328 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x66rb\" (UniqueName: \"kubernetes.io/projected/476b71ac-1050-46a6-b51c-168e75b1a350-kube-api-access-x66rb\") pod \"novaapic0fd-account-delete-w7ttc\" (UID: \"476b71ac-1050-46a6-b51c-168e75b1a350\") " pod="openstack/novaapic0fd-account-delete-w7ttc" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.035373 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts\") pod \"novaapic0fd-account-delete-w7ttc\" (UID: \"476b71ac-1050-46a6-b51c-168e75b1a350\") " pod="openstack/novaapic0fd-account-delete-w7ttc" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.072697 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell1c683-account-delete-46h9z"] Dec 01 14:59:46 crc kubenswrapper[4810]: E1201 14:59:46.073805 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-f9rn6 operator-scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/novacell1c683-account-delete-46h9z" podUID="068d65be-c5cf-4198-a7ce-0aedeb5e9985" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.115152 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell096e4-account-delete-qqlg4" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.120656 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.120988 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9acd8c29-93f5-4b0f-8b32-d8f3d576767b" containerName="nova-api-log" containerID="cri-o://6c951f3e299a57aa9f4c575d8d057fd04b80e714ae7050a14a4ca0758c9a0549" gracePeriod=30 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.122630 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9acd8c29-93f5-4b0f-8b32-d8f3d576767b" containerName="nova-api-api" containerID="cri-o://1f6241fff3fa641130500b0c5b06297d792485793ad36fdc4dd06993a79f221d" gracePeriod=30 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.136691 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x66rb\" (UniqueName: \"kubernetes.io/projected/476b71ac-1050-46a6-b51c-168e75b1a350-kube-api-access-x66rb\") pod \"novaapic0fd-account-delete-w7ttc\" (UID: \"476b71ac-1050-46a6-b51c-168e75b1a350\") " pod="openstack/novaapic0fd-account-delete-w7ttc" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.136750 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts\") pod \"novaapic0fd-account-delete-w7ttc\" (UID: \"476b71ac-1050-46a6-b51c-168e75b1a350\") " pod="openstack/novaapic0fd-account-delete-w7ttc" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.136801 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9rn6\" (UniqueName: \"kubernetes.io/projected/068d65be-c5cf-4198-a7ce-0aedeb5e9985-kube-api-access-f9rn6\") pod \"novacell1c683-account-delete-46h9z\" (UID: \"068d65be-c5cf-4198-a7ce-0aedeb5e9985\") " pod="openstack/novacell1c683-account-delete-46h9z" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.138011 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts\") pod \"novaapic0fd-account-delete-w7ttc\" (UID: \"476b71ac-1050-46a6-b51c-168e75b1a350\") " pod="openstack/novaapic0fd-account-delete-w7ttc" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.139305 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/068d65be-c5cf-4198-a7ce-0aedeb5e9985-operator-scripts\") pod \"novacell1c683-account-delete-46h9z\" (UID: \"068d65be-c5cf-4198-a7ce-0aedeb5e9985\") " pod="openstack/novacell1c683-account-delete-46h9z" Dec 01 14:59:46 crc kubenswrapper[4810]: E1201 14:59:46.140110 4810 secret.go:188] Couldn't get secret openstack/neutron-config: secret "neutron-config" not found Dec 01 14:59:46 crc kubenswrapper[4810]: E1201 14:59:46.140179 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-config podName:ba5f8a69-22cf-4196-bc89-ffa60eda9303 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:47.14016204 +0000 UTC m=+1552.903671643 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-config") pod "neutron-5477779d95-wg2vp" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303") : secret "neutron-config" not found Dec 01 14:59:46 crc kubenswrapper[4810]: E1201 14:59:46.140669 4810 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Dec 01 14:59:46 crc kubenswrapper[4810]: E1201 14:59:46.140697 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/068d65be-c5cf-4198-a7ce-0aedeb5e9985-operator-scripts podName:068d65be-c5cf-4198-a7ce-0aedeb5e9985 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:47.140688013 +0000 UTC m=+1552.904197616 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/068d65be-c5cf-4198-a7ce-0aedeb5e9985-operator-scripts") pod "novacell1c683-account-delete-46h9z" (UID: "068d65be-c5cf-4198-a7ce-0aedeb5e9985") : configmap "openstack-cell1-scripts" not found Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.145399 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="56bbf32f-35b5-4f92-bfc0-9266b0ecccac" containerName="rabbitmq" containerID="cri-o://82a26c84ad224e60a4fa117cf5e1d03643cfd46868d1eeb656fac71a268ad1e5" gracePeriod=604800 Dec 01 14:59:46 crc kubenswrapper[4810]: E1201 14:59:46.147140 4810 secret.go:188] Couldn't get secret openstack/neutron-httpd-config: secret "neutron-httpd-config" not found Dec 01 14:59:46 crc kubenswrapper[4810]: E1201 14:59:46.147244 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-httpd-config podName:ba5f8a69-22cf-4196-bc89-ffa60eda9303 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:47.147214211 +0000 UTC m=+1552.910723894 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "httpd-config" (UniqueName: "kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-httpd-config") pod "neutron-5477779d95-wg2vp" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303") : secret "neutron-httpd-config" not found Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.147289 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-c683-account-create-update-m9bvk"] Dec 01 14:59:46 crc kubenswrapper[4810]: E1201 14:59:46.154421 4810 projected.go:194] Error preparing data for projected volume kube-api-access-f9rn6 for pod openstack/novacell1c683-account-delete-46h9z: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 01 14:59:46 crc kubenswrapper[4810]: E1201 14:59:46.154656 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/068d65be-c5cf-4198-a7ce-0aedeb5e9985-kube-api-access-f9rn6 podName:068d65be-c5cf-4198-a7ce-0aedeb5e9985 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:47.154626322 +0000 UTC m=+1552.918135925 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-f9rn6" (UniqueName: "kubernetes.io/projected/068d65be-c5cf-4198-a7ce-0aedeb5e9985-kube-api-access-f9rn6") pod "novacell1c683-account-delete-46h9z" (UID: "068d65be-c5cf-4198-a7ce-0aedeb5e9985") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.157260 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-x82dv" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovs-vswitchd" containerID="cri-o://74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" gracePeriod=29 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.160070 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-c683-account-create-update-m9bvk"] Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.184651 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-r8tkk"] Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.222745 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x66rb\" (UniqueName: \"kubernetes.io/projected/476b71ac-1050-46a6-b51c-168e75b1a350-kube-api-access-x66rb\") pod \"novaapic0fd-account-delete-w7ttc\" (UID: \"476b71ac-1050-46a6-b51c-168e75b1a350\") " pod="openstack/novaapic0fd-account-delete-w7ttc" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.231911 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-6lqnm_fd2094a7-9a7b-417b-b4ea-b1da8b69cac6/openstack-network-exporter/0.log" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.231994 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.233857 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-r8tkk"] Dec 01 14:59:46 crc kubenswrapper[4810]: E1201 14:59:46.245858 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 01 14:59:46 crc kubenswrapper[4810]: E1201 14:59:46.245916 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-config-data podName:56bbf32f-35b5-4f92-bfc0-9266b0ecccac nodeName:}" failed. No retries permitted until 2025-12-01 14:59:48.245902974 +0000 UTC m=+1554.009412577 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-config-data") pod "rabbitmq-cell1-server-0" (UID: "56bbf32f-35b5-4f92-bfc0-9266b0ecccac") : configmap "rabbitmq-cell1-config-data" not found Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.251343 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.251603 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="0899579c-adaa-41fb-a36d-31ba657c6c1a" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://7f92191f00ad28697cca6ad61e70b41f045bb53297c7b28a21e7c29a04def64d" gracePeriod=30 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.264167 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.264383 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c" containerName="nova-cell1-conductor-conductor" containerID="cri-o://c0a6acaf5c3f0108d3d50cb630c0e2cb9470e93a6443905cbd9058d2ef094517" gracePeriod=30 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.291496 4810 generic.go:334] "Generic (PLEG): container finished" podID="d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" containerID="630e5f798fca1eedf921c4930a1d5c0f33b14ea41b717a6558b2a6c06e4903f1" exitCode=143 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.291986 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc","Type":"ContainerDied","Data":"630e5f798fca1eedf921c4930a1d5c0f33b14ea41b717a6558b2a6c06e4903f1"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.334091 4810 generic.go:334] "Generic (PLEG): container finished" podID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" exitCode=0 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.334174 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-x82dv" event={"ID":"9c699a0a-f268-419b-a829-a9ebaa52b322","Type":"ContainerDied","Data":"0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.346622 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-combined-ca-bundle\") pod \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.346702 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-metrics-certs-tls-certs\") pod \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.346999 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-config\") pod \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.347042 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-ovn-rundir\") pod \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.347273 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-ovs-rundir\") pod \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.347344 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bp2d\" (UniqueName: \"kubernetes.io/projected/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-kube-api-access-7bp2d\") pod \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\" (UID: \"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6\") " Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.358182 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-config" (OuterVolumeSpecName: "config") pod "fd2094a7-9a7b-417b-b4ea-b1da8b69cac6" (UID: "fd2094a7-9a7b-417b-b4ea-b1da8b69cac6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.360547 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "fd2094a7-9a7b-417b-b4ea-b1da8b69cac6" (UID: "fd2094a7-9a7b-417b-b4ea-b1da8b69cac6"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.360602 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "fd2094a7-9a7b-417b-b4ea-b1da8b69cac6" (UID: "fd2094a7-9a7b-417b-b4ea-b1da8b69cac6"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.360986 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3798bfc4-fb37-48d1-8146-42d177cf7861/ovsdbserver-nb/0.log" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.361024 4810 generic.go:334] "Generic (PLEG): container finished" podID="3798bfc4-fb37-48d1-8146-42d177cf7861" containerID="d74f9aed9f56dc44f786f3d28286286934f841d05512a1633b5fe3b4cc217204" exitCode=2 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.361043 4810 generic.go:334] "Generic (PLEG): container finished" podID="3798bfc4-fb37-48d1-8146-42d177cf7861" containerID="afd5408dd106d9875d318c4f58d0916d20a0514bcf3f71e7b57c27a91050bc80" exitCode=143 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.361188 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-f27f6"] Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.361215 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3798bfc4-fb37-48d1-8146-42d177cf7861","Type":"ContainerDied","Data":"d74f9aed9f56dc44f786f3d28286286934f841d05512a1633b5fe3b4cc217204"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.361236 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3798bfc4-fb37-48d1-8146-42d177cf7861","Type":"ContainerDied","Data":"afd5408dd106d9875d318c4f58d0916d20a0514bcf3f71e7b57c27a91050bc80"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.365903 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-f27f6"] Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.392714 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.392955 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="c73a61da-0c40-4a26-888c-ab5b994417d4" containerName="nova-cell0-conductor-conductor" containerID="cri-o://6cffd8cec4aa7e3f42f1a3b9cff8d7da52efafaece6f09b30bae42093876b093" gracePeriod=30 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.401425 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-kube-api-access-7bp2d" (OuterVolumeSpecName: "kube-api-access-7bp2d") pod "fd2094a7-9a7b-417b-b4ea-b1da8b69cac6" (UID: "fd2094a7-9a7b-417b-b4ea-b1da8b69cac6"). InnerVolumeSpecName "kube-api-access-7bp2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.424944 4810 generic.go:334] "Generic (PLEG): container finished" podID="129ae854-1430-465f-a177-0c459bd8b4c8" containerID="b8a19753910ff67350d9da738f60efda1e12ea5411279e28d972ad0cb28c8f79" exitCode=143 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.425030 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84864d5bd5-wmq86" event={"ID":"129ae854-1430-465f-a177-0c459bd8b4c8","Type":"ContainerDied","Data":"b8a19753910ff67350d9da738f60efda1e12ea5411279e28d972ad0cb28c8f79"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.427010 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd2094a7-9a7b-417b-b4ea-b1da8b69cac6" (UID: "fd2094a7-9a7b-417b-b4ea-b1da8b69cac6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.431327 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gnmm4"] Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.447719 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gnmm4"] Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.467740 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.468346 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" podUID="7478f42b-4be2-430b-9854-5930dccaee43" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.165:8080/healthcheck\": dial tcp 10.217.0.165:8080: connect: connection refused" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.468405 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="444ec27a-77d0-4b01-b59e-c6590c46a0c2" containerName="nova-scheduler-scheduler" containerID="cri-o://ffc643854f69e1e2f23a40254dbae2944125c570cca2c9e027760a19987124a6" gracePeriod=30 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.468541 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" podUID="7478f42b-4be2-430b-9854-5930dccaee43" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.165:8080/healthcheck\": dial tcp 10.217.0.165:8080: connect: connection refused" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.468753 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.468802 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.468816 4810 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-ovs-rundir\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.468825 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bp2d\" (UniqueName: \"kubernetes.io/projected/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-kube-api-access-7bp2d\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.468835 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.479602 4810 generic.go:334] "Generic (PLEG): container finished" podID="bdd90cbb-4242-4227-9de4-09f99e60a516" containerID="0af6bd36655b389f1090223a090ab3a8d47cace5e570943004f63de4c038edb3" exitCode=143 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.479665 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" event={"ID":"bdd90cbb-4242-4227-9de4-09f99e60a516","Type":"ContainerDied","Data":"0af6bd36655b389f1090223a090ab3a8d47cace5e570943004f63de4c038edb3"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.487109 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic0fd-account-delete-w7ttc" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.525673 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="d202704a-7a2e-46a6-9e20-2eeabdc11814" containerName="galera" containerID="cri-o://f21f812c44b7972a5f6ecdd8dbe2ffd46fb63a8fa481dd14afa8b978ba21edfb" gracePeriod=30 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.535841 4810 generic.go:334] "Generic (PLEG): container finished" podID="9dd705f6-0d8f-49df-aff6-813792212999" containerID="1077d64ab29e8e4a5e08c7d81078a1254c035e9de419585bf1944ac1e43a909c" exitCode=143 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.577880 4810 generic.go:334] "Generic (PLEG): container finished" podID="31b83b20-83c4-4e06-8eb0-1e1d83d4b160" containerID="ba0d20517bfab8cdb0b867e26e4238d3696e54b13598a77e2b77184f47eb47d0" exitCode=143 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.587390 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-6lqnm_fd2094a7-9a7b-417b-b4ea-b1da8b69cac6/openstack-network-exporter/0.log" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.588119 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "fd2094a7-9a7b-417b-b4ea-b1da8b69cac6" (UID: "fd2094a7-9a7b-417b-b4ea-b1da8b69cac6"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.588356 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.593128 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="077fd1b8-6283-4a1b-b159-437897afa556" path="/var/lib/kubelet/pods/077fd1b8-6283-4a1b-b159-437897afa556/volumes" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.593867 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09861728-4533-4ddf-a32d-3cb1b70fc325" path="/var/lib/kubelet/pods/09861728-4533-4ddf-a32d-3cb1b70fc325/volumes" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.594834 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b33096-654a-4684-971a-bd81b90d0a79" path="/var/lib/kubelet/pods/20b33096-654a-4684-971a-bd81b90d0a79/volumes" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.595651 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21fcb95f-afd8-43e5-90b2-afa5346fcd57" path="/var/lib/kubelet/pods/21fcb95f-afd8-43e5-90b2-afa5346fcd57/volumes" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.603525 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51938ba3-de39-414b-9777-fdc90cc14a97" path="/var/lib/kubelet/pods/51938ba3-de39-414b-9777-fdc90cc14a97/volumes" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.612949 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="636564e1-d025-4f10-bf66-22e848c21de0" path="/var/lib/kubelet/pods/636564e1-d025-4f10-bf66-22e848c21de0/volumes" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.614001 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bbc4102-9734-42a7-8bd5-9487e83d768b" path="/var/lib/kubelet/pods/8bbc4102-9734-42a7-8bd5-9487e83d768b/volumes" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.615012 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beec7c3d-1e8b-42c7-9e07-978388973b80" path="/var/lib/kubelet/pods/beec7c3d-1e8b-42c7-9e07-978388973b80/volumes" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.615438 4810 generic.go:334] "Generic (PLEG): container finished" podID="e79c5698-ab2d-4292-bda9-fab8dbbb2abb" containerID="ebb0282754daba83e3b7cf8ef6784b85c2feea8bbea0bae1fde73a5239ce6c9c" exitCode=143 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.617168 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef385349-2803-45cf-9307-8657305dc6cf" path="/var/lib/kubelet/pods/ef385349-2803-45cf-9307-8657305dc6cf/volumes" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.617844 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6346394-0edb-4b7b-85fd-47907ea08f08" path="/var/lib/kubelet/pods/f6346394-0edb-4b7b-85fd-47907ea08f08/volumes" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.618832 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88" path="/var/lib/kubelet/pods/fcc04db8-aad7-4ef7-abb2-c1f9f8a8ab88/volumes" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.619921 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9dd705f6-0d8f-49df-aff6-813792212999","Type":"ContainerDied","Data":"1077d64ab29e8e4a5e08c7d81078a1254c035e9de419585bf1944ac1e43a909c"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.619968 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6c5f549c7d-52fkz" event={"ID":"31b83b20-83c4-4e06-8eb0-1e1d83d4b160","Type":"ContainerDied","Data":"ba0d20517bfab8cdb0b867e26e4238d3696e54b13598a77e2b77184f47eb47d0"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.619986 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-6lqnm" event={"ID":"fd2094a7-9a7b-417b-b4ea-b1da8b69cac6","Type":"ContainerDied","Data":"d63d727c2af3793d4c63ed53b1bc16ab310573f3fe15ec6f8f985db2fab60875"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.620004 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e79c5698-ab2d-4292-bda9-fab8dbbb2abb","Type":"ContainerDied","Data":"ebb0282754daba83e3b7cf8ef6784b85c2feea8bbea0bae1fde73a5239ce6c9c"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.620040 4810 scope.go:117] "RemoveContainer" containerID="d78defca92531f2b0db0c17c5224178008057f1cf21db48f31f088bda1fc039b" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.678035 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.774912 4810 generic.go:334] "Generic (PLEG): container finished" podID="1f455986-23c9-4088-9829-51c74e5b7770" containerID="bfb0384344df9db35a4ff5ee4331e251ef264c0c11ab485df0d97f60541193de" exitCode=0 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775317 4810 generic.go:334] "Generic (PLEG): container finished" podID="1f455986-23c9-4088-9829-51c74e5b7770" containerID="10f0178b4b19110c943a68bcae88fac92f4d9354eaf37cbb5fce2607db759cb3" exitCode=0 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775332 4810 generic.go:334] "Generic (PLEG): container finished" podID="1f455986-23c9-4088-9829-51c74e5b7770" containerID="3eeb13544ba7950f14cd2ed7a1ec0e2b77f5688fbb09bf134178e7165abccdf9" exitCode=0 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775345 4810 generic.go:334] "Generic (PLEG): container finished" podID="1f455986-23c9-4088-9829-51c74e5b7770" containerID="18f2b5779be96d5d0ed49bece9ede2da2e81ab667fe59b1dcf9a253cd41860c6" exitCode=0 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775333 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerDied","Data":"bfb0384344df9db35a4ff5ee4331e251ef264c0c11ab485df0d97f60541193de"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775398 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerDied","Data":"10f0178b4b19110c943a68bcae88fac92f4d9354eaf37cbb5fce2607db759cb3"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775421 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerDied","Data":"3eeb13544ba7950f14cd2ed7a1ec0e2b77f5688fbb09bf134178e7165abccdf9"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775434 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerDied","Data":"18f2b5779be96d5d0ed49bece9ede2da2e81ab667fe59b1dcf9a253cd41860c6"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775445 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerDied","Data":"4d0480c8c52ca952777f48c3418a4710c30076e9fef37046f82aa4cf60de8531"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775356 4810 generic.go:334] "Generic (PLEG): container finished" podID="1f455986-23c9-4088-9829-51c74e5b7770" containerID="4d0480c8c52ca952777f48c3418a4710c30076e9fef37046f82aa4cf60de8531" exitCode=0 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775487 4810 generic.go:334] "Generic (PLEG): container finished" podID="1f455986-23c9-4088-9829-51c74e5b7770" containerID="ed49f712640c3ec14f04807a032c7cf926007968ca0dcd16ace0f1dc97a84d7a" exitCode=0 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775507 4810 generic.go:334] "Generic (PLEG): container finished" podID="1f455986-23c9-4088-9829-51c74e5b7770" containerID="0920962c564ec41c800c30de23663892437885ded40e3db2acbc5ff4420710d7" exitCode=0 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775516 4810 generic.go:334] "Generic (PLEG): container finished" podID="1f455986-23c9-4088-9829-51c74e5b7770" containerID="7c0fca921a04e4ac4e65934e2a8f95df9f334f4836f26bf901cf250082061924" exitCode=0 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775524 4810 generic.go:334] "Generic (PLEG): container finished" podID="1f455986-23c9-4088-9829-51c74e5b7770" containerID="7c19582265bd684bc9384473ffe961c1afdbc4a8cbb8d269b1882eb4dab174dc" exitCode=0 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775532 4810 generic.go:334] "Generic (PLEG): container finished" podID="1f455986-23c9-4088-9829-51c74e5b7770" containerID="184d4e1ea01b4832c5828fa2fe97577fb38a0e14ab4fe7d23a5339c79d094ef7" exitCode=0 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775540 4810 generic.go:334] "Generic (PLEG): container finished" podID="1f455986-23c9-4088-9829-51c74e5b7770" containerID="7a9772a6d14d26dd0282c93cad34c0f2f79a0710fc08b93b93ed46c435efb103" exitCode=0 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775549 4810 generic.go:334] "Generic (PLEG): container finished" podID="1f455986-23c9-4088-9829-51c74e5b7770" containerID="8124aab34c8595c54770f6fe4ffe9d15bf793940c0b2cdffcb037b35b9a1e5ce" exitCode=0 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775557 4810 generic.go:334] "Generic (PLEG): container finished" podID="1f455986-23c9-4088-9829-51c74e5b7770" containerID="7b27d0e70deed824a6d31268fa11d712beb18bff269cd2c61fb15665851547b6" exitCode=0 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775568 4810 generic.go:334] "Generic (PLEG): container finished" podID="1f455986-23c9-4088-9829-51c74e5b7770" containerID="cc265ca64611ebb2ffafcd64f73b893f0707c1a67a0e439ab6b7c8ef0dec3227" exitCode=0 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775683 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerDied","Data":"ed49f712640c3ec14f04807a032c7cf926007968ca0dcd16ace0f1dc97a84d7a"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775701 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerDied","Data":"0920962c564ec41c800c30de23663892437885ded40e3db2acbc5ff4420710d7"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775713 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerDied","Data":"7c0fca921a04e4ac4e65934e2a8f95df9f334f4836f26bf901cf250082061924"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775723 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerDied","Data":"7c19582265bd684bc9384473ffe961c1afdbc4a8cbb8d269b1882eb4dab174dc"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775733 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerDied","Data":"184d4e1ea01b4832c5828fa2fe97577fb38a0e14ab4fe7d23a5339c79d094ef7"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775743 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerDied","Data":"7a9772a6d14d26dd0282c93cad34c0f2f79a0710fc08b93b93ed46c435efb103"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775753 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerDied","Data":"8124aab34c8595c54770f6fe4ffe9d15bf793940c0b2cdffcb037b35b9a1e5ce"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775765 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerDied","Data":"7b27d0e70deed824a6d31268fa11d712beb18bff269cd2c61fb15665851547b6"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.775776 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerDied","Data":"cc265ca64611ebb2ffafcd64f73b893f0707c1a67a0e439ab6b7c8ef0dec3227"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.781650 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_081ac19d-af0e-48f7-9c2e-0d282a9bb7ba/ovsdbserver-sb/0.log" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.781744 4810 generic.go:334] "Generic (PLEG): container finished" podID="081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" containerID="6fcdb0e74e2281103d173be66fd77a226460e47c85adf65b5481538a371e0a73" exitCode=2 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.781764 4810 generic.go:334] "Generic (PLEG): container finished" podID="081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" containerID="2bb8d93447e17f91799220c99cae83b3109ca92b16e6bbce7a1bf9da7f081d60" exitCode=143 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.781855 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba","Type":"ContainerDied","Data":"6fcdb0e74e2281103d173be66fd77a226460e47c85adf65b5481538a371e0a73"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.781889 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba","Type":"ContainerDied","Data":"2bb8d93447e17f91799220c99cae83b3109ca92b16e6bbce7a1bf9da7f081d60"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.791983 4810 generic.go:334] "Generic (PLEG): container finished" podID="c35fc22d-59ab-4396-87c5-fadfc17ecc53" containerID="a303d0364f2f0ad764ac7d54b67330e695ac583a1627abd27d1f6447c9de1ee2" exitCode=137 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.807823 4810 generic.go:334] "Generic (PLEG): container finished" podID="3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" containerID="97cf7e0548b92abb2ae87397174ade605222b8ac75fdc9ae22ab5abf9a072ec8" exitCode=143 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.807946 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6","Type":"ContainerDied","Data":"97cf7e0548b92abb2ae87397174ade605222b8ac75fdc9ae22ab5abf9a072ec8"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.816605 4810 generic.go:334] "Generic (PLEG): container finished" podID="20964a12-27b5-46f7-a9b9-e68b8c7dfce2" containerID="5a5b7334cdb5668aa597f83c43e1f06755c33b97ba720234d356c46686f469f6" exitCode=143 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.816724 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" event={"ID":"20964a12-27b5-46f7-a9b9-e68b8c7dfce2","Type":"ContainerDied","Data":"5a5b7334cdb5668aa597f83c43e1f06755c33b97ba720234d356c46686f469f6"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.824512 4810 generic.go:334] "Generic (PLEG): container finished" podID="7478f42b-4be2-430b-9854-5930dccaee43" containerID="65c24e7fbab051767fd3a4645fd6dfed9e0a34a7bc48306bde5ba5dbb76b4aad" exitCode=0 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.824603 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell1c683-account-delete-46h9z" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.824921 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" event={"ID":"7478f42b-4be2-430b-9854-5930dccaee43","Type":"ContainerDied","Data":"65c24e7fbab051767fd3a4645fd6dfed9e0a34a7bc48306bde5ba5dbb76b4aad"} Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.825112 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5477779d95-wg2vp" podUID="ba5f8a69-22cf-4196-bc89-ffa60eda9303" containerName="neutron-api" containerID="cri-o://7b6de236d9a353033c45404d0d0dce626c48768c8ec55f0ea9bf63e5d411c381" gracePeriod=30 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.825218 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5477779d95-wg2vp" podUID="ba5f8a69-22cf-4196-bc89-ffa60eda9303" containerName="neutron-httpd" containerID="cri-o://8439e2334c700fa63f3562d0a3892b76d49cb65b3961834e76812a6b483afe47" gracePeriod=30 Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.860125 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance324e-account-delete-6s4dm"] Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.860493 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hfmpn" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.882040 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/88635a0d-cb9d-43c1-9086-5623c9a780f8-ovn-controller-tls-certs\") pod \"88635a0d-cb9d-43c1-9086-5623c9a780f8\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.886800 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88635a0d-cb9d-43c1-9086-5623c9a780f8-combined-ca-bundle\") pod \"88635a0d-cb9d-43c1-9086-5623c9a780f8\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.887198 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88635a0d-cb9d-43c1-9086-5623c9a780f8-scripts\") pod \"88635a0d-cb9d-43c1-9086-5623c9a780f8\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.890632 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88635a0d-cb9d-43c1-9086-5623c9a780f8-scripts" (OuterVolumeSpecName: "scripts") pod "88635a0d-cb9d-43c1-9086-5623c9a780f8" (UID: "88635a0d-cb9d-43c1-9086-5623c9a780f8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.891863 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/88635a0d-cb9d-43c1-9086-5623c9a780f8-var-log-ovn\") pod \"88635a0d-cb9d-43c1-9086-5623c9a780f8\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.891909 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/88635a0d-cb9d-43c1-9086-5623c9a780f8-var-run\") pod \"88635a0d-cb9d-43c1-9086-5623c9a780f8\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.892038 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6pz4\" (UniqueName: \"kubernetes.io/projected/88635a0d-cb9d-43c1-9086-5623c9a780f8-kube-api-access-b6pz4\") pod \"88635a0d-cb9d-43c1-9086-5623c9a780f8\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.892191 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/88635a0d-cb9d-43c1-9086-5623c9a780f8-var-run-ovn\") pod \"88635a0d-cb9d-43c1-9086-5623c9a780f8\" (UID: \"88635a0d-cb9d-43c1-9086-5623c9a780f8\") " Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.893074 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88635a0d-cb9d-43c1-9086-5623c9a780f8-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.893152 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88635a0d-cb9d-43c1-9086-5623c9a780f8-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "88635a0d-cb9d-43c1-9086-5623c9a780f8" (UID: "88635a0d-cb9d-43c1-9086-5623c9a780f8"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.893203 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88635a0d-cb9d-43c1-9086-5623c9a780f8-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "88635a0d-cb9d-43c1-9086-5623c9a780f8" (UID: "88635a0d-cb9d-43c1-9086-5623c9a780f8"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.893225 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88635a0d-cb9d-43c1-9086-5623c9a780f8-var-run" (OuterVolumeSpecName: "var-run") pod "88635a0d-cb9d-43c1-9086-5623c9a780f8" (UID: "88635a0d-cb9d-43c1-9086-5623c9a780f8"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.901842 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88635a0d-cb9d-43c1-9086-5623c9a780f8-kube-api-access-b6pz4" (OuterVolumeSpecName: "kube-api-access-b6pz4") pod "88635a0d-cb9d-43c1-9086-5623c9a780f8" (UID: "88635a0d-cb9d-43c1-9086-5623c9a780f8"). InnerVolumeSpecName "kube-api-access-b6pz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.932383 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.962428 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88635a0d-cb9d-43c1-9086-5623c9a780f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88635a0d-cb9d-43c1-9086-5623c9a780f8" (UID: "88635a0d-cb9d-43c1-9086-5623c9a780f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.967903 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_081ac19d-af0e-48f7-9c2e-0d282a9bb7ba/ovsdbserver-sb/0.log" Dec 01 14:59:46 crc kubenswrapper[4810]: I1201 14:59:46.968020 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.000437 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-ovsdb-rundir\") pod \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.000543 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-config\") pod \"b88404c2-d4bf-4257-986a-1b78178fca88\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.000587 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kczs5\" (UniqueName: \"kubernetes.io/projected/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-kube-api-access-kczs5\") pod \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.000637 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-dns-svc\") pod \"b88404c2-d4bf-4257-986a-1b78178fca88\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.000666 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-ovsdbserver-nb\") pod \"b88404c2-d4bf-4257-986a-1b78178fca88\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.000692 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-ovsdbserver-sb\") pod \"b88404c2-d4bf-4257-986a-1b78178fca88\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.000788 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-combined-ca-bundle\") pod \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.000825 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-config\") pod \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.000860 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-scripts\") pod \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.000888 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-ovsdbserver-sb-tls-certs\") pod \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.001069 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" (UID: "081ac19d-af0e-48f7-9c2e-0d282a9bb7ba"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.001111 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4b8fj\" (UniqueName: \"kubernetes.io/projected/b88404c2-d4bf-4257-986a-1b78178fca88-kube-api-access-4b8fj\") pod \"b88404c2-d4bf-4257-986a-1b78178fca88\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.001165 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-metrics-certs-tls-certs\") pod \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.001196 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\" (UID: \"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.001240 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-dns-swift-storage-0\") pod \"b88404c2-d4bf-4257-986a-1b78178fca88\" (UID: \"b88404c2-d4bf-4257-986a-1b78178fca88\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.001923 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6pz4\" (UniqueName: \"kubernetes.io/projected/88635a0d-cb9d-43c1-9086-5623c9a780f8-kube-api-access-b6pz4\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.001946 4810 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/88635a0d-cb9d-43c1-9086-5623c9a780f8-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.001956 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88635a0d-cb9d-43c1-9086-5623c9a780f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.001965 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.001976 4810 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/88635a0d-cb9d-43c1-9086-5623c9a780f8-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.001990 4810 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/88635a0d-cb9d-43c1-9086-5623c9a780f8-var-run\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.004349 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-config" (OuterVolumeSpecName: "config") pod "081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" (UID: "081ac19d-af0e-48f7-9c2e-0d282a9bb7ba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.010072 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.010347 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-scripts" (OuterVolumeSpecName: "scripts") pod "081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" (UID: "081ac19d-af0e-48f7-9c2e-0d282a9bb7ba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.021648 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" (UID: "081ac19d-af0e-48f7-9c2e-0d282a9bb7ba"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.023811 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-kube-api-access-kczs5" (OuterVolumeSpecName: "kube-api-access-kczs5") pod "081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" (UID: "081ac19d-af0e-48f7-9c2e-0d282a9bb7ba"). InnerVolumeSpecName "kube-api-access-kczs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.024357 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b88404c2-d4bf-4257-986a-1b78178fca88-kube-api-access-4b8fj" (OuterVolumeSpecName: "kube-api-access-4b8fj") pod "b88404c2-d4bf-4257-986a-1b78178fca88" (UID: "b88404c2-d4bf-4257-986a-1b78178fca88"). InnerVolumeSpecName "kube-api-access-4b8fj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.026818 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3798bfc4-fb37-48d1-8146-42d177cf7861/ovsdbserver-nb/0.log" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.026932 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.056111 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell1c683-account-delete-46h9z" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.062863 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88635a0d-cb9d-43c1-9086-5623c9a780f8-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "88635a0d-cb9d-43c1-9086-5623c9a780f8" (UID: "88635a0d-cb9d-43c1-9086-5623c9a780f8"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.076406 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b88404c2-d4bf-4257-986a-1b78178fca88" (UID: "b88404c2-d4bf-4257-986a-1b78178fca88"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.102881 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3798bfc4-fb37-48d1-8146-42d177cf7861-ovsdb-rundir\") pod \"3798bfc4-fb37-48d1-8146-42d177cf7861\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.102994 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gzl8\" (UniqueName: \"kubernetes.io/projected/3798bfc4-fb37-48d1-8146-42d177cf7861-kube-api-access-9gzl8\") pod \"3798bfc4-fb37-48d1-8146-42d177cf7861\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.103017 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3798bfc4-fb37-48d1-8146-42d177cf7861-ovsdbserver-nb-tls-certs\") pod \"3798bfc4-fb37-48d1-8146-42d177cf7861\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.103039 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3798bfc4-fb37-48d1-8146-42d177cf7861-metrics-certs-tls-certs\") pod \"3798bfc4-fb37-48d1-8146-42d177cf7861\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.103102 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c35fc22d-59ab-4396-87c5-fadfc17ecc53-openstack-config-secret\") pod \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\" (UID: \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.103140 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c35fc22d-59ab-4396-87c5-fadfc17ecc53-combined-ca-bundle\") pod \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\" (UID: \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.103168 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3798bfc4-fb37-48d1-8146-42d177cf7861-scripts\") pod \"3798bfc4-fb37-48d1-8146-42d177cf7861\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.103186 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3798bfc4-fb37-48d1-8146-42d177cf7861-config\") pod \"3798bfc4-fb37-48d1-8146-42d177cf7861\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.103252 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3798bfc4-fb37-48d1-8146-42d177cf7861-combined-ca-bundle\") pod \"3798bfc4-fb37-48d1-8146-42d177cf7861\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.103302 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c35fc22d-59ab-4396-87c5-fadfc17ecc53-openstack-config\") pod \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\" (UID: \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.103325 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kckmw\" (UniqueName: \"kubernetes.io/projected/c35fc22d-59ab-4396-87c5-fadfc17ecc53-kube-api-access-kckmw\") pod \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\" (UID: \"c35fc22d-59ab-4396-87c5-fadfc17ecc53\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.103372 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"3798bfc4-fb37-48d1-8146-42d177cf7861\" (UID: \"3798bfc4-fb37-48d1-8146-42d177cf7861\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.103773 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.103790 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.103799 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/88635a0d-cb9d-43c1-9086-5623c9a780f8-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.103808 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4b8fj\" (UniqueName: \"kubernetes.io/projected/b88404c2-d4bf-4257-986a-1b78178fca88-kube-api-access-4b8fj\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.103825 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.103836 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kczs5\" (UniqueName: \"kubernetes.io/projected/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-kube-api-access-kczs5\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.103844 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.108968 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement8657-account-delete-hx45b"] Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.109446 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3798bfc4-fb37-48d1-8146-42d177cf7861-config" (OuterVolumeSpecName: "config") pod "3798bfc4-fb37-48d1-8146-42d177cf7861" (UID: "3798bfc4-fb37-48d1-8146-42d177cf7861"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.109489 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c35fc22d-59ab-4396-87c5-fadfc17ecc53-kube-api-access-kckmw" (OuterVolumeSpecName: "kube-api-access-kckmw") pod "c35fc22d-59ab-4396-87c5-fadfc17ecc53" (UID: "c35fc22d-59ab-4396-87c5-fadfc17ecc53"). InnerVolumeSpecName "kube-api-access-kckmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.109630 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" (UID: "081ac19d-af0e-48f7-9c2e-0d282a9bb7ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.109622 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3798bfc4-fb37-48d1-8146-42d177cf7861-kube-api-access-9gzl8" (OuterVolumeSpecName: "kube-api-access-9gzl8") pod "3798bfc4-fb37-48d1-8146-42d177cf7861" (UID: "3798bfc4-fb37-48d1-8146-42d177cf7861"). InnerVolumeSpecName "kube-api-access-9gzl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.110982 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3798bfc4-fb37-48d1-8146-42d177cf7861-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "3798bfc4-fb37-48d1-8146-42d177cf7861" (UID: "3798bfc4-fb37-48d1-8146-42d177cf7861"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.114253 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3798bfc4-fb37-48d1-8146-42d177cf7861-scripts" (OuterVolumeSpecName: "scripts") pod "3798bfc4-fb37-48d1-8146-42d177cf7861" (UID: "3798bfc4-fb37-48d1-8146-42d177cf7861"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.122011 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "3798bfc4-fb37-48d1-8146-42d177cf7861" (UID: "3798bfc4-fb37-48d1-8146-42d177cf7861"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.129398 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-config" (OuterVolumeSpecName: "config") pod "b88404c2-d4bf-4257-986a-1b78178fca88" (UID: "b88404c2-d4bf-4257-986a-1b78178fca88"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.150168 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35fc22d-59ab-4396-87c5-fadfc17ecc53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c35fc22d-59ab-4396-87c5-fadfc17ecc53" (UID: "c35fc22d-59ab-4396-87c5-fadfc17ecc53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.163705 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c35fc22d-59ab-4396-87c5-fadfc17ecc53-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "c35fc22d-59ab-4396-87c5-fadfc17ecc53" (UID: "c35fc22d-59ab-4396-87c5-fadfc17ecc53"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.166330 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b88404c2-d4bf-4257-986a-1b78178fca88" (UID: "b88404c2-d4bf-4257-986a-1b78178fca88"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.205378 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/068d65be-c5cf-4198-a7ce-0aedeb5e9985-operator-scripts\") pod \"novacell1c683-account-delete-46h9z\" (UID: \"068d65be-c5cf-4198-a7ce-0aedeb5e9985\") " pod="openstack/novacell1c683-account-delete-46h9z" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.205489 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9rn6\" (UniqueName: \"kubernetes.io/projected/068d65be-c5cf-4198-a7ce-0aedeb5e9985-kube-api-access-f9rn6\") pod \"novacell1c683-account-delete-46h9z\" (UID: \"068d65be-c5cf-4198-a7ce-0aedeb5e9985\") " pod="openstack/novacell1c683-account-delete-46h9z" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.205585 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.205620 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c35fc22d-59ab-4396-87c5-fadfc17ecc53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.205632 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3798bfc4-fb37-48d1-8146-42d177cf7861-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.205640 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3798bfc4-fb37-48d1-8146-42d177cf7861-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.205649 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c35fc22d-59ab-4396-87c5-fadfc17ecc53-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.205658 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kckmw\" (UniqueName: \"kubernetes.io/projected/c35fc22d-59ab-4396-87c5-fadfc17ecc53-kube-api-access-kckmw\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.205682 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.205691 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3798bfc4-fb37-48d1-8146-42d177cf7861-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.205701 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.205709 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.205717 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gzl8\" (UniqueName: \"kubernetes.io/projected/3798bfc4-fb37-48d1-8146-42d177cf7861-kube-api-access-9gzl8\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: E1201 14:59:47.206191 4810 secret.go:188] Couldn't get secret openstack/neutron-httpd-config: secret "neutron-httpd-config" not found Dec 01 14:59:47 crc kubenswrapper[4810]: E1201 14:59:47.206225 4810 secret.go:188] Couldn't get secret openstack/neutron-config: secret "neutron-config" not found Dec 01 14:59:47 crc kubenswrapper[4810]: E1201 14:59:47.206269 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-httpd-config podName:ba5f8a69-22cf-4196-bc89-ffa60eda9303 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:49.206251056 +0000 UTC m=+1554.969760659 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "httpd-config" (UniqueName: "kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-httpd-config") pod "neutron-5477779d95-wg2vp" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303") : secret "neutron-httpd-config" not found Dec 01 14:59:47 crc kubenswrapper[4810]: E1201 14:59:47.206298 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-config podName:ba5f8a69-22cf-4196-bc89-ffa60eda9303 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:49.206279686 +0000 UTC m=+1554.969789289 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-config") pod "neutron-5477779d95-wg2vp" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303") : secret "neutron-config" not found Dec 01 14:59:47 crc kubenswrapper[4810]: E1201 14:59:47.207219 4810 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Dec 01 14:59:47 crc kubenswrapper[4810]: E1201 14:59:47.207392 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/068d65be-c5cf-4198-a7ce-0aedeb5e9985-operator-scripts podName:068d65be-c5cf-4198-a7ce-0aedeb5e9985 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:49.207381107 +0000 UTC m=+1554.970890710 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/068d65be-c5cf-4198-a7ce-0aedeb5e9985-operator-scripts") pod "novacell1c683-account-delete-46h9z" (UID: "068d65be-c5cf-4198-a7ce-0aedeb5e9985") : configmap "openstack-cell1-scripts" not found Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.211806 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" (UID: "081ac19d-af0e-48f7-9c2e-0d282a9bb7ba"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: E1201 14:59:47.212056 4810 projected.go:194] Error preparing data for projected volume kube-api-access-f9rn6 for pod openstack/novacell1c683-account-delete-46h9z: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 01 14:59:47 crc kubenswrapper[4810]: E1201 14:59:47.212111 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/068d65be-c5cf-4198-a7ce-0aedeb5e9985-kube-api-access-f9rn6 podName:068d65be-c5cf-4198-a7ce-0aedeb5e9985 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:49.212096034 +0000 UTC m=+1554.975605637 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-f9rn6" (UniqueName: "kubernetes.io/projected/068d65be-c5cf-4198-a7ce-0aedeb5e9985-kube-api-access-f9rn6") pod "novacell1c683-account-delete-46h9z" (UID: "068d65be-c5cf-4198-a7ce-0aedeb5e9985") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.212704 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b88404c2-d4bf-4257-986a-1b78178fca88" (UID: "b88404c2-d4bf-4257-986a-1b78178fca88"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.232385 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="d202704a-7a2e-46a6-9e20-2eeabdc11814" containerName="galera" probeResult="failure" output="" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.238462 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.270874 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicanff71-account-delete-2fnl8"] Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.278737 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35fc22d-59ab-4396-87c5-fadfc17ecc53-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "c35fc22d-59ab-4396-87c5-fadfc17ecc53" (UID: "c35fc22d-59ab-4396-87c5-fadfc17ecc53"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.301532 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.308504 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.308531 4810 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.308540 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.308549 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.308558 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c35fc22d-59ab-4396-87c5-fadfc17ecc53-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.313062 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b88404c2-d4bf-4257-986a-1b78178fca88" (UID: "b88404c2-d4bf-4257-986a-1b78178fca88"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: E1201 14:59:47.319515 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 01 14:59:47 crc kubenswrapper[4810]: E1201 14:59:47.319941 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 01 14:59:47 crc kubenswrapper[4810]: E1201 14:59:47.320362 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 01 14:59:47 crc kubenswrapper[4810]: E1201 14:59:47.320676 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 01 14:59:47 crc kubenswrapper[4810]: E1201 14:59:47.320699 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-x82dv" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovsdb-server" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.333597 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3798bfc4-fb37-48d1-8146-42d177cf7861-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3798bfc4-fb37-48d1-8146-42d177cf7861" (UID: "3798bfc4-fb37-48d1-8146-42d177cf7861"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: E1201 14:59:47.350079 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 01 14:59:47 crc kubenswrapper[4810]: E1201 14:59:47.360345 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 01 14:59:47 crc kubenswrapper[4810]: E1201 14:59:47.360414 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-x82dv" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovs-vswitchd" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.376010 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.409879 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7478f42b-4be2-430b-9854-5930dccaee43-log-httpd\") pod \"7478f42b-4be2-430b-9854-5930dccaee43\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.409927 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-config-data\") pod \"7478f42b-4be2-430b-9854-5930dccaee43\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.409945 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-public-tls-certs\") pod \"7478f42b-4be2-430b-9854-5930dccaee43\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.410037 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-combined-ca-bundle\") pod \"7478f42b-4be2-430b-9854-5930dccaee43\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.410059 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7478f42b-4be2-430b-9854-5930dccaee43-etc-swift\") pod \"7478f42b-4be2-430b-9854-5930dccaee43\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.410139 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrx4f\" (UniqueName: \"kubernetes.io/projected/7478f42b-4be2-430b-9854-5930dccaee43-kube-api-access-mrx4f\") pod \"7478f42b-4be2-430b-9854-5930dccaee43\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.410200 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7478f42b-4be2-430b-9854-5930dccaee43-run-httpd\") pod \"7478f42b-4be2-430b-9854-5930dccaee43\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.410231 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-internal-tls-certs\") pod \"7478f42b-4be2-430b-9854-5930dccaee43\" (UID: \"7478f42b-4be2-430b-9854-5930dccaee43\") " Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.410730 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b88404c2-d4bf-4257-986a-1b78178fca88-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.410744 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3798bfc4-fb37-48d1-8146-42d177cf7861-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.413182 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7478f42b-4be2-430b-9854-5930dccaee43-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7478f42b-4be2-430b-9854-5930dccaee43" (UID: "7478f42b-4be2-430b-9854-5930dccaee43"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.413656 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7478f42b-4be2-430b-9854-5930dccaee43-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7478f42b-4be2-430b-9854-5930dccaee43" (UID: "7478f42b-4be2-430b-9854-5930dccaee43"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.433629 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7478f42b-4be2-430b-9854-5930dccaee43-kube-api-access-mrx4f" (OuterVolumeSpecName: "kube-api-access-mrx4f") pod "7478f42b-4be2-430b-9854-5930dccaee43" (UID: "7478f42b-4be2-430b-9854-5930dccaee43"). InnerVolumeSpecName "kube-api-access-mrx4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.436766 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder17c6-account-delete-qktk6"] Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.452899 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7478f42b-4be2-430b-9854-5930dccaee43-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "7478f42b-4be2-430b-9854-5930dccaee43" (UID: "7478f42b-4be2-430b-9854-5930dccaee43"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: W1201 14:59:47.460595 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf15ccdf_1e4e_44ab_8ccb_819502935a7a.slice/crio-0a07bd9197943b1a845068c9151c0945af10f687ff4d4b2b37177546b1a72c14 WatchSource:0}: Error finding container 0a07bd9197943b1a845068c9151c0945af10f687ff4d4b2b37177546b1a72c14: Status 404 returned error can't find the container with id 0a07bd9197943b1a845068c9151c0945af10f687ff4d4b2b37177546b1a72c14 Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.529079 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7478f42b-4be2-430b-9854-5930dccaee43-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.529114 4810 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7478f42b-4be2-430b-9854-5930dccaee43-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.529124 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrx4f\" (UniqueName: \"kubernetes.io/projected/7478f42b-4be2-430b-9854-5930dccaee43-kube-api-access-mrx4f\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.529133 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7478f42b-4be2-430b-9854-5930dccaee43-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.570096 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell096e4-account-delete-qqlg4"] Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.594752 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutronc91e-account-delete-t585q"] Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.654970 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3798bfc4-fb37-48d1-8146-42d177cf7861-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "3798bfc4-fb37-48d1-8146-42d177cf7861" (UID: "3798bfc4-fb37-48d1-8146-42d177cf7861"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.665890 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-config-data" (OuterVolumeSpecName: "config-data") pod "7478f42b-4be2-430b-9854-5930dccaee43" (UID: "7478f42b-4be2-430b-9854-5930dccaee43"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.674614 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" (UID: "081ac19d-af0e-48f7-9c2e-0d282a9bb7ba"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.679865 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapic0fd-account-delete-w7ttc"] Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.683916 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3798bfc4-fb37-48d1-8146-42d177cf7861-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "3798bfc4-fb37-48d1-8146-42d177cf7861" (UID: "3798bfc4-fb37-48d1-8146-42d177cf7861"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.684459 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7478f42b-4be2-430b-9854-5930dccaee43" (UID: "7478f42b-4be2-430b-9854-5930dccaee43"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.685915 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7478f42b-4be2-430b-9854-5930dccaee43" (UID: "7478f42b-4be2-430b-9854-5930dccaee43"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.697560 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7478f42b-4be2-430b-9854-5930dccaee43" (UID: "7478f42b-4be2-430b-9854-5930dccaee43"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.737165 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3798bfc4-fb37-48d1-8146-42d177cf7861-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.737577 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3798bfc4-fb37-48d1-8146-42d177cf7861-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.737587 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.737595 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.737604 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.737616 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.737623 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7478f42b-4be2-430b-9854-5930dccaee43-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:47 crc kubenswrapper[4810]: E1201 14:59:47.737259 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 01 14:59:47 crc kubenswrapper[4810]: E1201 14:59:47.737690 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-config-data podName:3b81b6a9-0be6-4ce1-81b4-043edb1c221f nodeName:}" failed. No retries permitted until 2025-12-01 14:59:51.737673485 +0000 UTC m=+1557.501183088 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-config-data") pod "rabbitmq-server-0" (UID: "3b81b6a9-0be6-4ce1-81b4-043edb1c221f") : configmap "rabbitmq-config-data" not found Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.841721 4810 scope.go:117] "RemoveContainer" containerID="a303d0364f2f0ad764ac7d54b67330e695ac583a1627abd27d1f6447c9de1ee2" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.841744 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.849179 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder17c6-account-delete-qktk6" event={"ID":"bf15ccdf-1e4e-44ab-8ccb-819502935a7a","Type":"ContainerStarted","Data":"0a07bd9197943b1a845068c9151c0945af10f687ff4d4b2b37177546b1a72c14"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.860136 4810 generic.go:334] "Generic (PLEG): container finished" podID="d202704a-7a2e-46a6-9e20-2eeabdc11814" containerID="f21f812c44b7972a5f6ecdd8dbe2ffd46fb63a8fa481dd14afa8b978ba21edfb" exitCode=0 Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.860207 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d202704a-7a2e-46a6-9e20-2eeabdc11814","Type":"ContainerDied","Data":"f21f812c44b7972a5f6ecdd8dbe2ffd46fb63a8fa481dd14afa8b978ba21edfb"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.860233 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d202704a-7a2e-46a6-9e20-2eeabdc11814","Type":"ContainerDied","Data":"118df97aecfd418517e3007d6833b5d4e2b18bd1489f3992ba59f20ddf95abdf"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.860245 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="118df97aecfd418517e3007d6833b5d4e2b18bd1489f3992ba59f20ddf95abdf" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.865587 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanff71-account-delete-2fnl8" event={"ID":"1ef9fa82-f302-47e9-bed3-b02f34a2340f","Type":"ContainerStarted","Data":"c8a7aa776dc68e3946de920183358dd2dab73021442a06e65e4afc5eae2312fe"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.872535 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell096e4-account-delete-qqlg4" event={"ID":"4f7b5a25-eed0-4122-a036-093ac7adca22","Type":"ContainerStarted","Data":"b49f17b2795a2340ebd876f7a3c218b2cc7a3ea65b59733f36cbea047a113f55"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.876681 4810 generic.go:334] "Generic (PLEG): container finished" podID="bef8b0c7-66cd-4930-a78b-bf73804a73d2" containerID="c3e8f637d25ffa49a1106b343b7c3861c122e5ccdeb4f52b034a012b1bf12322" exitCode=0 Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.876741 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bef8b0c7-66cd-4930-a78b-bf73804a73d2","Type":"ContainerDied","Data":"c3e8f637d25ffa49a1106b343b7c3861c122e5ccdeb4f52b034a012b1bf12322"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.883376 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3798bfc4-fb37-48d1-8146-42d177cf7861/ovsdbserver-nb/0.log" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.883445 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3798bfc4-fb37-48d1-8146-42d177cf7861","Type":"ContainerDied","Data":"a30d683bfc270e3e352e7e4bd39449d321a965c5bf2257fce181fbfe593aa51c"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.883539 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.892185 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronc91e-account-delete-t585q" event={"ID":"f7b06553-ec05-4c6f-82c0-ef0b4dd3005a","Type":"ContainerStarted","Data":"6d62c108069647c18de078a2ba918fa95f1dc3fb75187b7cfcc958cebe7a3323"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.895543 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapic0fd-account-delete-w7ttc" event={"ID":"476b71ac-1050-46a6-b51c-168e75b1a350","Type":"ContainerStarted","Data":"6e9b6324957395c9efa53642231b0756aaf310606b0fbc1e11aee7cd643bdc77"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.897285 4810 generic.go:334] "Generic (PLEG): container finished" podID="91932480-76aa-4751-9251-3ad82d849882" containerID="fac78a09d1aaaaf2dc5fd71c00691039693074a0a7b1dee4f7775f2ca6566022" exitCode=0 Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.897331 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance324e-account-delete-6s4dm" event={"ID":"91932480-76aa-4751-9251-3ad82d849882","Type":"ContainerDied","Data":"fac78a09d1aaaaf2dc5fd71c00691039693074a0a7b1dee4f7775f2ca6566022"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.897347 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance324e-account-delete-6s4dm" event={"ID":"91932480-76aa-4751-9251-3ad82d849882","Type":"ContainerStarted","Data":"9e70e548e2a90c425ccea9e2119ed1893ced6c65cb1111e0b6cab49cbcbdf3ea"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.900901 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_081ac19d-af0e-48f7-9c2e-0d282a9bb7ba/ovsdbserver-sb/0.log" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.900952 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"081ac19d-af0e-48f7-9c2e-0d282a9bb7ba","Type":"ContainerDied","Data":"31554da4b756f2b0099d3a4a3d6c074cfcc4bca38c34e243462873eed47e271b"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.901006 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.914765 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" event={"ID":"b88404c2-d4bf-4257-986a-1b78178fca88","Type":"ContainerDied","Data":"ee8a52a8581a97da6378bf0cfba49cdaf6dd1156e9099284a43ed8cf18689572"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.914879 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.919680 4810 generic.go:334] "Generic (PLEG): container finished" podID="ba5f8a69-22cf-4196-bc89-ffa60eda9303" containerID="8439e2334c700fa63f3562d0a3892b76d49cb65b3961834e76812a6b483afe47" exitCode=0 Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.919737 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5477779d95-wg2vp" event={"ID":"ba5f8a69-22cf-4196-bc89-ffa60eda9303","Type":"ContainerDied","Data":"8439e2334c700fa63f3562d0a3892b76d49cb65b3961834e76812a6b483afe47"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.923401 4810 generic.go:334] "Generic (PLEG): container finished" podID="9acd8c29-93f5-4b0f-8b32-d8f3d576767b" containerID="6c951f3e299a57aa9f4c575d8d057fd04b80e714ae7050a14a4ca0758c9a0549" exitCode=143 Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.923452 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9acd8c29-93f5-4b0f-8b32-d8f3d576767b","Type":"ContainerDied","Data":"6c951f3e299a57aa9f4c575d8d057fd04b80e714ae7050a14a4ca0758c9a0549"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.925598 4810 generic.go:334] "Generic (PLEG): container finished" podID="7478f42b-4be2-430b-9854-5930dccaee43" containerID="1cc6810ebfaa94211940478b494df0cc163de04d06f714491ab4d71d2b9f032d" exitCode=0 Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.925632 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.925708 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" event={"ID":"7478f42b-4be2-430b-9854-5930dccaee43","Type":"ContainerDied","Data":"1cc6810ebfaa94211940478b494df0cc163de04d06f714491ab4d71d2b9f032d"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.925737 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6c8f8b5cf7-65pxh" event={"ID":"7478f42b-4be2-430b-9854-5930dccaee43","Type":"ContainerDied","Data":"4cfb038fda022537c9c1494bc93f9700a37720afcac62a202c221904f4195331"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.927230 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-hfmpn" event={"ID":"88635a0d-cb9d-43c1-9086-5623c9a780f8","Type":"ContainerDied","Data":"b83750ee8d6e70fe32615bc44082261ab5ed545cea6cd0bf0ff4b0eacb8f8dc5"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.927279 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-hfmpn" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.936374 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement8657-account-delete-hx45b" event={"ID":"2a3a8b5d-01e4-4396-8a9b-abd85dc55572","Type":"ContainerStarted","Data":"a54201a119ae4e648033f69b4c2cda9c0c2342bcf83446696e259213e7c43dfe"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.936433 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement8657-account-delete-hx45b" event={"ID":"2a3a8b5d-01e4-4396-8a9b-abd85dc55572","Type":"ContainerStarted","Data":"44593a6355c67c2f6ce5902d677ecce531982fb97d53a6bc836d3bdf9d804c4b"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.948674 4810 generic.go:334] "Generic (PLEG): container finished" podID="0899579c-adaa-41fb-a36d-31ba657c6c1a" containerID="7f92191f00ad28697cca6ad61e70b41f045bb53297c7b28a21e7c29a04def64d" exitCode=0 Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.948750 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell1c683-account-delete-46h9z" Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.949513 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0899579c-adaa-41fb-a36d-31ba657c6c1a","Type":"ContainerDied","Data":"7f92191f00ad28697cca6ad61e70b41f045bb53297c7b28a21e7c29a04def64d"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.950630 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0899579c-adaa-41fb-a36d-31ba657c6c1a","Type":"ContainerDied","Data":"2fca946a71c8696f06f213361f8621fab5f339501bd9ae641ecaca38a90be485"} Dec 01 14:59:47 crc kubenswrapper[4810]: I1201 14:59:47.950657 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fca946a71c8696f06f213361f8621fab5f339501bd9ae641ecaca38a90be485" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.025985 4810 scope.go:117] "RemoveContainer" containerID="d74f9aed9f56dc44f786f3d28286286934f841d05512a1633b5fe3b4cc217204" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.191406 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.209170 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.209458 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerName="ceilometer-central-agent" containerID="cri-o://b1b79938d10b71fee29a6515a87d1fb415850230bcaa5ce800b1e5ba6bb61661" gracePeriod=30 Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.209735 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerName="proxy-httpd" containerID="cri-o://ac4ab87e1191fa810eaa9e1774fc31d9ed1e7d6475958ea936337c2a47b7b3e0" gracePeriod=30 Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.209791 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerName="sg-core" containerID="cri-o://90f166a6125c50147c3cb78e63b7459cc98d8a4218a7a6ddd44ce0fb9b8a4fc2" gracePeriod=30 Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.209833 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerName="ceilometer-notification-agent" containerID="cri-o://3cb0e1b4c05926fcfa43bb4c227b62243d1a9c4234879cdb02f9c40af684935e" gracePeriod=30 Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.249863 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-vencrypt-tls-certs\") pod \"0899579c-adaa-41fb-a36d-31ba657c6c1a\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.250875 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-config-data\") pod \"0899579c-adaa-41fb-a36d-31ba657c6c1a\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.250975 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8fc8\" (UniqueName: \"kubernetes.io/projected/0899579c-adaa-41fb-a36d-31ba657c6c1a-kube-api-access-z8fc8\") pod \"0899579c-adaa-41fb-a36d-31ba657c6c1a\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.251092 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-combined-ca-bundle\") pod \"0899579c-adaa-41fb-a36d-31ba657c6c1a\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.251259 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-nova-novncproxy-tls-certs\") pod \"0899579c-adaa-41fb-a36d-31ba657c6c1a\" (UID: \"0899579c-adaa-41fb-a36d-31ba657c6c1a\") " Dec 01 14:59:48 crc kubenswrapper[4810]: E1201 14:59:48.251848 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 01 14:59:48 crc kubenswrapper[4810]: E1201 14:59:48.251974 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-config-data podName:56bbf32f-35b5-4f92-bfc0-9266b0ecccac nodeName:}" failed. No retries permitted until 2025-12-01 14:59:52.251958388 +0000 UTC m=+1558.015467991 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-config-data") pod "rabbitmq-cell1-server-0" (UID: "56bbf32f-35b5-4f92-bfc0-9266b0ecccac") : configmap "rabbitmq-cell1-config-data" not found Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.293265 4810 scope.go:117] "RemoveContainer" containerID="afd5408dd106d9875d318c4f58d0916d20a0514bcf3f71e7b57c27a91050bc80" Dec 01 14:59:48 crc kubenswrapper[4810]: E1201 14:59:48.302141 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="87439695570ef9b3e9da98681ae2d1dec7fb73d967ccb8912a2e40ca13529cf7" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 01 14:59:48 crc kubenswrapper[4810]: E1201 14:59:48.315903 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="87439695570ef9b3e9da98681ae2d1dec7fb73d967ccb8912a2e40ca13529cf7" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.332573 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0899579c-adaa-41fb-a36d-31ba657c6c1a-kube-api-access-z8fc8" (OuterVolumeSpecName: "kube-api-access-z8fc8") pod "0899579c-adaa-41fb-a36d-31ba657c6c1a" (UID: "0899579c-adaa-41fb-a36d-31ba657c6c1a"). InnerVolumeSpecName "kube-api-access-z8fc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:48 crc kubenswrapper[4810]: E1201 14:59:48.335158 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="87439695570ef9b3e9da98681ae2d1dec7fb73d967ccb8912a2e40ca13529cf7" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 01 14:59:48 crc kubenswrapper[4810]: E1201 14:59:48.335237 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="604a8b79-fcf3-40cf-82e6-8af2cf667b0f" containerName="ovn-northd" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.347515 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.361122 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8fc8\" (UniqueName: \"kubernetes.io/projected/0899579c-adaa-41fb-a36d-31ba657c6c1a-kube-api-access-z8fc8\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.412676 4810 scope.go:117] "RemoveContainer" containerID="6fcdb0e74e2281103d173be66fd77a226460e47c85adf65b5481538a371e0a73" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.427795 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0899579c-adaa-41fb-a36d-31ba657c6c1a" (UID: "0899579c-adaa-41fb-a36d-31ba657c6c1a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.463158 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d202704a-7a2e-46a6-9e20-2eeabdc11814-operator-scripts\") pod \"d202704a-7a2e-46a6-9e20-2eeabdc11814\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.463265 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8gbp\" (UniqueName: \"kubernetes.io/projected/d202704a-7a2e-46a6-9e20-2eeabdc11814-kube-api-access-d8gbp\") pod \"d202704a-7a2e-46a6-9e20-2eeabdc11814\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.463347 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d202704a-7a2e-46a6-9e20-2eeabdc11814-kolla-config\") pod \"d202704a-7a2e-46a6-9e20-2eeabdc11814\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.463383 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d202704a-7a2e-46a6-9e20-2eeabdc11814-galera-tls-certs\") pod \"d202704a-7a2e-46a6-9e20-2eeabdc11814\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.463451 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"d202704a-7a2e-46a6-9e20-2eeabdc11814\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.463514 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d202704a-7a2e-46a6-9e20-2eeabdc11814-combined-ca-bundle\") pod \"d202704a-7a2e-46a6-9e20-2eeabdc11814\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.463624 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d202704a-7a2e-46a6-9e20-2eeabdc11814-config-data-generated\") pod \"d202704a-7a2e-46a6-9e20-2eeabdc11814\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.463656 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d202704a-7a2e-46a6-9e20-2eeabdc11814-config-data-default\") pod \"d202704a-7a2e-46a6-9e20-2eeabdc11814\" (UID: \"d202704a-7a2e-46a6-9e20-2eeabdc11814\") " Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.478118 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d202704a-7a2e-46a6-9e20-2eeabdc11814-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "d202704a-7a2e-46a6-9e20-2eeabdc11814" (UID: "d202704a-7a2e-46a6-9e20-2eeabdc11814"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.481864 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.482302 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d202704a-7a2e-46a6-9e20-2eeabdc11814-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "d202704a-7a2e-46a6-9e20-2eeabdc11814" (UID: "d202704a-7a2e-46a6-9e20-2eeabdc11814"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.482306 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d202704a-7a2e-46a6-9e20-2eeabdc11814-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "d202704a-7a2e-46a6-9e20-2eeabdc11814" (UID: "d202704a-7a2e-46a6-9e20-2eeabdc11814"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.484422 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d202704a-7a2e-46a6-9e20-2eeabdc11814-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d202704a-7a2e-46a6-9e20-2eeabdc11814" (UID: "d202704a-7a2e-46a6-9e20-2eeabdc11814"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.518986 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d202704a-7a2e-46a6-9e20-2eeabdc11814-kube-api-access-d8gbp" (OuterVolumeSpecName: "kube-api-access-d8gbp") pod "d202704a-7a2e-46a6-9e20-2eeabdc11814" (UID: "d202704a-7a2e-46a6-9e20-2eeabdc11814"). InnerVolumeSpecName "kube-api-access-d8gbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.589532 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d202704a-7a2e-46a6-9e20-2eeabdc11814-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.589575 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d202704a-7a2e-46a6-9e20-2eeabdc11814-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.589589 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d202704a-7a2e-46a6-9e20-2eeabdc11814-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.589600 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8gbp\" (UniqueName: \"kubernetes.io/projected/d202704a-7a2e-46a6-9e20-2eeabdc11814-kube-api-access-d8gbp\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.589612 4810 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d202704a-7a2e-46a6-9e20-2eeabdc11814-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.654489 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "mysql-db") pod "d202704a-7a2e-46a6-9e20-2eeabdc11814" (UID: "d202704a-7a2e-46a6-9e20-2eeabdc11814"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.667517 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "0899579c-adaa-41fb-a36d-31ba657c6c1a" (UID: "0899579c-adaa-41fb-a36d-31ba657c6c1a"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.671443 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-config-data" (OuterVolumeSpecName: "config-data") pod "0899579c-adaa-41fb-a36d-31ba657c6c1a" (UID: "0899579c-adaa-41fb-a36d-31ba657c6c1a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.677199 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c35fc22d-59ab-4396-87c5-fadfc17ecc53" path="/var/lib/kubelet/pods/c35fc22d-59ab-4396-87c5-fadfc17ecc53/volumes" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.697095 4810 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.697484 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.697513 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 01 14:59:48 crc kubenswrapper[4810]: E1201 14:59:48.787005 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc35fc22d_59ab_4396_87c5_fadfc17ecc53.slice/crio-20d0aa4e26b165bc3f4355044df83264b401d34a1b27cc603c5dce4b4a255062\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3798bfc4_fb37_48d1_8146_42d177cf7861.slice/crio-a30d683bfc270e3e352e7e4bd39449d321a965c5bf2257fce181fbfe593aa51c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88635a0d_cb9d_43c1_9086_5623c9a780f8.slice/crio-b83750ee8d6e70fe32615bc44082261ab5ed545cea6cd0bf0ff4b0eacb8f8dc5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod081ac19d_af0e_48f7_9c2e_0d282a9bb7ba.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod068d65be_c5cf_4198_a7ce_0aedeb5e9985.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7478f42b_4be2_430b_9854_5930dccaee43.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb88404c2_d4bf_4257_986a_1b78178fca88.slice/crio-ee8a52a8581a97da6378bf0cfba49cdaf6dd1156e9099284a43ed8cf18689572\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod081ac19d_af0e_48f7_9c2e_0d282a9bb7ba.slice/crio-31554da4b756f2b0099d3a4a3d6c074cfcc4bca38c34e243462873eed47e271b\": RecentStats: unable to find data in memory cache]" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.851317 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d202704a-7a2e-46a6-9e20-2eeabdc11814-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d202704a-7a2e-46a6-9e20-2eeabdc11814" (UID: "d202704a-7a2e-46a6-9e20-2eeabdc11814"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.914352 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d202704a-7a2e-46a6-9e20-2eeabdc11814-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.923008 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 01 14:59:48 crc kubenswrapper[4810]: I1201 14:59:48.947829 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "0899579c-adaa-41fb-a36d-31ba657c6c1a" (UID: "0899579c-adaa-41fb-a36d-31ba657c6c1a"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.004707 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d202704a-7a2e-46a6-9e20-2eeabdc11814-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "d202704a-7a2e-46a6-9e20-2eeabdc11814" (UID: "d202704a-7a2e-46a6-9e20-2eeabdc11814"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.005676 4810 generic.go:334] "Generic (PLEG): container finished" podID="31b83b20-83c4-4e06-8eb0-1e1d83d4b160" containerID="c81abf769f22e0110d5df574ceaf9e47a00a5723bf172e5cebaebd7bde8ecee2" exitCode=0 Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.020228 4810 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d202704a-7a2e-46a6-9e20-2eeabdc11814-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.020280 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.020294 4810 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/0899579c-adaa-41fb-a36d-31ba657c6c1a-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.028975 4810 generic.go:334] "Generic (PLEG): container finished" podID="2a3a8b5d-01e4-4396-8a9b-abd85dc55572" containerID="a54201a119ae4e648033f69b4c2cda9c0c2342bcf83446696e259213e7c43dfe" exitCode=0 Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.047153 4810 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novaapic0fd-account-delete-w7ttc" secret="" err="secret \"galera-openstack-dockercfg-5jwrv\" not found" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.075893 4810 generic.go:334] "Generic (PLEG): container finished" podID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerID="ac4ab87e1191fa810eaa9e1774fc31d9ed1e7d6475958ea936337c2a47b7b3e0" exitCode=0 Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.075962 4810 generic.go:334] "Generic (PLEG): container finished" podID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerID="90f166a6125c50147c3cb78e63b7459cc98d8a4218a7a6ddd44ce0fb9b8a4fc2" exitCode=2 Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.075976 4810 generic.go:334] "Generic (PLEG): container finished" podID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerID="b1b79938d10b71fee29a6515a87d1fb415850230bcaa5ce800b1e5ba6bb61661" exitCode=0 Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.080233 4810 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/neutronc91e-account-delete-t585q" secret="" err="secret \"galera-openstack-dockercfg-5jwrv\" not found" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.100719 4810 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/cinder17c6-account-delete-qktk6" secret="" err="secret \"galera-openstack-dockercfg-5jwrv\" not found" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.125845 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.128563 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.128622 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts podName:476b71ac-1050-46a6-b51c-168e75b1a350 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:49.628607334 +0000 UTC m=+1555.392116937 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts") pod "novaapic0fd-account-delete-w7ttc" (UID: "476b71ac-1050-46a6-b51c-168e75b1a350") : configmap "openstack-scripts" not found Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.125846 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.134307 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" podUID="bdd90cbb-4242-4227-9de4-09f99e60a516" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:42390->10.217.0.156:9311: read: connection reset by peer" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.134424 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" podUID="bdd90cbb-4242-4227-9de4-09f99e60a516" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:42388->10.217.0.156:9311: read: connection reset by peer" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.135776 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6c5f549c7d-52fkz" event={"ID":"31b83b20-83c4-4e06-8eb0-1e1d83d4b160","Type":"ContainerDied","Data":"c81abf769f22e0110d5df574ceaf9e47a00a5723bf172e5cebaebd7bde8ecee2"} Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.135819 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.135837 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement8657-account-delete-hx45b" event={"ID":"2a3a8b5d-01e4-4396-8a9b-abd85dc55572","Type":"ContainerDied","Data":"a54201a119ae4e648033f69b4c2cda9c0c2342bcf83446696e259213e7c43dfe"} Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.135850 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.135861 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-6c8f8b5cf7-65pxh"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.135874 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-6c8f8b5cf7-65pxh"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.135890 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.135900 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.135910 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-v9d27"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.135919 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-tb6zm"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.135929 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-v9d27"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.135940 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-65bbfc7449-rtqm6"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.135953 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone8cd1-account-delete-c69f6"] Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.136292 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b88404c2-d4bf-4257-986a-1b78178fca88" containerName="dnsmasq-dns" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136305 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b88404c2-d4bf-4257-986a-1b78178fca88" containerName="dnsmasq-dns" Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.136316 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" containerName="openstack-network-exporter" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136322 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" containerName="openstack-network-exporter" Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.136332 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b88404c2-d4bf-4257-986a-1b78178fca88" containerName="init" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136338 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b88404c2-d4bf-4257-986a-1b78178fca88" containerName="init" Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.136352 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88635a0d-cb9d-43c1-9086-5623c9a780f8" containerName="ovn-controller" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136357 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="88635a0d-cb9d-43c1-9086-5623c9a780f8" containerName="ovn-controller" Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.136367 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7478f42b-4be2-430b-9854-5930dccaee43" containerName="proxy-server" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136372 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="7478f42b-4be2-430b-9854-5930dccaee43" containerName="proxy-server" Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.136384 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3798bfc4-fb37-48d1-8146-42d177cf7861" containerName="openstack-network-exporter" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136389 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3798bfc4-fb37-48d1-8146-42d177cf7861" containerName="openstack-network-exporter" Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.136404 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0899579c-adaa-41fb-a36d-31ba657c6c1a" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136409 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0899579c-adaa-41fb-a36d-31ba657c6c1a" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.136420 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" containerName="ovsdbserver-sb" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136426 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" containerName="ovsdbserver-sb" Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.136436 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d202704a-7a2e-46a6-9e20-2eeabdc11814" containerName="mysql-bootstrap" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136443 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d202704a-7a2e-46a6-9e20-2eeabdc11814" containerName="mysql-bootstrap" Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.136459 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d202704a-7a2e-46a6-9e20-2eeabdc11814" containerName="galera" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136483 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d202704a-7a2e-46a6-9e20-2eeabdc11814" containerName="galera" Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.136495 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3798bfc4-fb37-48d1-8146-42d177cf7861" containerName="ovsdbserver-nb" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136501 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3798bfc4-fb37-48d1-8146-42d177cf7861" containerName="ovsdbserver-nb" Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.136511 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7478f42b-4be2-430b-9854-5930dccaee43" containerName="proxy-httpd" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136516 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="7478f42b-4be2-430b-9854-5930dccaee43" containerName="proxy-httpd" Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.136524 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd2094a7-9a7b-417b-b4ea-b1da8b69cac6" containerName="openstack-network-exporter" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136530 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd2094a7-9a7b-417b-b4ea-b1da8b69cac6" containerName="openstack-network-exporter" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136692 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="7478f42b-4be2-430b-9854-5930dccaee43" containerName="proxy-httpd" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136707 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" containerName="ovsdbserver-sb" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136715 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="7478f42b-4be2-430b-9854-5930dccaee43" containerName="proxy-server" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136725 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0899579c-adaa-41fb-a36d-31ba657c6c1a" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136734 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="88635a0d-cb9d-43c1-9086-5623c9a780f8" containerName="ovn-controller" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136743 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d202704a-7a2e-46a6-9e20-2eeabdc11814" containerName="galera" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136750 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3798bfc4-fb37-48d1-8146-42d177cf7861" containerName="ovsdbserver-nb" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136759 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3798bfc4-fb37-48d1-8146-42d177cf7861" containerName="openstack-network-exporter" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136771 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" containerName="openstack-network-exporter" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136783 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b88404c2-d4bf-4257-986a-1b78178fca88" containerName="dnsmasq-dns" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.136796 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd2094a7-9a7b-417b-b4ea-b1da8b69cac6" containerName="openstack-network-exporter" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137355 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-tb6zm"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137372 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3","Type":"ContainerDied","Data":"ac4ab87e1191fa810eaa9e1774fc31d9ed1e7d6475958ea936337c2a47b7b3e0"} Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137387 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone8cd1-account-delete-c69f6"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137397 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137409 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137419 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137430 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3","Type":"ContainerDied","Data":"90f166a6125c50147c3cb78e63b7459cc98d8a4218a7a6ddd44ce0fb9b8a4fc2"} Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137442 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3","Type":"ContainerDied","Data":"b1b79938d10b71fee29a6515a87d1fb415850230bcaa5ce800b1e5ba6bb61661"} Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137451 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-gcsr9"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137461 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronc91e-account-delete-t585q" event={"ID":"f7b06553-ec05-4c6f-82c0-ef0b4dd3005a","Type":"ContainerStarted","Data":"2ba07f4b84a146833c2e8b31369eaa1e961a68fc47ee9b4c60c092208482cff9"} Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137489 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-gcsr9"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137500 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-8cd1-account-create-update-xj8tj"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137509 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-8cd1-account-create-update-xj8tj"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137519 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder17c6-account-delete-qktk6" event={"ID":"bf15ccdf-1e4e-44ab-8ccb-819502935a7a","Type":"ContainerStarted","Data":"afb9f73f83766bd0ed43d8d077992ad62bc5c10098a5767cbf7c3f030466f40f"} Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137530 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone8cd1-account-delete-c69f6"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137538 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-n85zz"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137548 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-n85zz"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137557 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-324e-account-create-update-rdb5j"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137715 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="39ce8411-05ee-4022-bd7a-87e46c2cee8d" containerName="kube-state-metrics" containerID="cri-o://030a775eb17300c273b82602d862651690419af2f0861ccfe4b4d6bdc64e3a5e" gracePeriod=30 Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.137884 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="e16305af-6573-4cfd-ae76-f596d410920b" containerName="memcached" containerID="cri-o://6a66d0aa3094f13ae73e9e79b4f1f5db114907ff336ae2e84fa9c09a936675a3" gracePeriod=30 Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.138058 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-65bbfc7449-rtqm6" podUID="5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0" containerName="keystone-api" containerID="cri-o://e064be1964d87f2e16398d0a402ce3f57b2caf18c89d6b179f3ef2a94e1374ac" gracePeriod=30 Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.138124 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone8cd1-account-delete-c69f6" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.160318 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novaapic0fd-account-delete-w7ttc" podStartSLOduration=4.160291656 podStartE2EDuration="4.160291656s" podCreationTimestamp="2025-12-01 14:59:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:59:49.077231748 +0000 UTC m=+1554.840741351" watchObservedRunningTime="2025-12-01 14:59:49.160291656 +0000 UTC m=+1554.923801259" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.183404 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-324e-account-create-update-rdb5j"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.191610 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": read tcp 10.217.0.2:40254->10.217.0.200:8775: read: connection reset by peer" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.191656 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": read tcp 10.217.0.2:40242->10.217.0.200:8775: read: connection reset by peer" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.202526 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance324e-account-delete-6s4dm"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.203146 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutronc91e-account-delete-t585q" podStartSLOduration=5.20312156 podStartE2EDuration="5.20312156s" podCreationTimestamp="2025-12-01 14:59:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:59:49.108154258 +0000 UTC m=+1554.871663861" watchObservedRunningTime="2025-12-01 14:59:49.20312156 +0000 UTC m=+1554.966631183" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.240260 4810 scope.go:117] "RemoveContainer" containerID="2bb8d93447e17f91799220c99cae83b3109ca92b16e6bbce7a1bf9da7f081d60" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.246791 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/068d65be-c5cf-4198-a7ce-0aedeb5e9985-operator-scripts\") pod \"novacell1c683-account-delete-46h9z\" (UID: \"068d65be-c5cf-4198-a7ce-0aedeb5e9985\") " pod="openstack/novacell1c683-account-delete-46h9z" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.246942 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9rn6\" (UniqueName: \"kubernetes.io/projected/068d65be-c5cf-4198-a7ce-0aedeb5e9985-kube-api-access-f9rn6\") pod \"novacell1c683-account-delete-46h9z\" (UID: \"068d65be-c5cf-4198-a7ce-0aedeb5e9985\") " pod="openstack/novacell1c683-account-delete-46h9z" Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.247972 4810 secret.go:188] Couldn't get secret openstack/neutron-config: secret "neutron-config" not found Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.248074 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-config podName:ba5f8a69-22cf-4196-bc89-ffa60eda9303 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:53.248048022 +0000 UTC m=+1559.011557625 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-config") pod "neutron-5477779d95-wg2vp" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303") : secret "neutron-config" not found Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.249044 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.260103 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts podName:bf15ccdf-1e4e-44ab-8ccb-819502935a7a nodeName:}" failed. No retries permitted until 2025-12-01 14:59:49.760059748 +0000 UTC m=+1555.523569361 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts") pod "cinder17c6-account-delete-qktk6" (UID: "bf15ccdf-1e4e-44ab-8ccb-819502935a7a") : configmap "openstack-scripts" not found Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.252177 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.260180 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts podName:f7b06553-ec05-4c6f-82c0-ef0b4dd3005a nodeName:}" failed. No retries permitted until 2025-12-01 14:59:49.760171951 +0000 UTC m=+1555.523681554 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts") pod "neutronc91e-account-delete-t585q" (UID: "f7b06553-ec05-4c6f-82c0-ef0b4dd3005a") : configmap "openstack-scripts" not found Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.254374 4810 secret.go:188] Couldn't get secret openstack/neutron-httpd-config: secret "neutron-httpd-config" not found Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.260226 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-httpd-config podName:ba5f8a69-22cf-4196-bc89-ffa60eda9303 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:53.260217712 +0000 UTC m=+1559.023727315 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "httpd-config" (UniqueName: "kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-httpd-config") pod "neutron-5477779d95-wg2vp" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303") : secret "neutron-httpd-config" not found Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.256077 4810 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: object "openstack"/"openstack-cell1-scripts" not registered Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.260278 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/068d65be-c5cf-4198-a7ce-0aedeb5e9985-operator-scripts podName:068d65be-c5cf-4198-a7ce-0aedeb5e9985 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:53.260258183 +0000 UTC m=+1559.023767786 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/068d65be-c5cf-4198-a7ce-0aedeb5e9985-operator-scripts") pod "novacell1c683-account-delete-46h9z" (UID: "068d65be-c5cf-4198-a7ce-0aedeb5e9985") : object "openstack"/"openstack-cell1-scripts" not registered Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.256343 4810 projected.go:194] Error preparing data for projected volume kube-api-access-f9rn6 for pod openstack/novacell1c683-account-delete-46h9z: failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.260320 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/068d65be-c5cf-4198-a7ce-0aedeb5e9985-kube-api-access-f9rn6 podName:068d65be-c5cf-4198-a7ce-0aedeb5e9985 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:53.260307815 +0000 UTC m=+1559.023817418 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-f9rn6" (UniqueName: "kubernetes.io/projected/068d65be-c5cf-4198-a7ce-0aedeb5e9985-kube-api-access-f9rn6") pod "novacell1c683-account-delete-46h9z" (UID: "068d65be-c5cf-4198-a7ce-0aedeb5e9985") : failed to fetch token: serviceaccounts "galera-openstack-cell1" not found Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.261988 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder17c6-account-delete-qktk6" podStartSLOduration=5.26197184 podStartE2EDuration="5.26197184s" podCreationTimestamp="2025-12-01 14:59:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:59:49.125796557 +0000 UTC m=+1554.889306160" watchObservedRunningTime="2025-12-01 14:59:49.26197184 +0000 UTC m=+1555.025481443" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.433871 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-hfmpn"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.443417 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="e79c5698-ab2d-4292-bda9-fab8dbbb2abb" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.162:8776/healthcheck\": read tcp 10.217.0.2:51022->10.217.0.162:8776: read: connection reset by peer" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.481545 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-hfmpn"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.514867 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.554531 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.629086 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell1c683-account-delete-46h9z"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.629147 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell1c683-account-delete-46h9z"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.635014 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-prlzg"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.645916 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-prlzg"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.660200 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9rn6\" (UniqueName: \"kubernetes.io/projected/068d65be-c5cf-4198-a7ce-0aedeb5e9985-kube-api-access-f9rn6\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.660224 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/068d65be-c5cf-4198-a7ce-0aedeb5e9985-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.660292 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.660333 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts podName:476b71ac-1050-46a6-b51c-168e75b1a350 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:50.660318901 +0000 UTC m=+1556.423828504 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts") pod "novaapic0fd-account-delete-w7ttc" (UID: "476b71ac-1050-46a6-b51c-168e75b1a350") : configmap "openstack-scripts" not found Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.675925 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-8657-account-create-update-jsmmz"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.682329 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement8657-account-delete-hx45b"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.683263 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="4c309abf-824f-4264-859b-ded1f0a9acaf" containerName="galera" containerID="cri-o://1b9b162f9bfb794c639b6d75ebcad8b08b8c31a6a252c25ffdb8e930eaa77acf" gracePeriod=30 Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.696295 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-8657-account-create-update-jsmmz"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.715700 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-wt46c"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.723020 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-wt46c"] Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.723779 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ffc643854f69e1e2f23a40254dbae2944125c570cca2c9e027760a19987124a6" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.734545 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-ff71-account-create-update-c62lr"] Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.756075 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ffc643854f69e1e2f23a40254dbae2944125c570cca2c9e027760a19987124a6" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.764126 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.764199 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts podName:f7b06553-ec05-4c6f-82c0-ef0b4dd3005a nodeName:}" failed. No retries permitted until 2025-12-01 14:59:50.764182185 +0000 UTC m=+1556.527691788 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts") pod "neutronc91e-account-delete-t585q" (UID: "f7b06553-ec05-4c6f-82c0-ef0b4dd3005a") : configmap "openstack-scripts" not found Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.764532 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.764612 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts podName:bf15ccdf-1e4e-44ab-8ccb-819502935a7a nodeName:}" failed. No retries permitted until 2025-12-01 14:59:50.764591876 +0000 UTC m=+1556.528101479 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts") pod "cinder17c6-account-delete-qktk6" (UID: "bf15ccdf-1e4e-44ab-8ccb-819502935a7a") : configmap "openstack-scripts" not found Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.772848 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ffc643854f69e1e2f23a40254dbae2944125c570cca2c9e027760a19987124a6" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.773066 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="444ec27a-77d0-4b01-b59e-c6590c46a0c2" containerName="nova-scheduler-scheduler" Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.773193 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicanff71-account-delete-2fnl8"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.803546 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-ff71-account-create-update-c62lr"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.840401 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-jvlml"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.843688 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-jvlml"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.869635 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder17c6-account-delete-qktk6"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.888750 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-17c6-account-create-update-kt4nm"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.906209 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-17c6-account-create-update-kt4nm"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.928137 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-tmqcq"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.947860 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-tmqcq"] Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.954023 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0a6acaf5c3f0108d3d50cb630c0e2cb9470e93a6443905cbd9058d2ef094517" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.961363 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c91e-account-create-update-wxssk"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.973206 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-c91e-account-create-update-wxssk"] Dec 01 14:59:49 crc kubenswrapper[4810]: E1201 14:59:49.984826 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0a6acaf5c3f0108d3d50cb630c0e2cb9470e93a6443905cbd9058d2ef094517" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 01 14:59:49 crc kubenswrapper[4810]: I1201 14:59:49.990192 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronc91e-account-delete-t585q"] Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.007661 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c0a6acaf5c3f0108d3d50cb630c0e2cb9470e93a6443905cbd9058d2ef094517" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.007716 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c" containerName="nova-cell1-conductor-conductor" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.146432 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance324e-account-delete-6s4dm" event={"ID":"91932480-76aa-4751-9251-3ad82d849882","Type":"ContainerDied","Data":"9e70e548e2a90c425ccea9e2119ed1893ced6c65cb1111e0b6cab49cbcbdf3ea"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.146463 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e70e548e2a90c425ccea9e2119ed1893ced6c65cb1111e0b6cab49cbcbdf3ea" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.148658 4810 generic.go:334] "Generic (PLEG): container finished" podID="129ae854-1430-465f-a177-0c459bd8b4c8" containerID="8c00c76d0791b2c2bc8a97328993ef7e1e8aa6997c54b9aa8cb7c3385b227dd7" exitCode=0 Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.148717 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84864d5bd5-wmq86" event={"ID":"129ae854-1430-465f-a177-0c459bd8b4c8","Type":"ContainerDied","Data":"8c00c76d0791b2c2bc8a97328993ef7e1e8aa6997c54b9aa8cb7c3385b227dd7"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.151456 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell096e4-account-delete-qqlg4" event={"ID":"4f7b5a25-eed0-4122-a036-093ac7adca22","Type":"ContainerStarted","Data":"ba6bbaa4e5865842e4d015f8e09df7c3af89eab522256f5f54fc25471d6b214f"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.153612 4810 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novacell096e4-account-delete-qqlg4" secret="" err="secret \"galera-openstack-dockercfg-5jwrv\" not found" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.155347 4810 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbicanff71-account-delete-2fnl8" secret="" err="secret \"galera-openstack-dockercfg-5jwrv\" not found" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.155767 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanff71-account-delete-2fnl8" event={"ID":"1ef9fa82-f302-47e9-bed3-b02f34a2340f","Type":"ContainerStarted","Data":"dfcba994df93f276679f0b982a3fda4b0ba3b97518b8e6d0c24012c34d33e149"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.162032 4810 generic.go:334] "Generic (PLEG): container finished" podID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerID="3cb0e1b4c05926fcfa43bb4c227b62243d1a9c4234879cdb02f9c40af684935e" exitCode=0 Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.162124 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3","Type":"ContainerDied","Data":"3cb0e1b4c05926fcfa43bb4c227b62243d1a9c4234879cdb02f9c40af684935e"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.164637 4810 generic.go:334] "Generic (PLEG): container finished" podID="992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c" containerID="c0a6acaf5c3f0108d3d50cb630c0e2cb9470e93a6443905cbd9058d2ef094517" exitCode=0 Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.164730 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c","Type":"ContainerDied","Data":"c0a6acaf5c3f0108d3d50cb630c0e2cb9470e93a6443905cbd9058d2ef094517"} Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.183612 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.183705 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts podName:4f7b5a25-eed0-4122-a036-093ac7adca22 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:50.683684842 +0000 UTC m=+1556.447194445 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts") pod "novacell096e4-account-delete-qqlg4" (UID: "4f7b5a25-eed0-4122-a036-093ac7adca22") : configmap "openstack-scripts" not found Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.184293 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.184348 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts podName:1ef9fa82-f302-47e9-bed3-b02f34a2340f nodeName:}" failed. No retries permitted until 2025-12-01 14:59:50.684332449 +0000 UTC m=+1556.447842052 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts") pod "barbicanff71-account-delete-2fnl8" (UID: "1ef9fa82-f302-47e9-bed3-b02f34a2340f") : configmap "openstack-scripts" not found Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.193981 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novacell096e4-account-delete-qqlg4" podStartSLOduration=5.193958031 podStartE2EDuration="5.193958031s" podCreationTimestamp="2025-12-01 14:59:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:59:50.179906298 +0000 UTC m=+1555.943415901" watchObservedRunningTime="2025-12-01 14:59:50.193958031 +0000 UTC m=+1555.957467634" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.231484 4810 generic.go:334] "Generic (PLEG): container finished" podID="9dd705f6-0d8f-49df-aff6-813792212999" containerID="ca8c16cdb56a677709c0ac603398c85e43e5a2f9e1cb3036f753b35de804a1f8" exitCode=0 Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.231783 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9dd705f6-0d8f-49df-aff6-813792212999","Type":"ContainerDied","Data":"ca8c16cdb56a677709c0ac603398c85e43e5a2f9e1cb3036f753b35de804a1f8"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.234649 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-c258c"] Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.240869 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.241120 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone8cd1-account-delete-c69f6" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.243984 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-c258c"] Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.244205 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbicanff71-account-delete-2fnl8" podStartSLOduration=6.244184806 podStartE2EDuration="6.244184806s" podCreationTimestamp="2025-12-01 14:59:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 14:59:50.210244484 +0000 UTC m=+1555.973754077" watchObservedRunningTime="2025-12-01 14:59:50.244184806 +0000 UTC m=+1556.007694409" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.245396 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6c5f549c7d-52fkz" event={"ID":"31b83b20-83c4-4e06-8eb0-1e1d83d4b160","Type":"ContainerDied","Data":"0fc0f37ab3b0b1ca3c876a18bef1e6d8eeb42170c39b40db211606adcf278956"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.249669 4810 scope.go:117] "RemoveContainer" containerID="4d129b7c0c3d768fea14f715dcd8a374cb51cf6191bbfa597636e8ce11c8d85e" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.258821 4810 generic.go:334] "Generic (PLEG): container finished" podID="3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" containerID="08362c9d20e6c5ffc7ef2c58a68a05a6da9d37da0f16242f1435dc93d6db4209" exitCode=0 Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.259023 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6","Type":"ContainerDied","Data":"08362c9d20e6c5ffc7ef2c58a68a05a6da9d37da0f16242f1435dc93d6db4209"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.273821 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-96e4-account-create-update-jgpfh"] Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.279771 4810 generic.go:334] "Generic (PLEG): container finished" podID="e16305af-6573-4cfd-ae76-f596d410920b" containerID="6a66d0aa3094f13ae73e9e79b4f1f5db114907ff336ae2e84fa9c09a936675a3" exitCode=0 Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.280046 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e16305af-6573-4cfd-ae76-f596d410920b","Type":"ContainerDied","Data":"6a66d0aa3094f13ae73e9e79b4f1f5db114907ff336ae2e84fa9c09a936675a3"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.284952 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement8657-account-delete-hx45b" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.286457 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-combined-ca-bundle\") pod \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.287390 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-public-tls-certs\") pod \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.287630 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-logs\") pod \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.287784 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-scripts\") pod \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.287910 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-internal-tls-certs\") pod \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.288022 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8hpz\" (UniqueName: \"kubernetes.io/projected/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-kube-api-access-q8hpz\") pod \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.288121 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-config-data\") pod \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\" (UID: \"31b83b20-83c4-4e06-8eb0-1e1d83d4b160\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.295811 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.296057 4810 generic.go:334] "Generic (PLEG): container finished" podID="c73a61da-0c40-4a26-888c-ab5b994417d4" containerID="6cffd8cec4aa7e3f42f1a3b9cff8d7da52efafaece6f09b30bae42093876b093" exitCode=0 Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.296377 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c73a61da-0c40-4a26-888c-ab5b994417d4","Type":"ContainerDied","Data":"6cffd8cec4aa7e3f42f1a3b9cff8d7da52efafaece6f09b30bae42093876b093"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.298375 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-logs" (OuterVolumeSpecName: "logs") pod "31b83b20-83c4-4e06-8eb0-1e1d83d4b160" (UID: "31b83b20-83c4-4e06-8eb0-1e1d83d4b160"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.309316 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance324e-account-delete-6s4dm" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.311526 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-scripts" (OuterVolumeSpecName: "scripts") pod "31b83b20-83c4-4e06-8eb0-1e1d83d4b160" (UID: "31b83b20-83c4-4e06-8eb0-1e1d83d4b160"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.311820 4810 generic.go:334] "Generic (PLEG): container finished" podID="e79c5698-ab2d-4292-bda9-fab8dbbb2abb" containerID="085bfd508911509acb2f401ff11dc8f197ed30234eac469830bf853fef1a201a" exitCode=0 Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.312017 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e79c5698-ab2d-4292-bda9-fab8dbbb2abb","Type":"ContainerDied","Data":"085bfd508911509acb2f401ff11dc8f197ed30234eac469830bf853fef1a201a"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.319203 4810 scope.go:117] "RemoveContainer" containerID="d8dada242e998791e9f2bfdb36e94b01560f068e320ef4e7eec1b062c6a1f3ff" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.320517 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-kube-api-access-q8hpz" (OuterVolumeSpecName: "kube-api-access-q8hpz") pod "31b83b20-83c4-4e06-8eb0-1e1d83d4b160" (UID: "31b83b20-83c4-4e06-8eb0-1e1d83d4b160"). InnerVolumeSpecName "kube-api-access-q8hpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.328554 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell096e4-account-delete-qqlg4"] Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.336142 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-96e4-account-create-update-jgpfh"] Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.348278 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.352138 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.359124 4810 generic.go:334] "Generic (PLEG): container finished" podID="bdd90cbb-4242-4227-9de4-09f99e60a516" containerID="4a8d673779dd5bbbb9060129ba8fe9ee012663d8194ce9496b1684874e3903dc" exitCode=0 Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.359230 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" event={"ID":"bdd90cbb-4242-4227-9de4-09f99e60a516","Type":"ContainerDied","Data":"4a8d673779dd5bbbb9060129ba8fe9ee012663d8194ce9496b1684874e3903dc"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.359256 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" event={"ID":"bdd90cbb-4242-4227-9de4-09f99e60a516","Type":"ContainerDied","Data":"ea546ae2e75fe338dbf8885d66eb7865f2845a29a179ed939efd251c8116fbf0"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.363185 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5b4ff7d6cb-m2g6c" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.369703 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.376116 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.381224 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.387617 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-config-data" (OuterVolumeSpecName: "config-data") pod "31b83b20-83c4-4e06-8eb0-1e1d83d4b160" (UID: "31b83b20-83c4-4e06-8eb0-1e1d83d4b160"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.394511 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdd90cbb-4242-4227-9de4-09f99e60a516-logs\") pod \"bdd90cbb-4242-4227-9de4-09f99e60a516\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.394580 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9sct\" (UniqueName: \"kubernetes.io/projected/39ce8411-05ee-4022-bd7a-87e46c2cee8d-kube-api-access-h9sct\") pod \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\" (UID: \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.394611 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-config-data\") pod \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.394647 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-combined-ca-bundle\") pod \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.394670 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a3a8b5d-01e4-4396-8a9b-abd85dc55572-operator-scripts\") pod \"2a3a8b5d-01e4-4396-8a9b-abd85dc55572\" (UID: \"2a3a8b5d-01e4-4396-8a9b-abd85dc55572\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.394713 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-combined-ca-bundle\") pod \"bdd90cbb-4242-4227-9de4-09f99e60a516\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.394740 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kx7mz\" (UniqueName: \"kubernetes.io/projected/9dd705f6-0d8f-49df-aff6-813792212999-kube-api-access-kx7mz\") pod \"9dd705f6-0d8f-49df-aff6-813792212999\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.394760 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-internal-tls-certs\") pod \"bdd90cbb-4242-4227-9de4-09f99e60a516\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.394789 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzlvt\" (UniqueName: \"kubernetes.io/projected/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-kube-api-access-hzlvt\") pod \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.394812 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/39ce8411-05ee-4022-bd7a-87e46c2cee8d-kube-state-metrics-tls-config\") pod \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\" (UID: \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.394819 4810 generic.go:334] "Generic (PLEG): container finished" podID="39ce8411-05ee-4022-bd7a-87e46c2cee8d" containerID="030a775eb17300c273b82602d862651690419af2f0861ccfe4b4d6bdc64e3a5e" exitCode=2 Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.394861 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-config-data-custom\") pod \"bdd90cbb-4242-4227-9de4-09f99e60a516\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.394898 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-public-tls-certs\") pod \"bdd90cbb-4242-4227-9de4-09f99e60a516\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.394920 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-config-data\") pod \"bdd90cbb-4242-4227-9de4-09f99e60a516\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.394942 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39ce8411-05ee-4022-bd7a-87e46c2cee8d-combined-ca-bundle\") pod \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\" (UID: \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.395018 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.396412 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd8zj\" (UniqueName: \"kubernetes.io/projected/91932480-76aa-4751-9251-3ad82d849882-kube-api-access-jd8zj\") pod \"91932480-76aa-4751-9251-3ad82d849882\" (UID: \"91932480-76aa-4751-9251-3ad82d849882\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.396444 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-nova-metadata-tls-certs\") pod \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.396551 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-logs\") pod \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\" (UID: \"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.397798 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdd90cbb-4242-4227-9de4-09f99e60a516-logs" (OuterVolumeSpecName: "logs") pod "bdd90cbb-4242-4227-9de4-09f99e60a516" (UID: "bdd90cbb-4242-4227-9de4-09f99e60a516"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.403834 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39ce8411-05ee-4022-bd7a-87e46c2cee8d-kube-api-access-h9sct" (OuterVolumeSpecName: "kube-api-access-h9sct") pod "39ce8411-05ee-4022-bd7a-87e46c2cee8d" (UID: "39ce8411-05ee-4022-bd7a-87e46c2cee8d"). InnerVolumeSpecName "kube-api-access-h9sct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.405973 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a3a8b5d-01e4-4396-8a9b-abd85dc55572-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2a3a8b5d-01e4-4396-8a9b-abd85dc55572" (UID: "2a3a8b5d-01e4-4396-8a9b-abd85dc55572"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.406865 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dd705f6-0d8f-49df-aff6-813792212999-kube-api-access-kx7mz" (OuterVolumeSpecName: "kube-api-access-kx7mz") pod "9dd705f6-0d8f-49df-aff6-813792212999" (UID: "9dd705f6-0d8f-49df-aff6-813792212999"). InnerVolumeSpecName "kube-api-access-kx7mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.415231 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement8657-account-delete-hx45b" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.394917 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"39ce8411-05ee-4022-bd7a-87e46c2cee8d","Type":"ContainerDied","Data":"030a775eb17300c273b82602d862651690419af2f0861ccfe4b4d6bdc64e3a5e"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.415619 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.415644 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"39ce8411-05ee-4022-bd7a-87e46c2cee8d","Type":"ContainerDied","Data":"5b4a8d4df900f9b4fede27035950a845b76dd2849a241b354a524387ce483969"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.415668 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement8657-account-delete-hx45b" event={"ID":"2a3a8b5d-01e4-4396-8a9b-abd85dc55572","Type":"ContainerDied","Data":"44593a6355c67c2f6ce5902d677ecce531982fb97d53a6bc836d3bdf9d804c4b"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.415684 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44593a6355c67c2f6ce5902d677ecce531982fb97d53a6bc836d3bdf9d804c4b" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.423128 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-logs" (OuterVolumeSpecName: "logs") pod "3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" (UID: "3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.396580 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/39ce8411-05ee-4022-bd7a-87e46c2cee8d-kube-state-metrics-tls-certs\") pod \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\" (UID: \"39ce8411-05ee-4022-bd7a-87e46c2cee8d\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.435317 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mrbz\" (UniqueName: \"kubernetes.io/projected/bdd90cbb-4242-4227-9de4-09f99e60a516-kube-api-access-9mrbz\") pod \"bdd90cbb-4242-4227-9de4-09f99e60a516\" (UID: \"bdd90cbb-4242-4227-9de4-09f99e60a516\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.435377 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gww9m\" (UniqueName: \"kubernetes.io/projected/2a3a8b5d-01e4-4396-8a9b-abd85dc55572-kube-api-access-gww9m\") pod \"2a3a8b5d-01e4-4396-8a9b-abd85dc55572\" (UID: \"2a3a8b5d-01e4-4396-8a9b-abd85dc55572\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.435400 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-public-tls-certs\") pod \"9dd705f6-0d8f-49df-aff6-813792212999\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.435434 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91932480-76aa-4751-9251-3ad82d849882-operator-scripts\") pod \"91932480-76aa-4751-9251-3ad82d849882\" (UID: \"91932480-76aa-4751-9251-3ad82d849882\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.435788 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91932480-76aa-4751-9251-3ad82d849882-kube-api-access-jd8zj" (OuterVolumeSpecName: "kube-api-access-jd8zj") pod "91932480-76aa-4751-9251-3ad82d849882" (UID: "91932480-76aa-4751-9251-3ad82d849882"). InnerVolumeSpecName "kube-api-access-jd8zj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.436451 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdd90cbb-4242-4227-9de4-09f99e60a516-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.469083 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9sct\" (UniqueName: \"kubernetes.io/projected/39ce8411-05ee-4022-bd7a-87e46c2cee8d-kube-api-access-h9sct\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.469109 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a3a8b5d-01e4-4396-8a9b-abd85dc55572-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.469120 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.469130 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kx7mz\" (UniqueName: \"kubernetes.io/projected/9dd705f6-0d8f-49df-aff6-813792212999-kube-api-access-kx7mz\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.469151 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.469161 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.469171 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8hpz\" (UniqueName: \"kubernetes.io/projected/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-kube-api-access-q8hpz\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.469180 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd8zj\" (UniqueName: \"kubernetes.io/projected/91932480-76aa-4751-9251-3ad82d849882-kube-api-access-jd8zj\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.469192 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.453027 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.469237 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapic0fd-account-delete-w7ttc" event={"ID":"476b71ac-1050-46a6-b51c-168e75b1a350","Type":"ContainerStarted","Data":"691a88e3e213a94e7564d8b43ac6e1b2e48afa2314fc5a29d364c71768f366fb"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.447038 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91932480-76aa-4751-9251-3ad82d849882-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "91932480-76aa-4751-9251-3ad82d849882" (UID: "91932480-76aa-4751-9251-3ad82d849882"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.467898 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bdd90cbb-4242-4227-9de4-09f99e60a516" (UID: "bdd90cbb-4242-4227-9de4-09f99e60a516"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.470803 4810 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novaapic0fd-account-delete-w7ttc" secret="" err="secret \"galera-openstack-dockercfg-5jwrv\" not found" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.476646 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a3a8b5d-01e4-4396-8a9b-abd85dc55572-kube-api-access-gww9m" (OuterVolumeSpecName: "kube-api-access-gww9m") pod "2a3a8b5d-01e4-4396-8a9b-abd85dc55572" (UID: "2a3a8b5d-01e4-4396-8a9b-abd85dc55572"). InnerVolumeSpecName "kube-api-access-gww9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.492394 4810 scope.go:117] "RemoveContainer" containerID="1cc6810ebfaa94211940478b494df0cc163de04d06f714491ab4d71d2b9f032d" Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.493742 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6cffd8cec4aa7e3f42f1a3b9cff8d7da52efafaece6f09b30bae42093876b093 is running failed: container process not found" containerID="6cffd8cec4aa7e3f42f1a3b9cff8d7da52efafaece6f09b30bae42093876b093" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.494132 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6cffd8cec4aa7e3f42f1a3b9cff8d7da52efafaece6f09b30bae42093876b093 is running failed: container process not found" containerID="6cffd8cec4aa7e3f42f1a3b9cff8d7da52efafaece6f09b30bae42093876b093" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.494412 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.494642 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6cffd8cec4aa7e3f42f1a3b9cff8d7da52efafaece6f09b30bae42093876b093 is running failed: container process not found" containerID="6cffd8cec4aa7e3f42f1a3b9cff8d7da52efafaece6f09b30bae42093876b093" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.494690 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6cffd8cec4aa7e3f42f1a3b9cff8d7da52efafaece6f09b30bae42093876b093 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="c73a61da-0c40-4a26-888c-ab5b994417d4" containerName="nova-cell0-conductor-conductor" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.497149 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.498814 4810 generic.go:334] "Generic (PLEG): container finished" podID="d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" containerID="aba75285c97143bda5c120fd54711afa0a88b59eeac5b761fa88578cc1f0f3b2" exitCode=0 Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.502857 4810 generic.go:334] "Generic (PLEG): container finished" podID="9acd8c29-93f5-4b0f-8b32-d8f3d576767b" containerID="1f6241fff3fa641130500b0c5b06297d792485793ad36fdc4dd06993a79f221d" exitCode=0 Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.504322 4810 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/cinder17c6-account-delete-qktk6" secret="" err="secret \"galera-openstack-dockercfg-5jwrv\" not found" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.504352 4810 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/neutronc91e-account-delete-t585q" secret="" err="secret \"galera-openstack-dockercfg-5jwrv\" not found" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.515563 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-kube-api-access-hzlvt" (OuterVolumeSpecName: "kube-api-access-hzlvt") pod "3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" (UID: "3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6"). InnerVolumeSpecName "kube-api-access-hzlvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.515656 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdd90cbb-4242-4227-9de4-09f99e60a516-kube-api-access-9mrbz" (OuterVolumeSpecName: "kube-api-access-9mrbz") pod "bdd90cbb-4242-4227-9de4-09f99e60a516" (UID: "bdd90cbb-4242-4227-9de4-09f99e60a516"). InnerVolumeSpecName "kube-api-access-9mrbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.543591 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01e51b04-eb2c-4102-9a86-d19ca8c566d6" path="/var/lib/kubelet/pods/01e51b04-eb2c-4102-9a86-d19ca8c566d6/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.544003 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="068d65be-c5cf-4198-a7ce-0aedeb5e9985" path="/var/lib/kubelet/pods/068d65be-c5cf-4198-a7ce-0aedeb5e9985/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.544402 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="081ac19d-af0e-48f7-9c2e-0d282a9bb7ba" path="/var/lib/kubelet/pods/081ac19d-af0e-48f7-9c2e-0d282a9bb7ba/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.544986 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0899579c-adaa-41fb-a36d-31ba657c6c1a" path="/var/lib/kubelet/pods/0899579c-adaa-41fb-a36d-31ba657c6c1a/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.546902 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e760d14-b5f0-48c8-927e-09d086591dce" path="/var/lib/kubelet/pods/1e760d14-b5f0-48c8-927e-09d086591dce/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.547571 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21d83a41-8ec9-424e-be10-fac93f693310" path="/var/lib/kubelet/pods/21d83a41-8ec9-424e-be10-fac93f693310/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.548201 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30a1b114-0b4e-43ff-815b-0983ca7a86e6" path="/var/lib/kubelet/pods/30a1b114-0b4e-43ff-815b-0983ca7a86e6/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.550010 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3798bfc4-fb37-48d1-8146-42d177cf7861" path="/var/lib/kubelet/pods/3798bfc4-fb37-48d1-8146-42d177cf7861/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.552292 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55732ab9-9610-4b4f-bae3-516b623e0608" path="/var/lib/kubelet/pods/55732ab9-9610-4b4f-bae3-516b623e0608/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.552918 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b341ba0-f0a3-40f7-a99b-7eb045ba4739" path="/var/lib/kubelet/pods/5b341ba0-f0a3-40f7-a99b-7eb045ba4739/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.555802 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72889b41-41e2-4d58-8e6d-f372f75793b7" path="/var/lib/kubelet/pods/72889b41-41e2-4d58-8e6d-f372f75793b7/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.556882 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7478f42b-4be2-430b-9854-5930dccaee43" path="/var/lib/kubelet/pods/7478f42b-4be2-430b-9854-5930dccaee43/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.557452 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88635a0d-cb9d-43c1-9086-5623c9a780f8" path="/var/lib/kubelet/pods/88635a0d-cb9d-43c1-9086-5623c9a780f8/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.558654 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fb6a553-c5fd-427f-96e6-b5bd6d3a4971" path="/var/lib/kubelet/pods/8fb6a553-c5fd-427f-96e6-b5bd6d3a4971/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.559168 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b88404c2-d4bf-4257-986a-1b78178fca88" path="/var/lib/kubelet/pods/b88404c2-d4bf-4257-986a-1b78178fca88/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.559728 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b96848ac-9038-4172-a36b-fe0889e3f6b6" path="/var/lib/kubelet/pods/b96848ac-9038-4172-a36b-fe0889e3f6b6/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.560676 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcd634cb-7e87-4e64-9eb7-894a15a047e2" path="/var/lib/kubelet/pods/bcd634cb-7e87-4e64-9eb7-894a15a047e2/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.561205 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d17c667c-7985-4132-88ee-f0e82ab7d25c" path="/var/lib/kubelet/pods/d17c667c-7985-4132-88ee-f0e82ab7d25c/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.561897 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d202704a-7a2e-46a6-9e20-2eeabdc11814" path="/var/lib/kubelet/pods/d202704a-7a2e-46a6-9e20-2eeabdc11814/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.562690 4810 scope.go:117] "RemoveContainer" containerID="65c24e7fbab051767fd3a4645fd6dfed9e0a34a7bc48306bde5ba5dbb76b4aad" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.562881 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3b0f7e7-9035-4c33-940b-8bdd905d97b5" path="/var/lib/kubelet/pods/d3b0f7e7-9035-4c33-940b-8bdd905d97b5/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.563365 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6a42d09-fb12-4466-a792-49998f29ccc4" path="/var/lib/kubelet/pods/d6a42d09-fb12-4466-a792-49998f29ccc4/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.563933 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e29ef9a3-9060-4301-ad2c-3d62fe66c6b9" path="/var/lib/kubelet/pods/e29ef9a3-9060-4301-ad2c-3d62fe66c6b9/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.564421 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0221e2e-817a-4fa9-9717-2bb622c966f0" path="/var/lib/kubelet/pods/f0221e2e-817a-4fa9-9717-2bb622c966f0/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.565378 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc8eec67-10ea-427f-b6f7-045c1dc7519f" path="/var/lib/kubelet/pods/fc8eec67-10ea-427f-b6f7-045c1dc7519f/volumes" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.570455 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"9dd705f6-0d8f-49df-aff6-813792212999\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.573545 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9dd705f6-0d8f-49df-aff6-813792212999-httpd-run\") pod \"9dd705f6-0d8f-49df-aff6-813792212999\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.573606 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-combined-ca-bundle\") pod \"9dd705f6-0d8f-49df-aff6-813792212999\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.573704 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-config-data\") pod \"9dd705f6-0d8f-49df-aff6-813792212999\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.574129 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dd705f6-0d8f-49df-aff6-813792212999-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9dd705f6-0d8f-49df-aff6-813792212999" (UID: "9dd705f6-0d8f-49df-aff6-813792212999"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.577969 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-scripts\") pod \"9dd705f6-0d8f-49df-aff6-813792212999\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.578036 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd705f6-0d8f-49df-aff6-813792212999-logs\") pod \"9dd705f6-0d8f-49df-aff6-813792212999\" (UID: \"9dd705f6-0d8f-49df-aff6-813792212999\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.588052 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dd705f6-0d8f-49df-aff6-813792212999-logs" (OuterVolumeSpecName: "logs") pod "9dd705f6-0d8f-49df-aff6-813792212999" (UID: "9dd705f6-0d8f-49df-aff6-813792212999"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.591941 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "9dd705f6-0d8f-49df-aff6-813792212999" (UID: "9dd705f6-0d8f-49df-aff6-813792212999"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.592638 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.592661 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9dd705f6-0d8f-49df-aff6-813792212999-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.592672 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mrbz\" (UniqueName: \"kubernetes.io/projected/bdd90cbb-4242-4227-9de4-09f99e60a516-kube-api-access-9mrbz\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.592683 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gww9m\" (UniqueName: \"kubernetes.io/projected/2a3a8b5d-01e4-4396-8a9b-abd85dc55572-kube-api-access-gww9m\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.592693 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91932480-76aa-4751-9251-3ad82d849882-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.592702 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd705f6-0d8f-49df-aff6-813792212999-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.592721 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.592749 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzlvt\" (UniqueName: \"kubernetes.io/projected/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-kube-api-access-hzlvt\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.596803 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.601042 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc","Type":"ContainerDied","Data":"aba75285c97143bda5c120fd54711afa0a88b59eeac5b761fa88578cc1f0f3b2"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.601074 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9acd8c29-93f5-4b0f-8b32-d8f3d576767b","Type":"ContainerDied","Data":"1f6241fff3fa641130500b0c5b06297d792485793ad36fdc4dd06993a79f221d"} Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.601093 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-2fflr"] Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.601107 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-2fflr"] Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.601125 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-c0fd-account-create-update-w9l6n"] Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.601135 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapic0fd-account-delete-w7ttc"] Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.607553 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-c0fd-account-create-update-w9l6n"] Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.622343 4810 scope.go:117] "RemoveContainer" containerID="1cc6810ebfaa94211940478b494df0cc163de04d06f714491ab4d71d2b9f032d" Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.623827 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cc6810ebfaa94211940478b494df0cc163de04d06f714491ab4d71d2b9f032d\": container with ID starting with 1cc6810ebfaa94211940478b494df0cc163de04d06f714491ab4d71d2b9f032d not found: ID does not exist" containerID="1cc6810ebfaa94211940478b494df0cc163de04d06f714491ab4d71d2b9f032d" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.623871 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cc6810ebfaa94211940478b494df0cc163de04d06f714491ab4d71d2b9f032d"} err="failed to get container status \"1cc6810ebfaa94211940478b494df0cc163de04d06f714491ab4d71d2b9f032d\": rpc error: code = NotFound desc = could not find container \"1cc6810ebfaa94211940478b494df0cc163de04d06f714491ab4d71d2b9f032d\": container with ID starting with 1cc6810ebfaa94211940478b494df0cc163de04d06f714491ab4d71d2b9f032d not found: ID does not exist" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.623899 4810 scope.go:117] "RemoveContainer" containerID="65c24e7fbab051767fd3a4645fd6dfed9e0a34a7bc48306bde5ba5dbb76b4aad" Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.627271 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65c24e7fbab051767fd3a4645fd6dfed9e0a34a7bc48306bde5ba5dbb76b4aad\": container with ID starting with 65c24e7fbab051767fd3a4645fd6dfed9e0a34a7bc48306bde5ba5dbb76b4aad not found: ID does not exist" containerID="65c24e7fbab051767fd3a4645fd6dfed9e0a34a7bc48306bde5ba5dbb76b4aad" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.627320 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65c24e7fbab051767fd3a4645fd6dfed9e0a34a7bc48306bde5ba5dbb76b4aad"} err="failed to get container status \"65c24e7fbab051767fd3a4645fd6dfed9e0a34a7bc48306bde5ba5dbb76b4aad\": rpc error: code = NotFound desc = could not find container \"65c24e7fbab051767fd3a4645fd6dfed9e0a34a7bc48306bde5ba5dbb76b4aad\": container with ID starting with 65c24e7fbab051767fd3a4645fd6dfed9e0a34a7bc48306bde5ba5dbb76b4aad not found: ID does not exist" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.627347 4810 scope.go:117] "RemoveContainer" containerID="1c3b60c932bcc8ce61ca6d0f28d631b382aa2b237627432ebf90c92ecb1485d0" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.641164 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-scripts" (OuterVolumeSpecName: "scripts") pod "9dd705f6-0d8f-49df-aff6-813792212999" (UID: "9dd705f6-0d8f-49df-aff6-813792212999"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.651499 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31b83b20-83c4-4e06-8eb0-1e1d83d4b160" (UID: "31b83b20-83c4-4e06-8eb0-1e1d83d4b160"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.688616 4810 scope.go:117] "RemoveContainer" containerID="c81abf769f22e0110d5df574ceaf9e47a00a5723bf172e5cebaebd7bde8ecee2" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.694462 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/129ae854-1430-465f-a177-0c459bd8b4c8-combined-ca-bundle\") pod \"129ae854-1430-465f-a177-0c459bd8b4c8\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.694651 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-combined-ca-bundle\") pod \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.694768 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-scripts\") pod \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.695679 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/129ae854-1430-465f-a177-0c459bd8b4c8-config-data\") pod \"129ae854-1430-465f-a177-0c459bd8b4c8\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.695838 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/129ae854-1430-465f-a177-0c459bd8b4c8-logs\") pod \"129ae854-1430-465f-a177-0c459bd8b4c8\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.695963 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/129ae854-1430-465f-a177-0c459bd8b4c8-config-data-custom\") pod \"129ae854-1430-465f-a177-0c459bd8b4c8\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.696099 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e16305af-6573-4cfd-ae76-f596d410920b-combined-ca-bundle\") pod \"e16305af-6573-4cfd-ae76-f596d410920b\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.696952 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tls46\" (UniqueName: \"kubernetes.io/projected/129ae854-1430-465f-a177-0c459bd8b4c8-kube-api-access-tls46\") pod \"129ae854-1430-465f-a177-0c459bd8b4c8\" (UID: \"129ae854-1430-465f-a177-0c459bd8b4c8\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.697041 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-run-httpd\") pod \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.697183 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-config-data\") pod \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.697291 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e16305af-6573-4cfd-ae76-f596d410920b-config-data\") pod \"e16305af-6573-4cfd-ae76-f596d410920b\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.697378 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-sg-core-conf-yaml\") pod \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.697639 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e16305af-6573-4cfd-ae76-f596d410920b-kolla-config\") pod \"e16305af-6573-4cfd-ae76-f596d410920b\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.697829 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-log-httpd\") pod \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.697949 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e16305af-6573-4cfd-ae76-f596d410920b-memcached-tls-certs\") pod \"e16305af-6573-4cfd-ae76-f596d410920b\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.698052 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-ceilometer-tls-certs\") pod \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.698140 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrlvv\" (UniqueName: \"kubernetes.io/projected/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-kube-api-access-zrlvv\") pod \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\" (UID: \"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.698221 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtzq4\" (UniqueName: \"kubernetes.io/projected/e16305af-6573-4cfd-ae76-f596d410920b-kube-api-access-wtzq4\") pod \"e16305af-6573-4cfd-ae76-f596d410920b\" (UID: \"e16305af-6573-4cfd-ae76-f596d410920b\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.698808 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.698940 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.695171 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.695507 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9dd705f6-0d8f-49df-aff6-813792212999" (UID: "9dd705f6-0d8f-49df-aff6-813792212999"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.700146 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/129ae854-1430-465f-a177-0c459bd8b4c8-logs" (OuterVolumeSpecName: "logs") pod "129ae854-1430-465f-a177-0c459bd8b4c8" (UID: "129ae854-1430-465f-a177-0c459bd8b4c8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.700324 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" (UID: "c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.702961 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.703013 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts podName:1ef9fa82-f302-47e9-bed3-b02f34a2340f nodeName:}" failed. No retries permitted until 2025-12-01 14:59:51.702994792 +0000 UTC m=+1557.466504395 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts") pod "barbicanff71-account-delete-2fnl8" (UID: "1ef9fa82-f302-47e9-bed3-b02f34a2340f") : configmap "openstack-scripts" not found Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.705345 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.705450 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts podName:4f7b5a25-eed0-4122-a036-093ac7adca22 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:51.705426627 +0000 UTC m=+1557.468936260 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts") pod "novacell096e4-account-delete-qqlg4" (UID: "4f7b5a25-eed0-4122-a036-093ac7adca22") : configmap "openstack-scripts" not found Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.705484 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.705568 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts podName:476b71ac-1050-46a6-b51c-168e75b1a350 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:52.70554877 +0000 UTC m=+1558.469058373 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts") pod "novaapic0fd-account-delete-w7ttc" (UID: "476b71ac-1050-46a6-b51c-168e75b1a350") : configmap "openstack-scripts" not found Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.706034 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e16305af-6573-4cfd-ae76-f596d410920b-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "e16305af-6573-4cfd-ae76-f596d410920b" (UID: "e16305af-6573-4cfd-ae76-f596d410920b"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.706457 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e16305af-6573-4cfd-ae76-f596d410920b-config-data" (OuterVolumeSpecName: "config-data") pod "e16305af-6573-4cfd-ae76-f596d410920b" (UID: "e16305af-6573-4cfd-ae76-f596d410920b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.707382 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" (UID: "c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.714014 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-scripts" (OuterVolumeSpecName: "scripts") pod "c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" (UID: "c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.718242 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/129ae854-1430-465f-a177-0c459bd8b4c8-kube-api-access-tls46" (OuterVolumeSpecName: "kube-api-access-tls46") pod "129ae854-1430-465f-a177-0c459bd8b4c8" (UID: "129ae854-1430-465f-a177-0c459bd8b4c8"). InnerVolumeSpecName "kube-api-access-tls46". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.718342 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/129ae854-1430-465f-a177-0c459bd8b4c8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "129ae854-1430-465f-a177-0c459bd8b4c8" (UID: "129ae854-1430-465f-a177-0c459bd8b4c8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.724588 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-kube-api-access-zrlvv" (OuterVolumeSpecName: "kube-api-access-zrlvv") pod "c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" (UID: "c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3"). InnerVolumeSpecName "kube-api-access-zrlvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.730232 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e16305af-6573-4cfd-ae76-f596d410920b-kube-api-access-wtzq4" (OuterVolumeSpecName: "kube-api-access-wtzq4") pod "e16305af-6573-4cfd-ae76-f596d410920b" (UID: "e16305af-6573-4cfd-ae76-f596d410920b"). InnerVolumeSpecName "kube-api-access-wtzq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.744713 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-config-data" (OuterVolumeSpecName: "config-data") pod "3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" (UID: "3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.775067 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.801443 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlz9r\" (UniqueName: \"kubernetes.io/projected/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-kube-api-access-nlz9r\") pod \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.801494 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-internal-tls-certs\") pod \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.801547 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-config-data-custom\") pod \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.801656 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-config-data\") pod \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.801677 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-logs\") pod \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.801746 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-scripts\") pod \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.801789 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-etc-machine-id\") pod \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.801810 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-public-tls-certs\") pod \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.801879 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-combined-ca-bundle\") pod \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\" (UID: \"e79c5698-ab2d-4292-bda9-fab8dbbb2abb\") " Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.802232 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39ce8411-05ee-4022-bd7a-87e46c2cee8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39ce8411-05ee-4022-bd7a-87e46c2cee8d" (UID: "39ce8411-05ee-4022-bd7a-87e46c2cee8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.802430 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.802723 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtzq4\" (UniqueName: \"kubernetes.io/projected/e16305af-6573-4cfd-ae76-f596d410920b-kube-api-access-wtzq4\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.802796 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.802867 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/129ae854-1430-465f-a177-0c459bd8b4c8-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.802925 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/129ae854-1430-465f-a177-0c459bd8b4c8-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.802991 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.803055 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.803426 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tls46\" (UniqueName: \"kubernetes.io/projected/129ae854-1430-465f-a177-0c459bd8b4c8-kube-api-access-tls46\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.803539 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts podName:f7b06553-ec05-4c6f-82c0-ef0b4dd3005a nodeName:}" failed. No retries permitted until 2025-12-01 14:59:52.803511804 +0000 UTC m=+1558.567021417 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts") pod "neutronc91e-account-delete-t585q" (UID: "f7b06553-ec05-4c6f-82c0-ef0b4dd3005a") : configmap "openstack-scripts" not found Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.803571 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.803586 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e16305af-6573-4cfd-ae76-f596d410920b-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.803599 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39ce8411-05ee-4022-bd7a-87e46c2cee8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.803614 4810 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e16305af-6573-4cfd-ae76-f596d410920b-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.803629 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.803641 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.803652 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrlvv\" (UniqueName: \"kubernetes.io/projected/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-kube-api-access-zrlvv\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.803736 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:50 crc kubenswrapper[4810]: E1201 14:59:50.803784 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts podName:bf15ccdf-1e4e-44ab-8ccb-819502935a7a nodeName:}" failed. No retries permitted until 2025-12-01 14:59:52.803773052 +0000 UTC m=+1558.567282665 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts") pod "cinder17c6-account-delete-qktk6" (UID: "bf15ccdf-1e4e-44ab-8ccb-819502935a7a") : configmap "openstack-scripts" not found Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.803786 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-logs" (OuterVolumeSpecName: "logs") pod "e79c5698-ab2d-4292-bda9-fab8dbbb2abb" (UID: "e79c5698-ab2d-4292-bda9-fab8dbbb2abb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.803825 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e79c5698-ab2d-4292-bda9-fab8dbbb2abb" (UID: "e79c5698-ab2d-4292-bda9-fab8dbbb2abb"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.806649 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-scripts" (OuterVolumeSpecName: "scripts") pod "e79c5698-ab2d-4292-bda9-fab8dbbb2abb" (UID: "e79c5698-ab2d-4292-bda9-fab8dbbb2abb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.846602 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-kube-api-access-nlz9r" (OuterVolumeSpecName: "kube-api-access-nlz9r") pod "e79c5698-ab2d-4292-bda9-fab8dbbb2abb" (UID: "e79c5698-ab2d-4292-bda9-fab8dbbb2abb"). InnerVolumeSpecName "kube-api-access-nlz9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.855095 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e79c5698-ab2d-4292-bda9-fab8dbbb2abb" (UID: "e79c5698-ab2d-4292-bda9-fab8dbbb2abb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.862336 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" (UID: "3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.905564 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.905596 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.905617 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.905633 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.905641 4810 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.905649 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlz9r\" (UniqueName: \"kubernetes.io/projected/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-kube-api-access-nlz9r\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.910772 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39ce8411-05ee-4022-bd7a-87e46c2cee8d-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "39ce8411-05ee-4022-bd7a-87e46c2cee8d" (UID: "39ce8411-05ee-4022-bd7a-87e46c2cee8d"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.950579 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e16305af-6573-4cfd-ae76-f596d410920b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e16305af-6573-4cfd-ae76-f596d410920b" (UID: "e16305af-6573-4cfd-ae76-f596d410920b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:50 crc kubenswrapper[4810]: I1201 14:59:50.980071 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9dd705f6-0d8f-49df-aff6-813792212999" (UID: "9dd705f6-0d8f-49df-aff6-813792212999"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.011411 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.012408 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e16305af-6573-4cfd-ae76-f596d410920b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.012565 4810 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/39ce8411-05ee-4022-bd7a-87e46c2cee8d-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.043653 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bdd90cbb-4242-4227-9de4-09f99e60a516" (UID: "bdd90cbb-4242-4227-9de4-09f99e60a516"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.054738 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-config-data" (OuterVolumeSpecName: "config-data") pod "bdd90cbb-4242-4227-9de4-09f99e60a516" (UID: "bdd90cbb-4242-4227-9de4-09f99e60a516"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.055175 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39ce8411-05ee-4022-bd7a-87e46c2cee8d-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "39ce8411-05ee-4022-bd7a-87e46c2cee8d" (UID: "39ce8411-05ee-4022-bd7a-87e46c2cee8d"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.075466 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "bdd90cbb-4242-4227-9de4-09f99e60a516" (UID: "bdd90cbb-4242-4227-9de4-09f99e60a516"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.076444 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e79c5698-ab2d-4292-bda9-fab8dbbb2abb" (UID: "e79c5698-ab2d-4292-bda9-fab8dbbb2abb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.082402 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" (UID: "3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.098562 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-config-data" (OuterVolumeSpecName: "config-data") pod "e79c5698-ab2d-4292-bda9-fab8dbbb2abb" (UID: "e79c5698-ab2d-4292-bda9-fab8dbbb2abb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.103664 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" (UID: "c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.105220 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e16305af-6573-4cfd-ae76-f596d410920b-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "e16305af-6573-4cfd-ae76-f596d410920b" (UID: "e16305af-6573-4cfd-ae76-f596d410920b"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.123447 4810 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/39ce8411-05ee-4022-bd7a-87e46c2cee8d-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.123491 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.123503 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.123511 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.123520 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.123527 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.123536 4810 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e16305af-6573-4cfd-ae76-f596d410920b-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.123544 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.123553 4810 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.124601 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "31b83b20-83c4-4e06-8eb0-1e1d83d4b160" (UID: "31b83b20-83c4-4e06-8eb0-1e1d83d4b160"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.124798 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/129ae854-1430-465f-a177-0c459bd8b4c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "129ae854-1430-465f-a177-0c459bd8b4c8" (UID: "129ae854-1430-465f-a177-0c459bd8b4c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.149837 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "bdd90cbb-4242-4227-9de4-09f99e60a516" (UID: "bdd90cbb-4242-4227-9de4-09f99e60a516"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.153920 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/129ae854-1430-465f-a177-0c459bd8b4c8-config-data" (OuterVolumeSpecName: "config-data") pod "129ae854-1430-465f-a177-0c459bd8b4c8" (UID: "129ae854-1430-465f-a177-0c459bd8b4c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.156892 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-config-data" (OuterVolumeSpecName: "config-data") pod "9dd705f6-0d8f-49df-aff6-813792212999" (UID: "9dd705f6-0d8f-49df-aff6-813792212999"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.161582 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" (UID: "c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.165646 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "31b83b20-83c4-4e06-8eb0-1e1d83d4b160" (UID: "31b83b20-83c4-4e06-8eb0-1e1d83d4b160"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.178539 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" (UID: "c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.183039 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e79c5698-ab2d-4292-bda9-fab8dbbb2abb" (UID: "e79c5698-ab2d-4292-bda9-fab8dbbb2abb"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.202045 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e79c5698-ab2d-4292-bda9-fab8dbbb2abb" (UID: "e79c5698-ab2d-4292-bda9-fab8dbbb2abb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.216939 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-config-data" (OuterVolumeSpecName: "config-data") pod "c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" (UID: "c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.224714 4810 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.224911 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/129ae854-1430-465f-a177-0c459bd8b4c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.225006 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.225132 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd705f6-0d8f-49df-aff6-813792212999-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.225224 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/129ae854-1430-465f-a177-0c459bd8b4c8-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.225309 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.225386 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.225461 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.225561 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31b83b20-83c4-4e06-8eb0-1e1d83d4b160-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.225619 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e79c5698-ab2d-4292-bda9-fab8dbbb2abb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.225674 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd90cbb-4242-4227-9de4-09f99e60a516-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.229013 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.232755 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.238541 4810 scope.go:117] "RemoveContainer" containerID="ba0d20517bfab8cdb0b867e26e4238d3696e54b13598a77e2b77184f47eb47d0" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.238871 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement8657-account-delete-hx45b"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.239889 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.257380 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement8657-account-delete-hx45b"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.262730 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.263315 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.271695 4810 scope.go:117] "RemoveContainer" containerID="4a8d673779dd5bbbb9060129ba8fe9ee012663d8194ce9496b1684874e3903dc" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.308877 4810 scope.go:117] "RemoveContainer" containerID="0af6bd36655b389f1090223a090ab3a8d47cace5e570943004f63de4c038edb3" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.327156 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-httpd-run\") pod \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.327461 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c-config-data\") pod \"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c\" (UID: \"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.327812 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" (UID: "d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.327929 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-config-data\") pod \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.328010 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-666rf\" (UniqueName: \"kubernetes.io/projected/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-kube-api-access-666rf\") pod \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.328083 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c73a61da-0c40-4a26-888c-ab5b994417d4-config-data\") pod \"c73a61da-0c40-4a26-888c-ab5b994417d4\" (UID: \"c73a61da-0c40-4a26-888c-ab5b994417d4\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.332423 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-combined-ca-bundle\") pod \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.333039 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-internal-tls-certs\") pod \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.333367 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9jdn\" (UniqueName: \"kubernetes.io/projected/c73a61da-0c40-4a26-888c-ab5b994417d4-kube-api-access-w9jdn\") pod \"c73a61da-0c40-4a26-888c-ab5b994417d4\" (UID: \"c73a61da-0c40-4a26-888c-ab5b994417d4\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.333567 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c73a61da-0c40-4a26-888c-ab5b994417d4-combined-ca-bundle\") pod \"c73a61da-0c40-4a26-888c-ab5b994417d4\" (UID: \"c73a61da-0c40-4a26-888c-ab5b994417d4\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.333692 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-public-tls-certs\") pod \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.333927 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-logs\") pod \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.334148 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-combined-ca-bundle\") pod \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.334318 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-config-data\") pod \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\" (UID: \"9acd8c29-93f5-4b0f-8b32-d8f3d576767b\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.340778 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-scripts\") pod \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.341768 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-logs" (OuterVolumeSpecName: "logs") pod "9acd8c29-93f5-4b0f-8b32-d8f3d576767b" (UID: "9acd8c29-93f5-4b0f-8b32-d8f3d576767b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.342256 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c-combined-ca-bundle\") pod \"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c\" (UID: \"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.343335 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzr6m\" (UniqueName: \"kubernetes.io/projected/444ec27a-77d0-4b01-b59e-c6590c46a0c2-kube-api-access-fzr6m\") pod \"444ec27a-77d0-4b01-b59e-c6590c46a0c2\" (UID: \"444ec27a-77d0-4b01-b59e-c6590c46a0c2\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.343390 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/444ec27a-77d0-4b01-b59e-c6590c46a0c2-config-data\") pod \"444ec27a-77d0-4b01-b59e-c6590c46a0c2\" (UID: \"444ec27a-77d0-4b01-b59e-c6590c46a0c2\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.343497 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.343529 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42pvr\" (UniqueName: \"kubernetes.io/projected/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-kube-api-access-42pvr\") pod \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.343551 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxs28\" (UniqueName: \"kubernetes.io/projected/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c-kube-api-access-dxs28\") pod \"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c\" (UID: \"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.343571 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/444ec27a-77d0-4b01-b59e-c6590c46a0c2-combined-ca-bundle\") pod \"444ec27a-77d0-4b01-b59e-c6590c46a0c2\" (UID: \"444ec27a-77d0-4b01-b59e-c6590c46a0c2\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.343596 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-internal-tls-certs\") pod \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.343625 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-logs\") pod \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\" (UID: \"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc\") " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.344159 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.344174 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.348662 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-logs" (OuterVolumeSpecName: "logs") pod "d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" (UID: "d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.352510 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5b4ff7d6cb-m2g6c"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.362979 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5b4ff7d6cb-m2g6c"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.365728 4810 scope.go:117] "RemoveContainer" containerID="4a8d673779dd5bbbb9060129ba8fe9ee012663d8194ce9496b1684874e3903dc" Dec 01 14:59:51 crc kubenswrapper[4810]: E1201 14:59:51.367034 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a8d673779dd5bbbb9060129ba8fe9ee012663d8194ce9496b1684874e3903dc\": container with ID starting with 4a8d673779dd5bbbb9060129ba8fe9ee012663d8194ce9496b1684874e3903dc not found: ID does not exist" containerID="4a8d673779dd5bbbb9060129ba8fe9ee012663d8194ce9496b1684874e3903dc" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.367165 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a8d673779dd5bbbb9060129ba8fe9ee012663d8194ce9496b1684874e3903dc"} err="failed to get container status \"4a8d673779dd5bbbb9060129ba8fe9ee012663d8194ce9496b1684874e3903dc\": rpc error: code = NotFound desc = could not find container \"4a8d673779dd5bbbb9060129ba8fe9ee012663d8194ce9496b1684874e3903dc\": container with ID starting with 4a8d673779dd5bbbb9060129ba8fe9ee012663d8194ce9496b1684874e3903dc not found: ID does not exist" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.367202 4810 scope.go:117] "RemoveContainer" containerID="0af6bd36655b389f1090223a090ab3a8d47cace5e570943004f63de4c038edb3" Dec 01 14:59:51 crc kubenswrapper[4810]: E1201 14:59:51.374064 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0af6bd36655b389f1090223a090ab3a8d47cace5e570943004f63de4c038edb3\": container with ID starting with 0af6bd36655b389f1090223a090ab3a8d47cace5e570943004f63de4c038edb3 not found: ID does not exist" containerID="0af6bd36655b389f1090223a090ab3a8d47cace5e570943004f63de4c038edb3" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.374132 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0af6bd36655b389f1090223a090ab3a8d47cace5e570943004f63de4c038edb3"} err="failed to get container status \"0af6bd36655b389f1090223a090ab3a8d47cace5e570943004f63de4c038edb3\": rpc error: code = NotFound desc = could not find container \"0af6bd36655b389f1090223a090ab3a8d47cace5e570943004f63de4c038edb3\": container with ID starting with 0af6bd36655b389f1090223a090ab3a8d47cace5e570943004f63de4c038edb3 not found: ID does not exist" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.374163 4810 scope.go:117] "RemoveContainer" containerID="030a775eb17300c273b82602d862651690419af2f0861ccfe4b4d6bdc64e3a5e" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.380179 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-kube-api-access-42pvr" (OuterVolumeSpecName: "kube-api-access-42pvr") pod "d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" (UID: "d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc"). InnerVolumeSpecName "kube-api-access-42pvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.384413 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c73a61da-0c40-4a26-888c-ab5b994417d4-kube-api-access-w9jdn" (OuterVolumeSpecName: "kube-api-access-w9jdn") pod "c73a61da-0c40-4a26-888c-ab5b994417d4" (UID: "c73a61da-0c40-4a26-888c-ab5b994417d4"). InnerVolumeSpecName "kube-api-access-w9jdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.384534 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/444ec27a-77d0-4b01-b59e-c6590c46a0c2-kube-api-access-fzr6m" (OuterVolumeSpecName: "kube-api-access-fzr6m") pod "444ec27a-77d0-4b01-b59e-c6590c46a0c2" (UID: "444ec27a-77d0-4b01-b59e-c6590c46a0c2"). InnerVolumeSpecName "kube-api-access-fzr6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.398857 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-scripts" (OuterVolumeSpecName: "scripts") pod "d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" (UID: "d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.399043 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-kube-api-access-666rf" (OuterVolumeSpecName: "kube-api-access-666rf") pod "9acd8c29-93f5-4b0f-8b32-d8f3d576767b" (UID: "9acd8c29-93f5-4b0f-8b32-d8f3d576767b"). InnerVolumeSpecName "kube-api-access-666rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.400063 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c-kube-api-access-dxs28" (OuterVolumeSpecName: "kube-api-access-dxs28") pod "992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c" (UID: "992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c"). InnerVolumeSpecName "kube-api-access-dxs28". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.406744 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" (UID: "d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.412322 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.423280 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.433975 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-config-data" (OuterVolumeSpecName: "config-data") pod "9acd8c29-93f5-4b0f-8b32-d8f3d576767b" (UID: "9acd8c29-93f5-4b0f-8b32-d8f3d576767b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.434066 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c-config-data" (OuterVolumeSpecName: "config-data") pod "992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c" (UID: "992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.436413 4810 scope.go:117] "RemoveContainer" containerID="030a775eb17300c273b82602d862651690419af2f0861ccfe4b4d6bdc64e3a5e" Dec 01 14:59:51 crc kubenswrapper[4810]: E1201 14:59:51.439093 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"030a775eb17300c273b82602d862651690419af2f0861ccfe4b4d6bdc64e3a5e\": container with ID starting with 030a775eb17300c273b82602d862651690419af2f0861ccfe4b4d6bdc64e3a5e not found: ID does not exist" containerID="030a775eb17300c273b82602d862651690419af2f0861ccfe4b4d6bdc64e3a5e" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.439134 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"030a775eb17300c273b82602d862651690419af2f0861ccfe4b4d6bdc64e3a5e"} err="failed to get container status \"030a775eb17300c273b82602d862651690419af2f0861ccfe4b4d6bdc64e3a5e\": rpc error: code = NotFound desc = could not find container \"030a775eb17300c273b82602d862651690419af2f0861ccfe4b4d6bdc64e3a5e\": container with ID starting with 030a775eb17300c273b82602d862651690419af2f0861ccfe4b4d6bdc64e3a5e not found: ID does not exist" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.441865 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" (UID: "d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.447717 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9acd8c29-93f5-4b0f-8b32-d8f3d576767b" (UID: "9acd8c29-93f5-4b0f-8b32-d8f3d576767b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.448379 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzr6m\" (UniqueName: \"kubernetes.io/projected/444ec27a-77d0-4b01-b59e-c6590c46a0c2-kube-api-access-fzr6m\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.448639 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.448659 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42pvr\" (UniqueName: \"kubernetes.io/projected/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-kube-api-access-42pvr\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.448669 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxs28\" (UniqueName: \"kubernetes.io/projected/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c-kube-api-access-dxs28\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.448677 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.448686 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.448694 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-666rf\" (UniqueName: \"kubernetes.io/projected/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-kube-api-access-666rf\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.448702 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9jdn\" (UniqueName: \"kubernetes.io/projected/c73a61da-0c40-4a26-888c-ab5b994417d4-kube-api-access-w9jdn\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.448711 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.448721 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.448731 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.448738 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.456347 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c73a61da-0c40-4a26-888c-ab5b994417d4-config-data" (OuterVolumeSpecName: "config-data") pod "c73a61da-0c40-4a26-888c-ab5b994417d4" (UID: "c73a61da-0c40-4a26-888c-ab5b994417d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.459717 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5d8fc4ccc9-tqkdf" podUID="b88404c2-d4bf-4257-986a-1b78178fca88" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.194:5353: i/o timeout" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.465172 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9acd8c29-93f5-4b0f-8b32-d8f3d576767b" (UID: "9acd8c29-93f5-4b0f-8b32-d8f3d576767b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.465726 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/444ec27a-77d0-4b01-b59e-c6590c46a0c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "444ec27a-77d0-4b01-b59e-c6590c46a0c2" (UID: "444ec27a-77d0-4b01-b59e-c6590c46a0c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.482749 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c" (UID: "992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.483234 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.484178 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/444ec27a-77d0-4b01-b59e-c6590c46a0c2-config-data" (OuterVolumeSpecName: "config-data") pod "444ec27a-77d0-4b01-b59e-c6590c46a0c2" (UID: "444ec27a-77d0-4b01-b59e-c6590c46a0c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.484339 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c73a61da-0c40-4a26-888c-ab5b994417d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c73a61da-0c40-4a26-888c-ab5b994417d4" (UID: "c73a61da-0c40-4a26-888c-ab5b994417d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.486252 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" (UID: "d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.487416 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9acd8c29-93f5-4b0f-8b32-d8f3d576767b" (UID: "9acd8c29-93f5-4b0f-8b32-d8f3d576767b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.497622 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-config-data" (OuterVolumeSpecName: "config-data") pod "d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" (UID: "d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.519880 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e79c5698-ab2d-4292-bda9-fab8dbbb2abb","Type":"ContainerDied","Data":"d3af5f0b7ed3a270d5b91e569b451e6bd182c3e93d804653e0703cad50ee10d6"} Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.519962 4810 scope.go:117] "RemoveContainer" containerID="085bfd508911509acb2f401ff11dc8f197ed30234eac469830bf853fef1a201a" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.520114 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.525085 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e16305af-6573-4cfd-ae76-f596d410920b","Type":"ContainerDied","Data":"6f18f95df324cde3ccd3262cfa697fd4db9f0bc111e923c5a0bca2ed88be7313"} Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.525265 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.531941 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3","Type":"ContainerDied","Data":"9516a9dad515dab02cb19edfecb587a3cf75167d1413868d1ee522450e715d7b"} Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.532095 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.546420 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6c5f549c7d-52fkz" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.551221 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.551256 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/444ec27a-77d0-4b01-b59e-c6590c46a0c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.551265 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.551276 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.551287 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c73a61da-0c40-4a26-888c-ab5b994417d4-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.551294 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.551302 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9acd8c29-93f5-4b0f-8b32-d8f3d576767b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.551310 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c73a61da-0c40-4a26-888c-ab5b994417d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.551318 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.551325 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/444ec27a-77d0-4b01-b59e-c6590c46a0c2-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.553085 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9dd705f6-0d8f-49df-aff6-813792212999","Type":"ContainerDied","Data":"16375e9d336a56f3aaf7a3394fb2b605a0b1e82cbab8e3ac3b7db8ab2f0f5f80"} Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.553170 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.565040 4810 scope.go:117] "RemoveContainer" containerID="ebb0282754daba83e3b7cf8ef6784b85c2feea8bbea0bae1fde73a5239ce6c9c" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.576140 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc","Type":"ContainerDied","Data":"8b54ceba4a7e600832243250c7d73799a28d24f2dca3d75e7a1f10997db13fb1"} Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.576332 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.582313 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.587214 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6","Type":"ContainerDied","Data":"3e6e63432073a7b17d3fb456f38359b231178a6af7b3213df4e0f311b915c146"} Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.587248 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.589855 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c73a61da-0c40-4a26-888c-ab5b994417d4","Type":"ContainerDied","Data":"0427c38a149f10233ec14f720e735356ae66eeb0a15ee032d631eff0f84f5475"} Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.590006 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.598728 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-84864d5bd5-wmq86" event={"ID":"129ae854-1430-465f-a177-0c459bd8b4c8","Type":"ContainerDied","Data":"46e8408cf96c668ecb1af9153b47b16715be23c99fb60a271b3559ab78086e7e"} Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.599316 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-84864d5bd5-wmq86" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.610455 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.616730 4810 scope.go:117] "RemoveContainer" containerID="6a66d0aa3094f13ae73e9e79b4f1f5db114907ff336ae2e84fa9c09a936675a3" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.623399 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6c5f549c7d-52fkz"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.629170 4810 generic.go:334] "Generic (PLEG): container finished" podID="444ec27a-77d0-4b01-b59e-c6590c46a0c2" containerID="ffc643854f69e1e2f23a40254dbae2944125c570cca2c9e027760a19987124a6" exitCode=0 Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.629548 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"444ec27a-77d0-4b01-b59e-c6590c46a0c2","Type":"ContainerDied","Data":"ffc643854f69e1e2f23a40254dbae2944125c570cca2c9e027760a19987124a6"} Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.629717 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"444ec27a-77d0-4b01-b59e-c6590c46a0c2","Type":"ContainerDied","Data":"a8047f5795ca6b6c992b59d4f4a759af0edb049e712a524429d10a5d1fdbc17f"} Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.629250 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.633159 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9acd8c29-93f5-4b0f-8b32-d8f3d576767b","Type":"ContainerDied","Data":"848a6230705ec969b06b50a8f363749de8fa626d9de709eafed5297752f8496d"} Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.633331 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.639842 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-6c5f549c7d-52fkz"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.640785 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c","Type":"ContainerDied","Data":"2fb0aac5102f60817c90c753fa782cb17744750246e5c4e2c759566245a5b0d6"} Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.640913 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.648237 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone8cd1-account-delete-c69f6" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.648282 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance324e-account-delete-6s4dm" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.648318 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbicanff71-account-delete-2fnl8" podUID="1ef9fa82-f302-47e9-bed3-b02f34a2340f" containerName="mariadb-account-delete" containerID="cri-o://dfcba994df93f276679f0b982a3fda4b0ba3b97518b8e6d0c24012c34d33e149" gracePeriod=30 Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.648443 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder17c6-account-delete-qktk6" podUID="bf15ccdf-1e4e-44ab-8ccb-819502935a7a" containerName="mariadb-account-delete" containerID="cri-o://afb9f73f83766bd0ed43d8d077992ad62bc5c10098a5767cbf7c3f030466f40f" gracePeriod=30 Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.648728 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutronc91e-account-delete-t585q" podUID="f7b06553-ec05-4c6f-82c0-ef0b4dd3005a" containerName="mariadb-account-delete" containerID="cri-o://2ba07f4b84a146833c2e8b31369eaa1e961a68fc47ee9b4c60c092208482cff9" gracePeriod=30 Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.652667 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novaapic0fd-account-delete-w7ttc" podUID="476b71ac-1050-46a6-b51c-168e75b1a350" containerName="mariadb-account-delete" containerID="cri-o://691a88e3e213a94e7564d8b43ac6e1b2e48afa2314fc5a29d364c71768f366fb" gracePeriod=30 Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.653550 4810 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novacell096e4-account-delete-qqlg4" secret="" err="secret \"galera-openstack-dockercfg-5jwrv\" not found" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.690945 4810 scope.go:117] "RemoveContainer" containerID="ac4ab87e1191fa810eaa9e1774fc31d9ed1e7d6475958ea936337c2a47b7b3e0" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.706678 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.712767 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: E1201 14:59:51.756892 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:51 crc kubenswrapper[4810]: E1201 14:59:51.756953 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts podName:4f7b5a25-eed0-4122-a036-093ac7adca22 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:53.756939104 +0000 UTC m=+1559.520448707 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts") pod "novacell096e4-account-delete-qqlg4" (UID: "4f7b5a25-eed0-4122-a036-093ac7adca22") : configmap "openstack-scripts" not found Dec 01 14:59:51 crc kubenswrapper[4810]: E1201 14:59:51.757272 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:51 crc kubenswrapper[4810]: E1201 14:59:51.757305 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts podName:1ef9fa82-f302-47e9-bed3-b02f34a2340f nodeName:}" failed. No retries permitted until 2025-12-01 14:59:53.757297624 +0000 UTC m=+1559.520807227 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts") pod "barbicanff71-account-delete-2fnl8" (UID: "1ef9fa82-f302-47e9-bed3-b02f34a2340f") : configmap "openstack-scripts" not found Dec 01 14:59:51 crc kubenswrapper[4810]: E1201 14:59:51.757330 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 01 14:59:51 crc kubenswrapper[4810]: E1201 14:59:51.757385 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-config-data podName:3b81b6a9-0be6-4ce1-81b4-043edb1c221f nodeName:}" failed. No retries permitted until 2025-12-01 14:59:59.757379666 +0000 UTC m=+1565.520889269 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-config-data") pod "rabbitmq-server-0" (UID: "3b81b6a9-0be6-4ce1-81b4-043edb1c221f") : configmap "rabbitmq-config-data" not found Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.766077 4810 scope.go:117] "RemoveContainer" containerID="90f166a6125c50147c3cb78e63b7459cc98d8a4218a7a6ddd44ce0fb9b8a4fc2" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.786678 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.792342 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.808075 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.817130 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.832834 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone8cd1-account-delete-c69f6"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.839705 4810 scope.go:117] "RemoveContainer" containerID="3cb0e1b4c05926fcfa43bb4c227b62243d1a9c4234879cdb02f9c40af684935e" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.840055 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone8cd1-account-delete-c69f6"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.848319 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.857177 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.866497 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance324e-account-delete-6s4dm"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.869224 4810 scope.go:117] "RemoveContainer" containerID="b1b79938d10b71fee29a6515a87d1fb415850230bcaa5ce800b1e5ba6bb61661" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.873527 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance324e-account-delete-6s4dm"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.881310 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-84864d5bd5-wmq86"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.889899 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-84864d5bd5-wmq86"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.896153 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.901715 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.911438 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.915031 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.923046 4810 scope.go:117] "RemoveContainer" containerID="ca8c16cdb56a677709c0ac603398c85e43e5a2f9e1cb3036f753b35de804a1f8" Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.923160 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.937302 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.943807 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.948344 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.954059 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.958614 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 14:59:51 crc kubenswrapper[4810]: I1201 14:59:51.972740 4810 scope.go:117] "RemoveContainer" containerID="1077d64ab29e8e4a5e08c7d81078a1254c035e9de419585bf1944ac1e43a909c" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.011775 4810 scope.go:117] "RemoveContainer" containerID="aba75285c97143bda5c120fd54711afa0a88b59eeac5b761fa88578cc1f0f3b2" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.072254 4810 scope.go:117] "RemoveContainer" containerID="630e5f798fca1eedf921c4930a1d5c0f33b14ea41b717a6558b2a6c06e4903f1" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.213289 4810 scope.go:117] "RemoveContainer" containerID="08362c9d20e6c5ffc7ef2c58a68a05a6da9d37da0f16242f1435dc93d6db4209" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.247874 4810 scope.go:117] "RemoveContainer" containerID="97cf7e0548b92abb2ae87397174ade605222b8ac75fdc9ae22ab5abf9a072ec8" Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.265526 4810 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.265604 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-config-data podName:56bbf32f-35b5-4f92-bfc0-9266b0ecccac nodeName:}" failed. No retries permitted until 2025-12-01 15:00:00.265587465 +0000 UTC m=+1566.029097068 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-config-data") pod "rabbitmq-cell1-server-0" (UID: "56bbf32f-35b5-4f92-bfc0-9266b0ecccac") : configmap "rabbitmq-cell1-config-data" not found Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.273303 4810 scope.go:117] "RemoveContainer" containerID="6cffd8cec4aa7e3f42f1a3b9cff8d7da52efafaece6f09b30bae42093876b093" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.312933 4810 scope.go:117] "RemoveContainer" containerID="8c00c76d0791b2c2bc8a97328993ef7e1e8aa6997c54b9aa8cb7c3385b227dd7" Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.324111 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.324522 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.324803 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.324861 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-x82dv" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovsdb-server" Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.324966 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.326210 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.327241 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.327260 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-x82dv" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovs-vswitchd" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.500250 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="129ae854-1430-465f-a177-0c459bd8b4c8" path="/var/lib/kubelet/pods/129ae854-1430-465f-a177-0c459bd8b4c8/volumes" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.501434 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a3a8b5d-01e4-4396-8a9b-abd85dc55572" path="/var/lib/kubelet/pods/2a3a8b5d-01e4-4396-8a9b-abd85dc55572/volumes" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.502093 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31b83b20-83c4-4e06-8eb0-1e1d83d4b160" path="/var/lib/kubelet/pods/31b83b20-83c4-4e06-8eb0-1e1d83d4b160/volumes" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.503382 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39ce8411-05ee-4022-bd7a-87e46c2cee8d" path="/var/lib/kubelet/pods/39ce8411-05ee-4022-bd7a-87e46c2cee8d/volumes" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.503953 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" path="/var/lib/kubelet/pods/3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6/volumes" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.504611 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="444ec27a-77d0-4b01-b59e-c6590c46a0c2" path="/var/lib/kubelet/pods/444ec27a-77d0-4b01-b59e-c6590c46a0c2/volumes" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.505517 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75872443-9a8a-4b7b-9c70-036dc558594e" path="/var/lib/kubelet/pods/75872443-9a8a-4b7b-9c70-036dc558594e/volumes" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.506010 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="916e9520-9e98-42a8-bbb5-cb5f4029daab" path="/var/lib/kubelet/pods/916e9520-9e98-42a8-bbb5-cb5f4029daab/volumes" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.506520 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91932480-76aa-4751-9251-3ad82d849882" path="/var/lib/kubelet/pods/91932480-76aa-4751-9251-3ad82d849882/volumes" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.507555 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c" path="/var/lib/kubelet/pods/992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c/volumes" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.508033 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9acd8c29-93f5-4b0f-8b32-d8f3d576767b" path="/var/lib/kubelet/pods/9acd8c29-93f5-4b0f-8b32-d8f3d576767b/volumes" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.508773 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dd705f6-0d8f-49df-aff6-813792212999" path="/var/lib/kubelet/pods/9dd705f6-0d8f-49df-aff6-813792212999/volumes" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.509954 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdd90cbb-4242-4227-9de4-09f99e60a516" path="/var/lib/kubelet/pods/bdd90cbb-4242-4227-9de4-09f99e60a516/volumes" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.510787 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" path="/var/lib/kubelet/pods/c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3/volumes" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.511430 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c73a61da-0c40-4a26-888c-ab5b994417d4" path="/var/lib/kubelet/pods/c73a61da-0c40-4a26-888c-ab5b994417d4/volumes" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.512447 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" path="/var/lib/kubelet/pods/d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc/volumes" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.513013 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e16305af-6573-4cfd-ae76-f596d410920b" path="/var/lib/kubelet/pods/e16305af-6573-4cfd-ae76-f596d410920b/volumes" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.514013 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e79c5698-ab2d-4292-bda9-fab8dbbb2abb" path="/var/lib/kubelet/pods/e79c5698-ab2d-4292-bda9-fab8dbbb2abb/volumes" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.600315 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="3b81b6a9-0be6-4ce1-81b4-043edb1c221f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.629309 4810 scope.go:117] "RemoveContainer" containerID="b8a19753910ff67350d9da738f60efda1e12ea5411279e28d972ad0cb28c8f79" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.633091 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.668487 4810 generic.go:334] "Generic (PLEG): container finished" podID="5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0" containerID="e064be1964d87f2e16398d0a402ce3f57b2caf18c89d6b179f3ef2a94e1374ac" exitCode=0 Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.668571 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-65bbfc7449-rtqm6" event={"ID":"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0","Type":"ContainerDied","Data":"e064be1964d87f2e16398d0a402ce3f57b2caf18c89d6b179f3ef2a94e1374ac"} Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.671635 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4c309abf-824f-4264-859b-ded1f0a9acaf-config-data-generated\") pod \"4c309abf-824f-4264-859b-ded1f0a9acaf\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.671691 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c309abf-824f-4264-859b-ded1f0a9acaf-operator-scripts\") pod \"4c309abf-824f-4264-859b-ded1f0a9acaf\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.671817 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c309abf-824f-4264-859b-ded1f0a9acaf-kolla-config\") pod \"4c309abf-824f-4264-859b-ded1f0a9acaf\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.671843 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"4c309abf-824f-4264-859b-ded1f0a9acaf\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.671888 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28jjc\" (UniqueName: \"kubernetes.io/projected/4c309abf-824f-4264-859b-ded1f0a9acaf-kube-api-access-28jjc\") pod \"4c309abf-824f-4264-859b-ded1f0a9acaf\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.671907 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c309abf-824f-4264-859b-ded1f0a9acaf-galera-tls-certs\") pod \"4c309abf-824f-4264-859b-ded1f0a9acaf\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.671937 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4c309abf-824f-4264-859b-ded1f0a9acaf-config-data-default\") pod \"4c309abf-824f-4264-859b-ded1f0a9acaf\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.671995 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c309abf-824f-4264-859b-ded1f0a9acaf-combined-ca-bundle\") pod \"4c309abf-824f-4264-859b-ded1f0a9acaf\" (UID: \"4c309abf-824f-4264-859b-ded1f0a9acaf\") " Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.672152 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c309abf-824f-4264-859b-ded1f0a9acaf-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "4c309abf-824f-4264-859b-ded1f0a9acaf" (UID: "4c309abf-824f-4264-859b-ded1f0a9acaf"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.672520 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4c309abf-824f-4264-859b-ded1f0a9acaf-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.672655 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c309abf-824f-4264-859b-ded1f0a9acaf-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "4c309abf-824f-4264-859b-ded1f0a9acaf" (UID: "4c309abf-824f-4264-859b-ded1f0a9acaf"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.672715 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c309abf-824f-4264-859b-ded1f0a9acaf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4c309abf-824f-4264-859b-ded1f0a9acaf" (UID: "4c309abf-824f-4264-859b-ded1f0a9acaf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.673500 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c309abf-824f-4264-859b-ded1f0a9acaf-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "4c309abf-824f-4264-859b-ded1f0a9acaf" (UID: "4c309abf-824f-4264-859b-ded1f0a9acaf"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.679994 4810 generic.go:334] "Generic (PLEG): container finished" podID="4c309abf-824f-4264-859b-ded1f0a9acaf" containerID="1b9b162f9bfb794c639b6d75ebcad8b08b8c31a6a252c25ffdb8e930eaa77acf" exitCode=0 Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.680059 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4c309abf-824f-4264-859b-ded1f0a9acaf","Type":"ContainerDied","Data":"1b9b162f9bfb794c639b6d75ebcad8b08b8c31a6a252c25ffdb8e930eaa77acf"} Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.680087 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"4c309abf-824f-4264-859b-ded1f0a9acaf","Type":"ContainerDied","Data":"37b9fc728b3901032eb1ebaa116b2a4617e7fa0ed61602d41ae1488dd53881a9"} Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.680159 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.683120 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_604a8b79-fcf3-40cf-82e6-8af2cf667b0f/ovn-northd/0.log" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.683148 4810 generic.go:334] "Generic (PLEG): container finished" podID="604a8b79-fcf3-40cf-82e6-8af2cf667b0f" containerID="87439695570ef9b3e9da98681ae2d1dec7fb73d967ccb8912a2e40ca13529cf7" exitCode=139 Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.683196 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"604a8b79-fcf3-40cf-82e6-8af2cf667b0f","Type":"ContainerDied","Data":"87439695570ef9b3e9da98681ae2d1dec7fb73d967ccb8912a2e40ca13529cf7"} Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.684881 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c309abf-824f-4264-859b-ded1f0a9acaf-kube-api-access-28jjc" (OuterVolumeSpecName: "kube-api-access-28jjc") pod "4c309abf-824f-4264-859b-ded1f0a9acaf" (UID: "4c309abf-824f-4264-859b-ded1f0a9acaf"). InnerVolumeSpecName "kube-api-access-28jjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.686763 4810 generic.go:334] "Generic (PLEG): container finished" podID="3b81b6a9-0be6-4ce1-81b4-043edb1c221f" containerID="74e6d90ac1ab1306dd26c5e8c6cd4cd8b45f47eb0ab68865c9c1575f423fce76" exitCode=0 Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.686797 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3b81b6a9-0be6-4ce1-81b4-043edb1c221f","Type":"ContainerDied","Data":"74e6d90ac1ab1306dd26c5e8c6cd4cd8b45f47eb0ab68865c9c1575f423fce76"} Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.689402 4810 scope.go:117] "RemoveContainer" containerID="ffc643854f69e1e2f23a40254dbae2944125c570cca2c9e027760a19987124a6" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.693285 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "mysql-db") pod "4c309abf-824f-4264-859b-ded1f0a9acaf" (UID: "4c309abf-824f-4264-859b-ded1f0a9acaf"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.699199 4810 generic.go:334] "Generic (PLEG): container finished" podID="56bbf32f-35b5-4f92-bfc0-9266b0ecccac" containerID="82a26c84ad224e60a4fa117cf5e1d03643cfd46868d1eeb656fac71a268ad1e5" exitCode=0 Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.699254 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"56bbf32f-35b5-4f92-bfc0-9266b0ecccac","Type":"ContainerDied","Data":"82a26c84ad224e60a4fa117cf5e1d03643cfd46868d1eeb656fac71a268ad1e5"} Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.700955 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novacell096e4-account-delete-qqlg4" podUID="4f7b5a25-eed0-4122-a036-093ac7adca22" containerName="mariadb-account-delete" containerID="cri-o://ba6bbaa4e5865842e4d015f8e09df7c3af89eab522256f5f54fc25471d6b214f" gracePeriod=30 Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.738669 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c309abf-824f-4264-859b-ded1f0a9acaf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c309abf-824f-4264-859b-ded1f0a9acaf" (UID: "4c309abf-824f-4264-859b-ded1f0a9acaf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.740030 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c309abf-824f-4264-859b-ded1f0a9acaf-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "4c309abf-824f-4264-859b-ded1f0a9acaf" (UID: "4c309abf-824f-4264-859b-ded1f0a9acaf"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.774493 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c309abf-824f-4264-859b-ded1f0a9acaf-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.774526 4810 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4c309abf-824f-4264-859b-ded1f0a9acaf-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.774551 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.774619 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts podName:476b71ac-1050-46a6-b51c-168e75b1a350 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:56.774603375 +0000 UTC m=+1562.538112978 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts") pod "novaapic0fd-account-delete-w7ttc" (UID: "476b71ac-1050-46a6-b51c-168e75b1a350") : configmap "openstack-scripts" not found Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.774558 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.774732 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28jjc\" (UniqueName: \"kubernetes.io/projected/4c309abf-824f-4264-859b-ded1f0a9acaf-kube-api-access-28jjc\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.774756 4810 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c309abf-824f-4264-859b-ded1f0a9acaf-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.774770 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4c309abf-824f-4264-859b-ded1f0a9acaf-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.774783 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c309abf-824f-4264-859b-ded1f0a9acaf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.791673 4810 scope.go:117] "RemoveContainer" containerID="ffc643854f69e1e2f23a40254dbae2944125c570cca2c9e027760a19987124a6" Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.792145 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffc643854f69e1e2f23a40254dbae2944125c570cca2c9e027760a19987124a6\": container with ID starting with ffc643854f69e1e2f23a40254dbae2944125c570cca2c9e027760a19987124a6 not found: ID does not exist" containerID="ffc643854f69e1e2f23a40254dbae2944125c570cca2c9e027760a19987124a6" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.792197 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffc643854f69e1e2f23a40254dbae2944125c570cca2c9e027760a19987124a6"} err="failed to get container status \"ffc643854f69e1e2f23a40254dbae2944125c570cca2c9e027760a19987124a6\": rpc error: code = NotFound desc = could not find container \"ffc643854f69e1e2f23a40254dbae2944125c570cca2c9e027760a19987124a6\": container with ID starting with ffc643854f69e1e2f23a40254dbae2944125c570cca2c9e027760a19987124a6 not found: ID does not exist" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.792230 4810 scope.go:117] "RemoveContainer" containerID="1f6241fff3fa641130500b0c5b06297d792485793ad36fdc4dd06993a79f221d" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.796487 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.811664 4810 scope.go:117] "RemoveContainer" containerID="6c951f3e299a57aa9f4c575d8d057fd04b80e714ae7050a14a4ca0758c9a0549" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.839717 4810 scope.go:117] "RemoveContainer" containerID="c0a6acaf5c3f0108d3d50cb630c0e2cb9470e93a6443905cbd9058d2ef094517" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.872548 4810 scope.go:117] "RemoveContainer" containerID="1b9b162f9bfb794c639b6d75ebcad8b08b8c31a6a252c25ffdb8e930eaa77acf" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.876817 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.876904 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.876952 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts podName:f7b06553-ec05-4c6f-82c0-ef0b4dd3005a nodeName:}" failed. No retries permitted until 2025-12-01 14:59:56.876936146 +0000 UTC m=+1562.640445749 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts") pod "neutronc91e-account-delete-t585q" (UID: "f7b06553-ec05-4c6f-82c0-ef0b4dd3005a") : configmap "openstack-scripts" not found Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.877348 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.877376 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts podName:bf15ccdf-1e4e-44ab-8ccb-819502935a7a nodeName:}" failed. No retries permitted until 2025-12-01 14:59:56.877367628 +0000 UTC m=+1562.640877231 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts") pod "cinder17c6-account-delete-qktk6" (UID: "bf15ccdf-1e4e-44ab-8ccb-819502935a7a") : configmap "openstack-scripts" not found Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.911065 4810 scope.go:117] "RemoveContainer" containerID="b5e8843e09485339399f86646edf6056d7c953642f1b41d38018280215001cb5" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.942081 4810 scope.go:117] "RemoveContainer" containerID="1b9b162f9bfb794c639b6d75ebcad8b08b8c31a6a252c25ffdb8e930eaa77acf" Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.942589 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b9b162f9bfb794c639b6d75ebcad8b08b8c31a6a252c25ffdb8e930eaa77acf\": container with ID starting with 1b9b162f9bfb794c639b6d75ebcad8b08b8c31a6a252c25ffdb8e930eaa77acf not found: ID does not exist" containerID="1b9b162f9bfb794c639b6d75ebcad8b08b8c31a6a252c25ffdb8e930eaa77acf" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.942619 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b9b162f9bfb794c639b6d75ebcad8b08b8c31a6a252c25ffdb8e930eaa77acf"} err="failed to get container status \"1b9b162f9bfb794c639b6d75ebcad8b08b8c31a6a252c25ffdb8e930eaa77acf\": rpc error: code = NotFound desc = could not find container \"1b9b162f9bfb794c639b6d75ebcad8b08b8c31a6a252c25ffdb8e930eaa77acf\": container with ID starting with 1b9b162f9bfb794c639b6d75ebcad8b08b8c31a6a252c25ffdb8e930eaa77acf not found: ID does not exist" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.942637 4810 scope.go:117] "RemoveContainer" containerID="b5e8843e09485339399f86646edf6056d7c953642f1b41d38018280215001cb5" Dec 01 14:59:52 crc kubenswrapper[4810]: E1201 14:59:52.942904 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5e8843e09485339399f86646edf6056d7c953642f1b41d38018280215001cb5\": container with ID starting with b5e8843e09485339399f86646edf6056d7c953642f1b41d38018280215001cb5 not found: ID does not exist" containerID="b5e8843e09485339399f86646edf6056d7c953642f1b41d38018280215001cb5" Dec 01 14:59:52 crc kubenswrapper[4810]: I1201 14:59:52.942953 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5e8843e09485339399f86646edf6056d7c953642f1b41d38018280215001cb5"} err="failed to get container status \"b5e8843e09485339399f86646edf6056d7c953642f1b41d38018280215001cb5\": rpc error: code = NotFound desc = could not find container \"b5e8843e09485339399f86646edf6056d7c953642f1b41d38018280215001cb5\": container with ID starting with b5e8843e09485339399f86646edf6056d7c953642f1b41d38018280215001cb5 not found: ID does not exist" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.017974 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.028397 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 14:59:53 crc kubenswrapper[4810]: E1201 14:59:53.268059 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 87439695570ef9b3e9da98681ae2d1dec7fb73d967ccb8912a2e40ca13529cf7 is running failed: container process not found" containerID="87439695570ef9b3e9da98681ae2d1dec7fb73d967ccb8912a2e40ca13529cf7" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 01 14:59:53 crc kubenswrapper[4810]: E1201 14:59:53.269115 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 87439695570ef9b3e9da98681ae2d1dec7fb73d967ccb8912a2e40ca13529cf7 is running failed: container process not found" containerID="87439695570ef9b3e9da98681ae2d1dec7fb73d967ccb8912a2e40ca13529cf7" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 01 14:59:53 crc kubenswrapper[4810]: E1201 14:59:53.269392 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 87439695570ef9b3e9da98681ae2d1dec7fb73d967ccb8912a2e40ca13529cf7 is running failed: container process not found" containerID="87439695570ef9b3e9da98681ae2d1dec7fb73d967ccb8912a2e40ca13529cf7" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 01 14:59:53 crc kubenswrapper[4810]: E1201 14:59:53.269421 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 87439695570ef9b3e9da98681ae2d1dec7fb73d967ccb8912a2e40ca13529cf7 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="604a8b79-fcf3-40cf-82e6-8af2cf667b0f" containerName="ovn-northd" Dec 01 14:59:53 crc kubenswrapper[4810]: E1201 14:59:53.322505 4810 secret.go:188] Couldn't get secret openstack/neutron-httpd-config: secret "neutron-httpd-config" not found Dec 01 14:59:53 crc kubenswrapper[4810]: E1201 14:59:53.322578 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-httpd-config podName:ba5f8a69-22cf-4196-bc89-ffa60eda9303 nodeName:}" failed. No retries permitted until 2025-12-01 15:00:01.322560015 +0000 UTC m=+1567.086069618 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "httpd-config" (UniqueName: "kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-httpd-config") pod "neutron-5477779d95-wg2vp" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303") : secret "neutron-httpd-config" not found Dec 01 14:59:53 crc kubenswrapper[4810]: E1201 14:59:53.322953 4810 secret.go:188] Couldn't get secret openstack/neutron-config: secret "neutron-config" not found Dec 01 14:59:53 crc kubenswrapper[4810]: E1201 14:59:53.322985 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-config podName:ba5f8a69-22cf-4196-bc89-ffa60eda9303 nodeName:}" failed. No retries permitted until 2025-12-01 15:00:01.322975666 +0000 UTC m=+1567.086485269 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-config") pod "neutron-5477779d95-wg2vp" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303") : secret "neutron-config" not found Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.412963 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.450708 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_604a8b79-fcf3-40cf-82e6-8af2cf667b0f/ovn-northd/0.log" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.450798 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.507495 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.517001 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.525411 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-plugins-conf\") pod \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.525514 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8cqd\" (UniqueName: \"kubernetes.io/projected/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-kube-api-access-f8cqd\") pod \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.525553 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-ovn-rundir\") pod \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.525591 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-ovn-northd-tls-certs\") pod \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.529758 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9hct\" (UniqueName: \"kubernetes.io/projected/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-kube-api-access-d9hct\") pod \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.529806 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-pod-info\") pod \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.529868 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-scripts\") pod \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.529941 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-erlang-cookie\") pod \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.529994 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-erlang-cookie-secret\") pod \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.530089 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-server-conf\") pod \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.530135 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.530204 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-confd\") pod \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.530238 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-plugins\") pod \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.530294 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-config\") pod \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.530747 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-metrics-certs-tls-certs\") pod \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.530786 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-combined-ca-bundle\") pod \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\" (UID: \"604a8b79-fcf3-40cf-82e6-8af2cf667b0f\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.530826 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-config-data\") pod \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.530860 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-tls\") pod \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\" (UID: \"56bbf32f-35b5-4f92-bfc0-9266b0ecccac\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.525963 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "56bbf32f-35b5-4f92-bfc0-9266b0ecccac" (UID: "56bbf32f-35b5-4f92-bfc0-9266b0ecccac"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.528173 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "604a8b79-fcf3-40cf-82e6-8af2cf667b0f" (UID: "604a8b79-fcf3-40cf-82e6-8af2cf667b0f"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.532556 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-kube-api-access-f8cqd" (OuterVolumeSpecName: "kube-api-access-f8cqd") pod "604a8b79-fcf3-40cf-82e6-8af2cf667b0f" (UID: "604a8b79-fcf3-40cf-82e6-8af2cf667b0f"). InnerVolumeSpecName "kube-api-access-f8cqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.533988 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "56bbf32f-35b5-4f92-bfc0-9266b0ecccac" (UID: "56bbf32f-35b5-4f92-bfc0-9266b0ecccac"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.535244 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-kube-api-access-d9hct" (OuterVolumeSpecName: "kube-api-access-d9hct") pod "56bbf32f-35b5-4f92-bfc0-9266b0ecccac" (UID: "56bbf32f-35b5-4f92-bfc0-9266b0ecccac"). InnerVolumeSpecName "kube-api-access-d9hct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.536973 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "56bbf32f-35b5-4f92-bfc0-9266b0ecccac" (UID: "56bbf32f-35b5-4f92-bfc0-9266b0ecccac"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.537777 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "56bbf32f-35b5-4f92-bfc0-9266b0ecccac" (UID: "56bbf32f-35b5-4f92-bfc0-9266b0ecccac"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.538071 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "56bbf32f-35b5-4f92-bfc0-9266b0ecccac" (UID: "56bbf32f-35b5-4f92-bfc0-9266b0ecccac"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.538620 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-scripts" (OuterVolumeSpecName: "scripts") pod "604a8b79-fcf3-40cf-82e6-8af2cf667b0f" (UID: "604a8b79-fcf3-40cf-82e6-8af2cf667b0f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.539037 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-config" (OuterVolumeSpecName: "config") pod "604a8b79-fcf3-40cf-82e6-8af2cf667b0f" (UID: "604a8b79-fcf3-40cf-82e6-8af2cf667b0f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.539546 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-pod-info" (OuterVolumeSpecName: "pod-info") pod "56bbf32f-35b5-4f92-bfc0-9266b0ecccac" (UID: "56bbf32f-35b5-4f92-bfc0-9266b0ecccac"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.543344 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "56bbf32f-35b5-4f92-bfc0-9266b0ecccac" (UID: "56bbf32f-35b5-4f92-bfc0-9266b0ecccac"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.561600 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-config-data" (OuterVolumeSpecName: "config-data") pod "56bbf32f-35b5-4f92-bfc0-9266b0ecccac" (UID: "56bbf32f-35b5-4f92-bfc0-9266b0ecccac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.571411 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "604a8b79-fcf3-40cf-82e6-8af2cf667b0f" (UID: "604a8b79-fcf3-40cf-82e6-8af2cf667b0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.589719 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-server-conf" (OuterVolumeSpecName: "server-conf") pod "56bbf32f-35b5-4f92-bfc0-9266b0ecccac" (UID: "56bbf32f-35b5-4f92-bfc0-9266b0ecccac"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.595846 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "604a8b79-fcf3-40cf-82e6-8af2cf667b0f" (UID: "604a8b79-fcf3-40cf-82e6-8af2cf667b0f"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.613579 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "604a8b79-fcf3-40cf-82e6-8af2cf667b0f" (UID: "604a8b79-fcf3-40cf-82e6-8af2cf667b0f"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.634940 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7kzl\" (UniqueName: \"kubernetes.io/projected/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-kube-api-access-p7kzl\") pod \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.634992 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-tls\") pod \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.635074 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-plugins-conf\") pod \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.635633 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "3b81b6a9-0be6-4ce1-81b4-043edb1c221f" (UID: "3b81b6a9-0be6-4ce1-81b4-043edb1c221f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.635639 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-config-data\") pod \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.635691 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-fernet-keys\") pod \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.635724 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-pod-info\") pod \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.635754 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwvm6\" (UniqueName: \"kubernetes.io/projected/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-kube-api-access-zwvm6\") pod \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.635771 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-confd\") pod \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.635801 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-internal-tls-certs\") pod \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.635820 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-erlang-cookie\") pod \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.635855 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-public-tls-certs\") pod \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.635884 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.635906 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-server-conf\") pod \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.635924 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-credential-keys\") pod \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.635945 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-config-data\") pod \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.635974 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-combined-ca-bundle\") pod \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.635992 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-erlang-cookie-secret\") pod \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636011 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-plugins\") pod \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\" (UID: \"3b81b6a9-0be6-4ce1-81b4-043edb1c221f\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636048 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-scripts\") pod \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\" (UID: \"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636405 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636416 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-config\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636426 4810 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636436 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636444 4810 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636452 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636461 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636483 4810 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636492 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8cqd\" (UniqueName: \"kubernetes.io/projected/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-kube-api-access-f8cqd\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636500 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636508 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636517 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9hct\" (UniqueName: \"kubernetes.io/projected/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-kube-api-access-d9hct\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636526 4810 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-pod-info\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636534 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/604a8b79-fcf3-40cf-82e6-8af2cf667b0f-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636544 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636553 4810 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636561 4810 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-server-conf\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.636580 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.638089 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "3b81b6a9-0be6-4ce1-81b4-043edb1c221f" (UID: "3b81b6a9-0be6-4ce1-81b4-043edb1c221f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.639377 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-kube-api-access-p7kzl" (OuterVolumeSpecName: "kube-api-access-p7kzl") pod "5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0" (UID: "5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0"). InnerVolumeSpecName "kube-api-access-p7kzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.640981 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "3b81b6a9-0be6-4ce1-81b4-043edb1c221f" (UID: "3b81b6a9-0be6-4ce1-81b4-043edb1c221f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.641560 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "3b81b6a9-0be6-4ce1-81b4-043edb1c221f" (UID: "3b81b6a9-0be6-4ce1-81b4-043edb1c221f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.641967 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0" (UID: "5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.642063 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-pod-info" (OuterVolumeSpecName: "pod-info") pod "3b81b6a9-0be6-4ce1-81b4-043edb1c221f" (UID: "3b81b6a9-0be6-4ce1-81b4-043edb1c221f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.642106 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0" (UID: "5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.642561 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "3b81b6a9-0be6-4ce1-81b4-043edb1c221f" (UID: "3b81b6a9-0be6-4ce1-81b4-043edb1c221f"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.643223 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-scripts" (OuterVolumeSpecName: "scripts") pod "5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0" (UID: "5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.643884 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "3b81b6a9-0be6-4ce1-81b4-043edb1c221f" (UID: "3b81b6a9-0be6-4ce1-81b4-043edb1c221f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.646528 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "56bbf32f-35b5-4f92-bfc0-9266b0ecccac" (UID: "56bbf32f-35b5-4f92-bfc0-9266b0ecccac"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.648240 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-kube-api-access-zwvm6" (OuterVolumeSpecName: "kube-api-access-zwvm6") pod "3b81b6a9-0be6-4ce1-81b4-043edb1c221f" (UID: "3b81b6a9-0be6-4ce1-81b4-043edb1c221f"). InnerVolumeSpecName "kube-api-access-zwvm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.662131 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.662842 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-config-data" (OuterVolumeSpecName: "config-data") pod "3b81b6a9-0be6-4ce1-81b4-043edb1c221f" (UID: "3b81b6a9-0be6-4ce1-81b4-043edb1c221f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.666101 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-config-data" (OuterVolumeSpecName: "config-data") pod "5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0" (UID: "5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.685946 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0" (UID: "5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.693401 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0" (UID: "5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.695129 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0" (UID: "5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.695134 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-server-conf" (OuterVolumeSpecName: "server-conf") pod "3b81b6a9-0be6-4ce1-81b4-043edb1c221f" (UID: "3b81b6a9-0be6-4ce1-81b4-043edb1c221f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.710552 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "3b81b6a9-0be6-4ce1-81b4-043edb1c221f" (UID: "3b81b6a9-0be6-4ce1-81b4-043edb1c221f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.712327 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"56bbf32f-35b5-4f92-bfc0-9266b0ecccac","Type":"ContainerDied","Data":"93aeae8258afd7bcc2b35dc20a4f79760421b80ce5ada14ec46ca0fa7970ab90"} Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.712351 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.712377 4810 scope.go:117] "RemoveContainer" containerID="82a26c84ad224e60a4fa117cf5e1d03643cfd46868d1eeb656fac71a268ad1e5" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.715380 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_604a8b79-fcf3-40cf-82e6-8af2cf667b0f/ovn-northd/0.log" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.715578 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"604a8b79-fcf3-40cf-82e6-8af2cf667b0f","Type":"ContainerDied","Data":"1610573c62e58cd66296d7120033026e8ac8365ced8ff8d8bc935cffdb387816"} Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.715610 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.725752 4810 generic.go:334] "Generic (PLEG): container finished" podID="bef8b0c7-66cd-4930-a78b-bf73804a73d2" containerID="d01cd11ccf42943023a23e361f67460aa3a8387758c28bbf2c4f4a6dba26d397" exitCode=0 Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.725823 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bef8b0c7-66cd-4930-a78b-bf73804a73d2","Type":"ContainerDied","Data":"d01cd11ccf42943023a23e361f67460aa3a8387758c28bbf2c4f4a6dba26d397"} Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.729232 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-65bbfc7449-rtqm6" event={"ID":"5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0","Type":"ContainerDied","Data":"0f78ed302304f1a4ae276cf04f45b339eb2e16c4d7fe8c01f8a3c129526d347d"} Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.729307 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-65bbfc7449-rtqm6" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.733006 4810 generic.go:334] "Generic (PLEG): container finished" podID="20964a12-27b5-46f7-a9b9-e68b8c7dfce2" containerID="a40c19049624e61426f700a570719d59ec7a3b73107e4fbc2d3c42482e3793b0" exitCode=0 Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.733072 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" event={"ID":"20964a12-27b5-46f7-a9b9-e68b8c7dfce2","Type":"ContainerDied","Data":"a40c19049624e61426f700a570719d59ec7a3b73107e4fbc2d3c42482e3793b0"} Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.733098 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" event={"ID":"20964a12-27b5-46f7-a9b9-e68b8c7dfce2","Type":"ContainerDied","Data":"7f009842a256ec9e51a6a68edbc02dff2e8fa77481ed53d9a21a09772491ce71"} Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.733112 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f009842a256ec9e51a6a68edbc02dff2e8fa77481ed53d9a21a09772491ce71" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.734892 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3b81b6a9-0be6-4ce1-81b4-043edb1c221f","Type":"ContainerDied","Data":"e48c8a12d183e258596aa4c10eacf8ff2cbe94f89874e3e923e2c7d69077f67d"} Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.734946 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.739306 4810 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.739330 4810 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-pod-info\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.739340 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwvm6\" (UniqueName: \"kubernetes.io/projected/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-kube-api-access-zwvm6\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.739351 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.739359 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.739367 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.739375 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.739409 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.739419 4810 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-server-conf\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.739427 4810 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.739435 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.739443 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.746583 4810 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.746623 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.746634 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.746645 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/56bbf32f-35b5-4f92-bfc0-9266b0ecccac-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.746656 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.746665 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7kzl\" (UniqueName: \"kubernetes.io/projected/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0-kube-api-access-p7kzl\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.746675 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.746688 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b81b6a9-0be6-4ce1-81b4-043edb1c221f-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.773389 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.815953 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.821308 4810 scope.go:117] "RemoveContainer" containerID="e23662080306dd3c3cc28ceb3e4c2195c124ff9bccb4a3034f0a68f6534ad3d3" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.849483 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: E1201 14:59:53.849580 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:53 crc kubenswrapper[4810]: E1201 14:59:53.849640 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts podName:1ef9fa82-f302-47e9-bed3-b02f34a2340f nodeName:}" failed. No retries permitted until 2025-12-01 14:59:57.849619215 +0000 UTC m=+1563.613128838 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts") pod "barbicanff71-account-delete-2fnl8" (UID: "1ef9fa82-f302-47e9-bed3-b02f34a2340f") : configmap "openstack-scripts" not found Dec 01 14:59:53 crc kubenswrapper[4810]: E1201 14:59:53.849986 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:53 crc kubenswrapper[4810]: E1201 14:59:53.850028 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts podName:4f7b5a25-eed0-4122-a036-093ac7adca22 nodeName:}" failed. No retries permitted until 2025-12-01 14:59:57.850017976 +0000 UTC m=+1563.613527579 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts") pod "novacell096e4-account-delete-qqlg4" (UID: "4f7b5a25-eed0-4122-a036-093ac7adca22") : configmap "openstack-scripts" not found Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.868645 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.879490 4810 scope.go:117] "RemoveContainer" containerID="4403ed72fa83e2083ad29edab75aeefc505950c9b561718246753461398f9c33" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.879624 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.889524 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.906709 4810 scope.go:117] "RemoveContainer" containerID="87439695570ef9b3e9da98681ae2d1dec7fb73d967ccb8912a2e40ca13529cf7" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.907533 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.917925 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-65bbfc7449-rtqm6"] Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.923512 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-65bbfc7449-rtqm6"] Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.934184 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.938419 4810 scope.go:117] "RemoveContainer" containerID="e064be1964d87f2e16398d0a402ce3f57b2caf18c89d6b179f3ef2a94e1374ac" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.940302 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.950383 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-config-data-custom\") pod \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.950663 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-combined-ca-bundle\") pod \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.950745 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9xp8\" (UniqueName: \"kubernetes.io/projected/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-kube-api-access-d9xp8\") pod \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.950780 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-logs\") pod \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.950814 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-config-data\") pod \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\" (UID: \"20964a12-27b5-46f7-a9b9-e68b8c7dfce2\") " Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.951509 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-logs" (OuterVolumeSpecName: "logs") pod "20964a12-27b5-46f7-a9b9-e68b8c7dfce2" (UID: "20964a12-27b5-46f7-a9b9-e68b8c7dfce2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.951871 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-logs\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.954518 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "20964a12-27b5-46f7-a9b9-e68b8c7dfce2" (UID: "20964a12-27b5-46f7-a9b9-e68b8c7dfce2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.955146 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-kube-api-access-d9xp8" (OuterVolumeSpecName: "kube-api-access-d9xp8") pod "20964a12-27b5-46f7-a9b9-e68b8c7dfce2" (UID: "20964a12-27b5-46f7-a9b9-e68b8c7dfce2"). InnerVolumeSpecName "kube-api-access-d9xp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.973700 4810 scope.go:117] "RemoveContainer" containerID="74e6d90ac1ab1306dd26c5e8c6cd4cd8b45f47eb0ab68865c9c1575f423fce76" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.983087 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20964a12-27b5-46f7-a9b9-e68b8c7dfce2" (UID: "20964a12-27b5-46f7-a9b9-e68b8c7dfce2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:53 crc kubenswrapper[4810]: I1201 14:59:53.996241 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-config-data" (OuterVolumeSpecName: "config-data") pod "20964a12-27b5-46f7-a9b9-e68b8c7dfce2" (UID: "20964a12-27b5-46f7-a9b9-e68b8c7dfce2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.023789 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.033705 4810 scope.go:117] "RemoveContainer" containerID="823b399b01a75b82b70c3653a5f597fb68c1ca3af1fbf06c04a2d840c74494cb" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.054264 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.054289 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.054299 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9xp8\" (UniqueName: \"kubernetes.io/projected/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-kube-api-access-d9xp8\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.054308 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20964a12-27b5-46f7-a9b9-e68b8c7dfce2-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.155381 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bef8b0c7-66cd-4930-a78b-bf73804a73d2-etc-machine-id\") pod \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.155488 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-config-data-custom\") pod \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.155518 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-combined-ca-bundle\") pod \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.155543 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-config-data\") pod \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.155652 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-scripts\") pod \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.155691 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v485w\" (UniqueName: \"kubernetes.io/projected/bef8b0c7-66cd-4930-a78b-bf73804a73d2-kube-api-access-v485w\") pod \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\" (UID: \"bef8b0c7-66cd-4930-a78b-bf73804a73d2\") " Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.155965 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bef8b0c7-66cd-4930-a78b-bf73804a73d2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bef8b0c7-66cd-4930-a78b-bf73804a73d2" (UID: "bef8b0c7-66cd-4930-a78b-bf73804a73d2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.156657 4810 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bef8b0c7-66cd-4930-a78b-bf73804a73d2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.159257 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-scripts" (OuterVolumeSpecName: "scripts") pod "bef8b0c7-66cd-4930-a78b-bf73804a73d2" (UID: "bef8b0c7-66cd-4930-a78b-bf73804a73d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.160295 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bef8b0c7-66cd-4930-a78b-bf73804a73d2" (UID: "bef8b0c7-66cd-4930-a78b-bf73804a73d2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.160527 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bef8b0c7-66cd-4930-a78b-bf73804a73d2-kube-api-access-v485w" (OuterVolumeSpecName: "kube-api-access-v485w") pod "bef8b0c7-66cd-4930-a78b-bf73804a73d2" (UID: "bef8b0c7-66cd-4930-a78b-bf73804a73d2"). InnerVolumeSpecName "kube-api-access-v485w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.194161 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bef8b0c7-66cd-4930-a78b-bf73804a73d2" (UID: "bef8b0c7-66cd-4930-a78b-bf73804a73d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.230122 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-config-data" (OuterVolumeSpecName: "config-data") pod "bef8b0c7-66cd-4930-a78b-bf73804a73d2" (UID: "bef8b0c7-66cd-4930-a78b-bf73804a73d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.258450 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.258518 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v485w\" (UniqueName: \"kubernetes.io/projected/bef8b0c7-66cd-4930-a78b-bf73804a73d2-kube-api-access-v485w\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.258539 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.258554 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.258599 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bef8b0c7-66cd-4930-a78b-bf73804a73d2-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.537562 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b81b6a9-0be6-4ce1-81b4-043edb1c221f" path="/var/lib/kubelet/pods/3b81b6a9-0be6-4ce1-81b4-043edb1c221f/volumes" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.538690 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c309abf-824f-4264-859b-ded1f0a9acaf" path="/var/lib/kubelet/pods/4c309abf-824f-4264-859b-ded1f0a9acaf/volumes" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.541680 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56bbf32f-35b5-4f92-bfc0-9266b0ecccac" path="/var/lib/kubelet/pods/56bbf32f-35b5-4f92-bfc0-9266b0ecccac/volumes" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.542566 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0" path="/var/lib/kubelet/pods/5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0/volumes" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.543154 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="604a8b79-fcf3-40cf-82e6-8af2cf667b0f" path="/var/lib/kubelet/pods/604a8b79-fcf3-40cf-82e6-8af2cf667b0f/volumes" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.791164 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.791437 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bef8b0c7-66cd-4930-a78b-bf73804a73d2","Type":"ContainerDied","Data":"f7dff30ba20cc0a4c5812b5cf1ac8ac65e4d45ec8049c46527f8c1a87513a936"} Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.791501 4810 scope.go:117] "RemoveContainer" containerID="c3e8f637d25ffa49a1106b343b7c3861c122e5ccdeb4f52b034a012b1bf12322" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.799129 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6d8fb448c4-2hg9g" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.821932 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.827912 4810 scope.go:117] "RemoveContainer" containerID="d01cd11ccf42943023a23e361f67460aa3a8387758c28bbf2c4f4a6dba26d397" Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.833974 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.840807 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-6d8fb448c4-2hg9g"] Dec 01 14:59:54 crc kubenswrapper[4810]: I1201 14:59:54.857931 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-6d8fb448c4-2hg9g"] Dec 01 14:59:56 crc kubenswrapper[4810]: I1201 14:59:56.501234 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20964a12-27b5-46f7-a9b9-e68b8c7dfce2" path="/var/lib/kubelet/pods/20964a12-27b5-46f7-a9b9-e68b8c7dfce2/volumes" Dec 01 14:59:56 crc kubenswrapper[4810]: I1201 14:59:56.502108 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bef8b0c7-66cd-4930-a78b-bf73804a73d2" path="/var/lib/kubelet/pods/bef8b0c7-66cd-4930-a78b-bf73804a73d2/volumes" Dec 01 14:59:56 crc kubenswrapper[4810]: E1201 14:59:56.815209 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:56 crc kubenswrapper[4810]: E1201 14:59:56.815304 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts podName:476b71ac-1050-46a6-b51c-168e75b1a350 nodeName:}" failed. No retries permitted until 2025-12-01 15:00:04.815287916 +0000 UTC m=+1570.578797529 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts") pod "novaapic0fd-account-delete-w7ttc" (UID: "476b71ac-1050-46a6-b51c-168e75b1a350") : configmap "openstack-scripts" not found Dec 01 14:59:56 crc kubenswrapper[4810]: E1201 14:59:56.916527 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:56 crc kubenswrapper[4810]: E1201 14:59:56.916613 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts podName:bf15ccdf-1e4e-44ab-8ccb-819502935a7a nodeName:}" failed. No retries permitted until 2025-12-01 15:00:04.916594189 +0000 UTC m=+1570.680103792 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts") pod "cinder17c6-account-delete-qktk6" (UID: "bf15ccdf-1e4e-44ab-8ccb-819502935a7a") : configmap "openstack-scripts" not found Dec 01 14:59:56 crc kubenswrapper[4810]: E1201 14:59:56.916616 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:56 crc kubenswrapper[4810]: E1201 14:59:56.916727 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts podName:f7b06553-ec05-4c6f-82c0-ef0b4dd3005a nodeName:}" failed. No retries permitted until 2025-12-01 15:00:04.916704093 +0000 UTC m=+1570.680213706 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts") pod "neutronc91e-account-delete-t585q" (UID: "f7b06553-ec05-4c6f-82c0-ef0b4dd3005a") : configmap "openstack-scripts" not found Dec 01 14:59:57 crc kubenswrapper[4810]: E1201 14:59:57.319991 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 01 14:59:57 crc kubenswrapper[4810]: E1201 14:59:57.320456 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 01 14:59:57 crc kubenswrapper[4810]: E1201 14:59:57.321087 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 01 14:59:57 crc kubenswrapper[4810]: E1201 14:59:57.321193 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 01 14:59:57 crc kubenswrapper[4810]: E1201 14:59:57.321238 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-x82dv" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovsdb-server" Dec 01 14:59:57 crc kubenswrapper[4810]: E1201 14:59:57.323656 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 01 14:59:57 crc kubenswrapper[4810]: E1201 14:59:57.328605 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 01 14:59:57 crc kubenswrapper[4810]: E1201 14:59:57.328654 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-x82dv" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovs-vswitchd" Dec 01 14:59:57 crc kubenswrapper[4810]: E1201 14:59:57.930332 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:57 crc kubenswrapper[4810]: E1201 14:59:57.930430 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts podName:1ef9fa82-f302-47e9-bed3-b02f34a2340f nodeName:}" failed. No retries permitted until 2025-12-01 15:00:05.930408356 +0000 UTC m=+1571.693917989 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts") pod "barbicanff71-account-delete-2fnl8" (UID: "1ef9fa82-f302-47e9-bed3-b02f34a2340f") : configmap "openstack-scripts" not found Dec 01 14:59:57 crc kubenswrapper[4810]: E1201 14:59:57.930333 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 14:59:57 crc kubenswrapper[4810]: E1201 14:59:57.930544 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts podName:4f7b5a25-eed0-4122-a036-093ac7adca22 nodeName:}" failed. No retries permitted until 2025-12-01 15:00:05.930529079 +0000 UTC m=+1571.694038682 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts") pod "novacell096e4-account-delete-qqlg4" (UID: "4f7b5a25-eed0-4122-a036-093ac7adca22") : configmap "openstack-scripts" not found Dec 01 14:59:58 crc kubenswrapper[4810]: I1201 14:59:58.220189 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="56bbf32f-35b5-4f92-bfc0-9266b0ecccac" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: i/o timeout" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134263 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r"] Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134583 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="129ae854-1430-465f-a177-0c459bd8b4c8" containerName="barbican-worker-log" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134594 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="129ae854-1430-465f-a177-0c459bd8b4c8" containerName="barbican-worker-log" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134610 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bef8b0c7-66cd-4930-a78b-bf73804a73d2" containerName="probe" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134615 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bef8b0c7-66cd-4930-a78b-bf73804a73d2" containerName="probe" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134631 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56bbf32f-35b5-4f92-bfc0-9266b0ecccac" containerName="rabbitmq" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134637 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="56bbf32f-35b5-4f92-bfc0-9266b0ecccac" containerName="rabbitmq" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134646 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c73a61da-0c40-4a26-888c-ab5b994417d4" containerName="nova-cell0-conductor-conductor" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134652 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c73a61da-0c40-4a26-888c-ab5b994417d4" containerName="nova-cell0-conductor-conductor" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134661 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c" containerName="nova-cell1-conductor-conductor" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134666 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c" containerName="nova-cell1-conductor-conductor" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134675 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31b83b20-83c4-4e06-8eb0-1e1d83d4b160" containerName="placement-log" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134681 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="31b83b20-83c4-4e06-8eb0-1e1d83d4b160" containerName="placement-log" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134692 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e79c5698-ab2d-4292-bda9-fab8dbbb2abb" containerName="cinder-api" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134697 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e79c5698-ab2d-4292-bda9-fab8dbbb2abb" containerName="cinder-api" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134709 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b81b6a9-0be6-4ce1-81b4-043edb1c221f" containerName="rabbitmq" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134714 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b81b6a9-0be6-4ce1-81b4-043edb1c221f" containerName="rabbitmq" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134725 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91932480-76aa-4751-9251-3ad82d849882" containerName="mariadb-account-delete" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134730 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="91932480-76aa-4751-9251-3ad82d849882" containerName="mariadb-account-delete" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134737 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a3a8b5d-01e4-4396-8a9b-abd85dc55572" containerName="mariadb-account-delete" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134744 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a3a8b5d-01e4-4396-8a9b-abd85dc55572" containerName="mariadb-account-delete" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134751 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd705f6-0d8f-49df-aff6-813792212999" containerName="glance-log" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134756 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd705f6-0d8f-49df-aff6-813792212999" containerName="glance-log" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134765 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerName="ceilometer-central-agent" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134771 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerName="ceilometer-central-agent" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134777 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56bbf32f-35b5-4f92-bfc0-9266b0ecccac" containerName="setup-container" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134782 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="56bbf32f-35b5-4f92-bfc0-9266b0ecccac" containerName="setup-container" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134796 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9acd8c29-93f5-4b0f-8b32-d8f3d576767b" containerName="nova-api-log" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134801 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9acd8c29-93f5-4b0f-8b32-d8f3d576767b" containerName="nova-api-log" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134808 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="444ec27a-77d0-4b01-b59e-c6590c46a0c2" containerName="nova-scheduler-scheduler" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134813 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="444ec27a-77d0-4b01-b59e-c6590c46a0c2" containerName="nova-scheduler-scheduler" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134824 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0" containerName="keystone-api" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134829 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0" containerName="keystone-api" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134839 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerName="sg-core" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134845 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerName="sg-core" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134854 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c309abf-824f-4264-859b-ded1f0a9acaf" containerName="mysql-bootstrap" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134860 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c309abf-824f-4264-859b-ded1f0a9acaf" containerName="mysql-bootstrap" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134870 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdd90cbb-4242-4227-9de4-09f99e60a516" containerName="barbican-api-log" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134875 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdd90cbb-4242-4227-9de4-09f99e60a516" containerName="barbican-api-log" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134885 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="604a8b79-fcf3-40cf-82e6-8af2cf667b0f" containerName="openstack-network-exporter" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134890 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="604a8b79-fcf3-40cf-82e6-8af2cf667b0f" containerName="openstack-network-exporter" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134902 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" containerName="glance-httpd" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134908 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" containerName="glance-httpd" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134917 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c309abf-824f-4264-859b-ded1f0a9acaf" containerName="galera" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134922 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c309abf-824f-4264-859b-ded1f0a9acaf" containerName="galera" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134930 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" containerName="glance-log" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134936 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" containerName="glance-log" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134943 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39ce8411-05ee-4022-bd7a-87e46c2cee8d" containerName="kube-state-metrics" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134950 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="39ce8411-05ee-4022-bd7a-87e46c2cee8d" containerName="kube-state-metrics" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134957 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerName="ceilometer-notification-agent" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134963 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerName="ceilometer-notification-agent" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134972 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31b83b20-83c4-4e06-8eb0-1e1d83d4b160" containerName="placement-api" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134978 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="31b83b20-83c4-4e06-8eb0-1e1d83d4b160" containerName="placement-api" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134983 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerName="proxy-httpd" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.134988 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerName="proxy-httpd" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.134995 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e79c5698-ab2d-4292-bda9-fab8dbbb2abb" containerName="cinder-api-log" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135001 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e79c5698-ab2d-4292-bda9-fab8dbbb2abb" containerName="cinder-api-log" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.135010 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9acd8c29-93f5-4b0f-8b32-d8f3d576767b" containerName="nova-api-api" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135015 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9acd8c29-93f5-4b0f-8b32-d8f3d576767b" containerName="nova-api-api" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.135025 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20964a12-27b5-46f7-a9b9-e68b8c7dfce2" containerName="barbican-keystone-listener-log" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135030 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="20964a12-27b5-46f7-a9b9-e68b8c7dfce2" containerName="barbican-keystone-listener-log" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.135042 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd705f6-0d8f-49df-aff6-813792212999" containerName="glance-httpd" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135048 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd705f6-0d8f-49df-aff6-813792212999" containerName="glance-httpd" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.135059 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" containerName="nova-metadata-log" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135064 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" containerName="nova-metadata-log" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.135073 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdd90cbb-4242-4227-9de4-09f99e60a516" containerName="barbican-api" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135078 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdd90cbb-4242-4227-9de4-09f99e60a516" containerName="barbican-api" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.135087 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="129ae854-1430-465f-a177-0c459bd8b4c8" containerName="barbican-worker" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135094 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="129ae854-1430-465f-a177-0c459bd8b4c8" containerName="barbican-worker" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.135106 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20964a12-27b5-46f7-a9b9-e68b8c7dfce2" containerName="barbican-keystone-listener" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135114 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="20964a12-27b5-46f7-a9b9-e68b8c7dfce2" containerName="barbican-keystone-listener" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.135127 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b81b6a9-0be6-4ce1-81b4-043edb1c221f" containerName="setup-container" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135136 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b81b6a9-0be6-4ce1-81b4-043edb1c221f" containerName="setup-container" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.135148 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e16305af-6573-4cfd-ae76-f596d410920b" containerName="memcached" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135158 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e16305af-6573-4cfd-ae76-f596d410920b" containerName="memcached" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.135169 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bef8b0c7-66cd-4930-a78b-bf73804a73d2" containerName="cinder-scheduler" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135177 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bef8b0c7-66cd-4930-a78b-bf73804a73d2" containerName="cinder-scheduler" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.135189 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="604a8b79-fcf3-40cf-82e6-8af2cf667b0f" containerName="ovn-northd" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135196 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="604a8b79-fcf3-40cf-82e6-8af2cf667b0f" containerName="ovn-northd" Dec 01 15:00:00 crc kubenswrapper[4810]: E1201 15:00:00.135205 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" containerName="nova-metadata-metadata" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135212 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" containerName="nova-metadata-metadata" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135373 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bef8b0c7-66cd-4930-a78b-bf73804a73d2" containerName="cinder-scheduler" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135384 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="91932480-76aa-4751-9251-3ad82d849882" containerName="mariadb-account-delete" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135395 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bef8b0c7-66cd-4930-a78b-bf73804a73d2" containerName="probe" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135409 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="604a8b79-fcf3-40cf-82e6-8af2cf667b0f" containerName="openstack-network-exporter" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135417 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="56bbf32f-35b5-4f92-bfc0-9266b0ecccac" containerName="rabbitmq" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135426 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="20964a12-27b5-46f7-a9b9-e68b8c7dfce2" containerName="barbican-keystone-listener" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135436 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dd705f6-0d8f-49df-aff6-813792212999" containerName="glance-log" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135443 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9acd8c29-93f5-4b0f-8b32-d8f3d576767b" containerName="nova-api-log" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135449 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="992a1c9d-4ccb-4aa3-9cf4-aaed69d4cd7c" containerName="nova-cell1-conductor-conductor" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135456 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="444ec27a-77d0-4b01-b59e-c6590c46a0c2" containerName="nova-scheduler-scheduler" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135463 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerName="proxy-httpd" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135506 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="604a8b79-fcf3-40cf-82e6-8af2cf667b0f" containerName="ovn-northd" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135513 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdd90cbb-4242-4227-9de4-09f99e60a516" containerName="barbican-api-log" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135521 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e16305af-6573-4cfd-ae76-f596d410920b" containerName="memcached" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135530 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" containerName="nova-metadata-metadata" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135539 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d0ef8e6-d5f3-491f-9d3d-ed18954dc3f0" containerName="keystone-api" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135547 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerName="ceilometer-central-agent" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135557 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e79c5698-ab2d-4292-bda9-fab8dbbb2abb" containerName="cinder-api" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135565 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b81b6a9-0be6-4ce1-81b4-043edb1c221f" containerName="rabbitmq" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135576 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dd705f6-0d8f-49df-aff6-813792212999" containerName="glance-httpd" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135585 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c73a61da-0c40-4a26-888c-ab5b994417d4" containerName="nova-cell0-conductor-conductor" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135595 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="129ae854-1430-465f-a177-0c459bd8b4c8" containerName="barbican-worker" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135605 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="39ce8411-05ee-4022-bd7a-87e46c2cee8d" containerName="kube-state-metrics" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135615 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ae9e416-c20a-47d7-a8fe-b3cdd0e1a2e6" containerName="nova-metadata-log" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135621 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerName="sg-core" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135627 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="129ae854-1430-465f-a177-0c459bd8b4c8" containerName="barbican-worker-log" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135636 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c32cbe3a-7d7b-44b9-a3a1-945dfaf765b3" containerName="ceilometer-notification-agent" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135643 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="31b83b20-83c4-4e06-8eb0-1e1d83d4b160" containerName="placement-log" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135653 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdd90cbb-4242-4227-9de4-09f99e60a516" containerName="barbican-api" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135659 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e79c5698-ab2d-4292-bda9-fab8dbbb2abb" containerName="cinder-api-log" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135666 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" containerName="glance-httpd" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135675 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="20964a12-27b5-46f7-a9b9-e68b8c7dfce2" containerName="barbican-keystone-listener-log" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135682 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1d037bd-2eb1-4894-ac87-1a8ac1a1aabc" containerName="glance-log" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135690 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="31b83b20-83c4-4e06-8eb0-1e1d83d4b160" containerName="placement-api" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135700 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c309abf-824f-4264-859b-ded1f0a9acaf" containerName="galera" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135709 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9acd8c29-93f5-4b0f-8b32-d8f3d576767b" containerName="nova-api-api" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.135717 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a3a8b5d-01e4-4396-8a9b-abd85dc55572" containerName="mariadb-account-delete" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.136171 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.139041 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.139263 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.184105 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r"] Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.278939 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79dmh\" (UniqueName: \"kubernetes.io/projected/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25-kube-api-access-79dmh\") pod \"collect-profiles-29410020-zvw8r\" (UID: \"ec7ca8a4-04cb-4b40-be20-6f2d872f8b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.279066 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25-config-volume\") pod \"collect-profiles-29410020-zvw8r\" (UID: \"ec7ca8a4-04cb-4b40-be20-6f2d872f8b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.279199 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25-secret-volume\") pod \"collect-profiles-29410020-zvw8r\" (UID: \"ec7ca8a4-04cb-4b40-be20-6f2d872f8b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.380853 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79dmh\" (UniqueName: \"kubernetes.io/projected/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25-kube-api-access-79dmh\") pod \"collect-profiles-29410020-zvw8r\" (UID: \"ec7ca8a4-04cb-4b40-be20-6f2d872f8b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.380945 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25-config-volume\") pod \"collect-profiles-29410020-zvw8r\" (UID: \"ec7ca8a4-04cb-4b40-be20-6f2d872f8b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.381028 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25-secret-volume\") pod \"collect-profiles-29410020-zvw8r\" (UID: \"ec7ca8a4-04cb-4b40-be20-6f2d872f8b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.381951 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25-config-volume\") pod \"collect-profiles-29410020-zvw8r\" (UID: \"ec7ca8a4-04cb-4b40-be20-6f2d872f8b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.398551 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25-secret-volume\") pod \"collect-profiles-29410020-zvw8r\" (UID: \"ec7ca8a4-04cb-4b40-be20-6f2d872f8b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.414277 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79dmh\" (UniqueName: \"kubernetes.io/projected/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25-kube-api-access-79dmh\") pod \"collect-profiles-29410020-zvw8r\" (UID: \"ec7ca8a4-04cb-4b40-be20-6f2d872f8b25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.460375 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r" Dec 01 15:00:00 crc kubenswrapper[4810]: I1201 15:00:00.917646 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r"] Dec 01 15:00:01 crc kubenswrapper[4810]: E1201 15:00:01.407364 4810 secret.go:188] Couldn't get secret openstack/neutron-config: secret "neutron-config" not found Dec 01 15:00:01 crc kubenswrapper[4810]: E1201 15:00:01.407456 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-config podName:ba5f8a69-22cf-4196-bc89-ffa60eda9303 nodeName:}" failed. No retries permitted until 2025-12-01 15:00:17.407435042 +0000 UTC m=+1583.170944655 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-config") pod "neutron-5477779d95-wg2vp" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303") : secret "neutron-config" not found Dec 01 15:00:01 crc kubenswrapper[4810]: E1201 15:00:01.407574 4810 secret.go:188] Couldn't get secret openstack/neutron-httpd-config: secret "neutron-httpd-config" not found Dec 01 15:00:01 crc kubenswrapper[4810]: E1201 15:00:01.407677 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-httpd-config podName:ba5f8a69-22cf-4196-bc89-ffa60eda9303 nodeName:}" failed. No retries permitted until 2025-12-01 15:00:17.407655798 +0000 UTC m=+1583.171165421 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "httpd-config" (UniqueName: "kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-httpd-config") pod "neutron-5477779d95-wg2vp" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303") : secret "neutron-httpd-config" not found Dec 01 15:00:01 crc kubenswrapper[4810]: I1201 15:00:01.875687 4810 generic.go:334] "Generic (PLEG): container finished" podID="ec7ca8a4-04cb-4b40-be20-6f2d872f8b25" containerID="16c524e58e80b9b3ae10fd0e7efafe0e4568dc81b9a83ba564662a50e1693504" exitCode=0 Dec 01 15:00:01 crc kubenswrapper[4810]: I1201 15:00:01.875880 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r" event={"ID":"ec7ca8a4-04cb-4b40-be20-6f2d872f8b25","Type":"ContainerDied","Data":"16c524e58e80b9b3ae10fd0e7efafe0e4568dc81b9a83ba564662a50e1693504"} Dec 01 15:00:01 crc kubenswrapper[4810]: I1201 15:00:01.875961 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r" event={"ID":"ec7ca8a4-04cb-4b40-be20-6f2d872f8b25","Type":"ContainerStarted","Data":"e419cd05c8b54cf652b6c441437be68e0b23ba401f763d5437239b0eb69324fe"} Dec 01 15:00:02 crc kubenswrapper[4810]: E1201 15:00:02.320407 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 01 15:00:02 crc kubenswrapper[4810]: E1201 15:00:02.320723 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 01 15:00:02 crc kubenswrapper[4810]: E1201 15:00:02.321147 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 01 15:00:02 crc kubenswrapper[4810]: E1201 15:00:02.321211 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-x82dv" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovsdb-server" Dec 01 15:00:02 crc kubenswrapper[4810]: E1201 15:00:02.321827 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 01 15:00:02 crc kubenswrapper[4810]: E1201 15:00:02.325120 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 01 15:00:02 crc kubenswrapper[4810]: E1201 15:00:02.328235 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 01 15:00:02 crc kubenswrapper[4810]: E1201 15:00:02.328291 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-x82dv" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovs-vswitchd" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.266119 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.443104 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25-secret-volume\") pod \"ec7ca8a4-04cb-4b40-be20-6f2d872f8b25\" (UID: \"ec7ca8a4-04cb-4b40-be20-6f2d872f8b25\") " Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.443164 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25-config-volume\") pod \"ec7ca8a4-04cb-4b40-be20-6f2d872f8b25\" (UID: \"ec7ca8a4-04cb-4b40-be20-6f2d872f8b25\") " Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.443195 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79dmh\" (UniqueName: \"kubernetes.io/projected/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25-kube-api-access-79dmh\") pod \"ec7ca8a4-04cb-4b40-be20-6f2d872f8b25\" (UID: \"ec7ca8a4-04cb-4b40-be20-6f2d872f8b25\") " Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.444183 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25-config-volume" (OuterVolumeSpecName: "config-volume") pod "ec7ca8a4-04cb-4b40-be20-6f2d872f8b25" (UID: "ec7ca8a4-04cb-4b40-be20-6f2d872f8b25"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.449562 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25-kube-api-access-79dmh" (OuterVolumeSpecName: "kube-api-access-79dmh") pod "ec7ca8a4-04cb-4b40-be20-6f2d872f8b25" (UID: "ec7ca8a4-04cb-4b40-be20-6f2d872f8b25"). InnerVolumeSpecName "kube-api-access-79dmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.452700 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ec7ca8a4-04cb-4b40-be20-6f2d872f8b25" (UID: "ec7ca8a4-04cb-4b40-be20-6f2d872f8b25"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.524174 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5477779d95-wg2vp" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.545353 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.545559 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.545617 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79dmh\" (UniqueName: \"kubernetes.io/projected/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25-kube-api-access-79dmh\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.646716 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-config\") pod \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.646771 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mrmd\" (UniqueName: \"kubernetes.io/projected/ba5f8a69-22cf-4196-bc89-ffa60eda9303-kube-api-access-6mrmd\") pod \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.646795 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-ovndb-tls-certs\") pod \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.647426 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-combined-ca-bundle\") pod \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.647453 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-httpd-config\") pod \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.647500 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-public-tls-certs\") pod \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.647911 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-internal-tls-certs\") pod \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\" (UID: \"ba5f8a69-22cf-4196-bc89-ffa60eda9303\") " Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.650111 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba5f8a69-22cf-4196-bc89-ffa60eda9303-kube-api-access-6mrmd" (OuterVolumeSpecName: "kube-api-access-6mrmd") pod "ba5f8a69-22cf-4196-bc89-ffa60eda9303" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303"). InnerVolumeSpecName "kube-api-access-6mrmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.651202 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "ba5f8a69-22cf-4196-bc89-ffa60eda9303" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.678485 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-config" (OuterVolumeSpecName: "config") pod "ba5f8a69-22cf-4196-bc89-ffa60eda9303" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.681613 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ba5f8a69-22cf-4196-bc89-ffa60eda9303" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.683077 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ba5f8a69-22cf-4196-bc89-ffa60eda9303" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.684868 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ba5f8a69-22cf-4196-bc89-ffa60eda9303" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.696466 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "ba5f8a69-22cf-4196-bc89-ffa60eda9303" (UID: "ba5f8a69-22cf-4196-bc89-ffa60eda9303"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.749760 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.749798 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.749813 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mrmd\" (UniqueName: \"kubernetes.io/projected/ba5f8a69-22cf-4196-bc89-ffa60eda9303-kube-api-access-6mrmd\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.749827 4810 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.749839 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.749851 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.749862 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba5f8a69-22cf-4196-bc89-ffa60eda9303-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.896714 4810 generic.go:334] "Generic (PLEG): container finished" podID="ba5f8a69-22cf-4196-bc89-ffa60eda9303" containerID="7b6de236d9a353033c45404d0d0dce626c48768c8ec55f0ea9bf63e5d411c381" exitCode=0 Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.896779 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5477779d95-wg2vp" event={"ID":"ba5f8a69-22cf-4196-bc89-ffa60eda9303","Type":"ContainerDied","Data":"7b6de236d9a353033c45404d0d0dce626c48768c8ec55f0ea9bf63e5d411c381"} Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.896807 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5477779d95-wg2vp" event={"ID":"ba5f8a69-22cf-4196-bc89-ffa60eda9303","Type":"ContainerDied","Data":"8c05d4229dd218e5bf9e4223256a65e9728e52b9a83b569c8efa023476792fe0"} Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.896825 4810 scope.go:117] "RemoveContainer" containerID="8439e2334c700fa63f3562d0a3892b76d49cb65b3961834e76812a6b483afe47" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.896938 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5477779d95-wg2vp" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.900119 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r" event={"ID":"ec7ca8a4-04cb-4b40-be20-6f2d872f8b25","Type":"ContainerDied","Data":"e419cd05c8b54cf652b6c441437be68e0b23ba401f763d5437239b0eb69324fe"} Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.900165 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e419cd05c8b54cf652b6c441437be68e0b23ba401f763d5437239b0eb69324fe" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.900229 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.935356 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5477779d95-wg2vp"] Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.943779 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5477779d95-wg2vp"] Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.944735 4810 scope.go:117] "RemoveContainer" containerID="7b6de236d9a353033c45404d0d0dce626c48768c8ec55f0ea9bf63e5d411c381" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.975201 4810 scope.go:117] "RemoveContainer" containerID="8439e2334c700fa63f3562d0a3892b76d49cb65b3961834e76812a6b483afe47" Dec 01 15:00:03 crc kubenswrapper[4810]: E1201 15:00:03.976251 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8439e2334c700fa63f3562d0a3892b76d49cb65b3961834e76812a6b483afe47\": container with ID starting with 8439e2334c700fa63f3562d0a3892b76d49cb65b3961834e76812a6b483afe47 not found: ID does not exist" containerID="8439e2334c700fa63f3562d0a3892b76d49cb65b3961834e76812a6b483afe47" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.976307 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8439e2334c700fa63f3562d0a3892b76d49cb65b3961834e76812a6b483afe47"} err="failed to get container status \"8439e2334c700fa63f3562d0a3892b76d49cb65b3961834e76812a6b483afe47\": rpc error: code = NotFound desc = could not find container \"8439e2334c700fa63f3562d0a3892b76d49cb65b3961834e76812a6b483afe47\": container with ID starting with 8439e2334c700fa63f3562d0a3892b76d49cb65b3961834e76812a6b483afe47 not found: ID does not exist" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.976349 4810 scope.go:117] "RemoveContainer" containerID="7b6de236d9a353033c45404d0d0dce626c48768c8ec55f0ea9bf63e5d411c381" Dec 01 15:00:03 crc kubenswrapper[4810]: E1201 15:00:03.976914 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b6de236d9a353033c45404d0d0dce626c48768c8ec55f0ea9bf63e5d411c381\": container with ID starting with 7b6de236d9a353033c45404d0d0dce626c48768c8ec55f0ea9bf63e5d411c381 not found: ID does not exist" containerID="7b6de236d9a353033c45404d0d0dce626c48768c8ec55f0ea9bf63e5d411c381" Dec 01 15:00:03 crc kubenswrapper[4810]: I1201 15:00:03.976944 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b6de236d9a353033c45404d0d0dce626c48768c8ec55f0ea9bf63e5d411c381"} err="failed to get container status \"7b6de236d9a353033c45404d0d0dce626c48768c8ec55f0ea9bf63e5d411c381\": rpc error: code = NotFound desc = could not find container \"7b6de236d9a353033c45404d0d0dce626c48768c8ec55f0ea9bf63e5d411c381\": container with ID starting with 7b6de236d9a353033c45404d0d0dce626c48768c8ec55f0ea9bf63e5d411c381 not found: ID does not exist" Dec 01 15:00:04 crc kubenswrapper[4810]: I1201 15:00:04.503860 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba5f8a69-22cf-4196-bc89-ffa60eda9303" path="/var/lib/kubelet/pods/ba5f8a69-22cf-4196-bc89-ffa60eda9303/volumes" Dec 01 15:00:04 crc kubenswrapper[4810]: E1201 15:00:04.867848 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 15:00:04 crc kubenswrapper[4810]: E1201 15:00:04.867972 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts podName:476b71ac-1050-46a6-b51c-168e75b1a350 nodeName:}" failed. No retries permitted until 2025-12-01 15:00:20.867942058 +0000 UTC m=+1586.631451881 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts") pod "novaapic0fd-account-delete-w7ttc" (UID: "476b71ac-1050-46a6-b51c-168e75b1a350") : configmap "openstack-scripts" not found Dec 01 15:00:04 crc kubenswrapper[4810]: E1201 15:00:04.969211 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 15:00:04 crc kubenswrapper[4810]: E1201 15:00:04.969291 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 15:00:04 crc kubenswrapper[4810]: E1201 15:00:04.969304 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts podName:bf15ccdf-1e4e-44ab-8ccb-819502935a7a nodeName:}" failed. No retries permitted until 2025-12-01 15:00:20.969279902 +0000 UTC m=+1586.732789585 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts") pod "cinder17c6-account-delete-qktk6" (UID: "bf15ccdf-1e4e-44ab-8ccb-819502935a7a") : configmap "openstack-scripts" not found Dec 01 15:00:04 crc kubenswrapper[4810]: E1201 15:00:04.969413 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts podName:f7b06553-ec05-4c6f-82c0-ef0b4dd3005a nodeName:}" failed. No retries permitted until 2025-12-01 15:00:20.969391585 +0000 UTC m=+1586.732901238 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts") pod "neutronc91e-account-delete-t585q" (UID: "f7b06553-ec05-4c6f-82c0-ef0b4dd3005a") : configmap "openstack-scripts" not found Dec 01 15:00:05 crc kubenswrapper[4810]: E1201 15:00:05.991014 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 15:00:05 crc kubenswrapper[4810]: E1201 15:00:05.991164 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts podName:4f7b5a25-eed0-4122-a036-093ac7adca22 nodeName:}" failed. No retries permitted until 2025-12-01 15:00:21.991110547 +0000 UTC m=+1587.754620190 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts") pod "novacell096e4-account-delete-qqlg4" (UID: "4f7b5a25-eed0-4122-a036-093ac7adca22") : configmap "openstack-scripts" not found Dec 01 15:00:05 crc kubenswrapper[4810]: E1201 15:00:05.991389 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 15:00:05 crc kubenswrapper[4810]: E1201 15:00:05.991666 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts podName:1ef9fa82-f302-47e9-bed3-b02f34a2340f nodeName:}" failed. No retries permitted until 2025-12-01 15:00:21.991627541 +0000 UTC m=+1587.755137184 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts") pod "barbicanff71-account-delete-2fnl8" (UID: "1ef9fa82-f302-47e9-bed3-b02f34a2340f") : configmap "openstack-scripts" not found Dec 01 15:00:07 crc kubenswrapper[4810]: E1201 15:00:07.320607 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 01 15:00:07 crc kubenswrapper[4810]: E1201 15:00:07.321144 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 01 15:00:07 crc kubenswrapper[4810]: E1201 15:00:07.321383 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 01 15:00:07 crc kubenswrapper[4810]: E1201 15:00:07.321415 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-x82dv" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovsdb-server" Dec 01 15:00:07 crc kubenswrapper[4810]: E1201 15:00:07.321868 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 01 15:00:07 crc kubenswrapper[4810]: E1201 15:00:07.323013 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 01 15:00:07 crc kubenswrapper[4810]: E1201 15:00:07.324101 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 01 15:00:07 crc kubenswrapper[4810]: E1201 15:00:07.324139 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-x82dv" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovs-vswitchd" Dec 01 15:00:12 crc kubenswrapper[4810]: E1201 15:00:12.320456 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 01 15:00:12 crc kubenswrapper[4810]: E1201 15:00:12.321197 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 01 15:00:12 crc kubenswrapper[4810]: E1201 15:00:12.322023 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 01 15:00:12 crc kubenswrapper[4810]: E1201 15:00:12.322026 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 01 15:00:12 crc kubenswrapper[4810]: E1201 15:00:12.322062 4810 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-x82dv" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovsdb-server" Dec 01 15:00:12 crc kubenswrapper[4810]: E1201 15:00:12.323820 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 01 15:00:12 crc kubenswrapper[4810]: E1201 15:00:12.326261 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 01 15:00:12 crc kubenswrapper[4810]: E1201 15:00:12.326311 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-x82dv" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovs-vswitchd" Dec 01 15:00:15 crc kubenswrapper[4810]: I1201 15:00:15.740273 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 01 15:00:15 crc kubenswrapper[4810]: I1201 15:00:15.846855 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"1f455986-23c9-4088-9829-51c74e5b7770\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " Dec 01 15:00:15 crc kubenswrapper[4810]: I1201 15:00:15.846916 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1f455986-23c9-4088-9829-51c74e5b7770-cache\") pod \"1f455986-23c9-4088-9829-51c74e5b7770\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " Dec 01 15:00:15 crc kubenswrapper[4810]: I1201 15:00:15.846975 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift\") pod \"1f455986-23c9-4088-9829-51c74e5b7770\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " Dec 01 15:00:15 crc kubenswrapper[4810]: I1201 15:00:15.847005 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1f455986-23c9-4088-9829-51c74e5b7770-lock\") pod \"1f455986-23c9-4088-9829-51c74e5b7770\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " Dec 01 15:00:15 crc kubenswrapper[4810]: I1201 15:00:15.847031 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mq9v\" (UniqueName: \"kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-kube-api-access-5mq9v\") pod \"1f455986-23c9-4088-9829-51c74e5b7770\" (UID: \"1f455986-23c9-4088-9829-51c74e5b7770\") " Dec 01 15:00:15 crc kubenswrapper[4810]: I1201 15:00:15.848648 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f455986-23c9-4088-9829-51c74e5b7770-cache" (OuterVolumeSpecName: "cache") pod "1f455986-23c9-4088-9829-51c74e5b7770" (UID: "1f455986-23c9-4088-9829-51c74e5b7770"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:00:15 crc kubenswrapper[4810]: I1201 15:00:15.848858 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f455986-23c9-4088-9829-51c74e5b7770-lock" (OuterVolumeSpecName: "lock") pod "1f455986-23c9-4088-9829-51c74e5b7770" (UID: "1f455986-23c9-4088-9829-51c74e5b7770"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:00:15 crc kubenswrapper[4810]: I1201 15:00:15.852623 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-kube-api-access-5mq9v" (OuterVolumeSpecName: "kube-api-access-5mq9v") pod "1f455986-23c9-4088-9829-51c74e5b7770" (UID: "1f455986-23c9-4088-9829-51c74e5b7770"). InnerVolumeSpecName "kube-api-access-5mq9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:00:15 crc kubenswrapper[4810]: I1201 15:00:15.853836 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1f455986-23c9-4088-9829-51c74e5b7770" (UID: "1f455986-23c9-4088-9829-51c74e5b7770"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:00:15 crc kubenswrapper[4810]: I1201 15:00:15.854489 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "swift") pod "1f455986-23c9-4088-9829-51c74e5b7770" (UID: "1f455986-23c9-4088-9829-51c74e5b7770"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 15:00:15 crc kubenswrapper[4810]: I1201 15:00:15.949206 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 01 15:00:15 crc kubenswrapper[4810]: I1201 15:00:15.949250 4810 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1f455986-23c9-4088-9829-51c74e5b7770-cache\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:15 crc kubenswrapper[4810]: I1201 15:00:15.949263 4810 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:15 crc kubenswrapper[4810]: I1201 15:00:15.949274 4810 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1f455986-23c9-4088-9829-51c74e5b7770-lock\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:15 crc kubenswrapper[4810]: I1201 15:00:15.949286 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mq9v\" (UniqueName: \"kubernetes.io/projected/1f455986-23c9-4088-9829-51c74e5b7770-kube-api-access-5mq9v\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:15 crc kubenswrapper[4810]: I1201 15:00:15.963893 4810 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.015486 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-x82dv_9c699a0a-f268-419b-a829-a9ebaa52b322/ovs-vswitchd/0.log" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.016347 4810 generic.go:334] "Generic (PLEG): container finished" podID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" exitCode=137 Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.016435 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-x82dv" event={"ID":"9c699a0a-f268-419b-a829-a9ebaa52b322","Type":"ContainerDied","Data":"74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6"} Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.027499 4810 generic.go:334] "Generic (PLEG): container finished" podID="1f455986-23c9-4088-9829-51c74e5b7770" containerID="24cce379393b557c9eef88db330dd60281782cf5c7918b19088e9f8504b09390" exitCode=137 Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.027540 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerDied","Data":"24cce379393b557c9eef88db330dd60281782cf5c7918b19088e9f8504b09390"} Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.027565 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1f455986-23c9-4088-9829-51c74e5b7770","Type":"ContainerDied","Data":"2f3e61aff3858aad8d5c49a32975a02fe7fc56a522b919189eb31626e1b14ee9"} Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.027582 4810 scope.go:117] "RemoveContainer" containerID="24cce379393b557c9eef88db330dd60281782cf5c7918b19088e9f8504b09390" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.027769 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.050402 4810 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.051328 4810 scope.go:117] "RemoveContainer" containerID="bfb0384344df9db35a4ff5ee4331e251ef264c0c11ab485df0d97f60541193de" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.092027 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.098623 4810 scope.go:117] "RemoveContainer" containerID="10f0178b4b19110c943a68bcae88fac92f4d9354eaf37cbb5fce2607db759cb3" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.102761 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.117301 4810 scope.go:117] "RemoveContainer" containerID="3eeb13544ba7950f14cd2ed7a1ec0e2b77f5688fbb09bf134178e7165abccdf9" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.144722 4810 scope.go:117] "RemoveContainer" containerID="18f2b5779be96d5d0ed49bece9ede2da2e81ab667fe59b1dcf9a253cd41860c6" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.163306 4810 scope.go:117] "RemoveContainer" containerID="4d0480c8c52ca952777f48c3418a4710c30076e9fef37046f82aa4cf60de8531" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.180212 4810 scope.go:117] "RemoveContainer" containerID="ed49f712640c3ec14f04807a032c7cf926007968ca0dcd16ace0f1dc97a84d7a" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.204833 4810 scope.go:117] "RemoveContainer" containerID="0920962c564ec41c800c30de23663892437885ded40e3db2acbc5ff4420710d7" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.222939 4810 scope.go:117] "RemoveContainer" containerID="7c0fca921a04e4ac4e65934e2a8f95df9f334f4836f26bf901cf250082061924" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.282692 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-x82dv_9c699a0a-f268-419b-a829-a9ebaa52b322/ovs-vswitchd/0.log" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.283682 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-x82dv" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.291634 4810 scope.go:117] "RemoveContainer" containerID="7c19582265bd684bc9384473ffe961c1afdbc4a8cbb8d269b1882eb4dab174dc" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.311096 4810 scope.go:117] "RemoveContainer" containerID="184d4e1ea01b4832c5828fa2fe97577fb38a0e14ab4fe7d23a5339c79d094ef7" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.328518 4810 scope.go:117] "RemoveContainer" containerID="7a9772a6d14d26dd0282c93cad34c0f2f79a0710fc08b93b93ed46c435efb103" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.345650 4810 scope.go:117] "RemoveContainer" containerID="8124aab34c8595c54770f6fe4ffe9d15bf793940c0b2cdffcb037b35b9a1e5ce" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.363379 4810 scope.go:117] "RemoveContainer" containerID="7b27d0e70deed824a6d31268fa11d712beb18bff269cd2c61fb15665851547b6" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.378274 4810 scope.go:117] "RemoveContainer" containerID="cc265ca64611ebb2ffafcd64f73b893f0707c1a67a0e439ab6b7c8ef0dec3227" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.392593 4810 scope.go:117] "RemoveContainer" containerID="24cce379393b557c9eef88db330dd60281782cf5c7918b19088e9f8504b09390" Dec 01 15:00:16 crc kubenswrapper[4810]: E1201 15:00:16.392993 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24cce379393b557c9eef88db330dd60281782cf5c7918b19088e9f8504b09390\": container with ID starting with 24cce379393b557c9eef88db330dd60281782cf5c7918b19088e9f8504b09390 not found: ID does not exist" containerID="24cce379393b557c9eef88db330dd60281782cf5c7918b19088e9f8504b09390" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.393031 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24cce379393b557c9eef88db330dd60281782cf5c7918b19088e9f8504b09390"} err="failed to get container status \"24cce379393b557c9eef88db330dd60281782cf5c7918b19088e9f8504b09390\": rpc error: code = NotFound desc = could not find container \"24cce379393b557c9eef88db330dd60281782cf5c7918b19088e9f8504b09390\": container with ID starting with 24cce379393b557c9eef88db330dd60281782cf5c7918b19088e9f8504b09390 not found: ID does not exist" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.393056 4810 scope.go:117] "RemoveContainer" containerID="bfb0384344df9db35a4ff5ee4331e251ef264c0c11ab485df0d97f60541193de" Dec 01 15:00:16 crc kubenswrapper[4810]: E1201 15:00:16.393423 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfb0384344df9db35a4ff5ee4331e251ef264c0c11ab485df0d97f60541193de\": container with ID starting with bfb0384344df9db35a4ff5ee4331e251ef264c0c11ab485df0d97f60541193de not found: ID does not exist" containerID="bfb0384344df9db35a4ff5ee4331e251ef264c0c11ab485df0d97f60541193de" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.393455 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfb0384344df9db35a4ff5ee4331e251ef264c0c11ab485df0d97f60541193de"} err="failed to get container status \"bfb0384344df9db35a4ff5ee4331e251ef264c0c11ab485df0d97f60541193de\": rpc error: code = NotFound desc = could not find container \"bfb0384344df9db35a4ff5ee4331e251ef264c0c11ab485df0d97f60541193de\": container with ID starting with bfb0384344df9db35a4ff5ee4331e251ef264c0c11ab485df0d97f60541193de not found: ID does not exist" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.393515 4810 scope.go:117] "RemoveContainer" containerID="10f0178b4b19110c943a68bcae88fac92f4d9354eaf37cbb5fce2607db759cb3" Dec 01 15:00:16 crc kubenswrapper[4810]: E1201 15:00:16.393825 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10f0178b4b19110c943a68bcae88fac92f4d9354eaf37cbb5fce2607db759cb3\": container with ID starting with 10f0178b4b19110c943a68bcae88fac92f4d9354eaf37cbb5fce2607db759cb3 not found: ID does not exist" containerID="10f0178b4b19110c943a68bcae88fac92f4d9354eaf37cbb5fce2607db759cb3" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.393871 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10f0178b4b19110c943a68bcae88fac92f4d9354eaf37cbb5fce2607db759cb3"} err="failed to get container status \"10f0178b4b19110c943a68bcae88fac92f4d9354eaf37cbb5fce2607db759cb3\": rpc error: code = NotFound desc = could not find container \"10f0178b4b19110c943a68bcae88fac92f4d9354eaf37cbb5fce2607db759cb3\": container with ID starting with 10f0178b4b19110c943a68bcae88fac92f4d9354eaf37cbb5fce2607db759cb3 not found: ID does not exist" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.393895 4810 scope.go:117] "RemoveContainer" containerID="3eeb13544ba7950f14cd2ed7a1ec0e2b77f5688fbb09bf134178e7165abccdf9" Dec 01 15:00:16 crc kubenswrapper[4810]: E1201 15:00:16.394172 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eeb13544ba7950f14cd2ed7a1ec0e2b77f5688fbb09bf134178e7165abccdf9\": container with ID starting with 3eeb13544ba7950f14cd2ed7a1ec0e2b77f5688fbb09bf134178e7165abccdf9 not found: ID does not exist" containerID="3eeb13544ba7950f14cd2ed7a1ec0e2b77f5688fbb09bf134178e7165abccdf9" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.394192 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eeb13544ba7950f14cd2ed7a1ec0e2b77f5688fbb09bf134178e7165abccdf9"} err="failed to get container status \"3eeb13544ba7950f14cd2ed7a1ec0e2b77f5688fbb09bf134178e7165abccdf9\": rpc error: code = NotFound desc = could not find container \"3eeb13544ba7950f14cd2ed7a1ec0e2b77f5688fbb09bf134178e7165abccdf9\": container with ID starting with 3eeb13544ba7950f14cd2ed7a1ec0e2b77f5688fbb09bf134178e7165abccdf9 not found: ID does not exist" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.394203 4810 scope.go:117] "RemoveContainer" containerID="18f2b5779be96d5d0ed49bece9ede2da2e81ab667fe59b1dcf9a253cd41860c6" Dec 01 15:00:16 crc kubenswrapper[4810]: E1201 15:00:16.394514 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18f2b5779be96d5d0ed49bece9ede2da2e81ab667fe59b1dcf9a253cd41860c6\": container with ID starting with 18f2b5779be96d5d0ed49bece9ede2da2e81ab667fe59b1dcf9a253cd41860c6 not found: ID does not exist" containerID="18f2b5779be96d5d0ed49bece9ede2da2e81ab667fe59b1dcf9a253cd41860c6" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.394539 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18f2b5779be96d5d0ed49bece9ede2da2e81ab667fe59b1dcf9a253cd41860c6"} err="failed to get container status \"18f2b5779be96d5d0ed49bece9ede2da2e81ab667fe59b1dcf9a253cd41860c6\": rpc error: code = NotFound desc = could not find container \"18f2b5779be96d5d0ed49bece9ede2da2e81ab667fe59b1dcf9a253cd41860c6\": container with ID starting with 18f2b5779be96d5d0ed49bece9ede2da2e81ab667fe59b1dcf9a253cd41860c6 not found: ID does not exist" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.394555 4810 scope.go:117] "RemoveContainer" containerID="4d0480c8c52ca952777f48c3418a4710c30076e9fef37046f82aa4cf60de8531" Dec 01 15:00:16 crc kubenswrapper[4810]: E1201 15:00:16.394885 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d0480c8c52ca952777f48c3418a4710c30076e9fef37046f82aa4cf60de8531\": container with ID starting with 4d0480c8c52ca952777f48c3418a4710c30076e9fef37046f82aa4cf60de8531 not found: ID does not exist" containerID="4d0480c8c52ca952777f48c3418a4710c30076e9fef37046f82aa4cf60de8531" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.394915 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d0480c8c52ca952777f48c3418a4710c30076e9fef37046f82aa4cf60de8531"} err="failed to get container status \"4d0480c8c52ca952777f48c3418a4710c30076e9fef37046f82aa4cf60de8531\": rpc error: code = NotFound desc = could not find container \"4d0480c8c52ca952777f48c3418a4710c30076e9fef37046f82aa4cf60de8531\": container with ID starting with 4d0480c8c52ca952777f48c3418a4710c30076e9fef37046f82aa4cf60de8531 not found: ID does not exist" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.394932 4810 scope.go:117] "RemoveContainer" containerID="ed49f712640c3ec14f04807a032c7cf926007968ca0dcd16ace0f1dc97a84d7a" Dec 01 15:00:16 crc kubenswrapper[4810]: E1201 15:00:16.395189 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed49f712640c3ec14f04807a032c7cf926007968ca0dcd16ace0f1dc97a84d7a\": container with ID starting with ed49f712640c3ec14f04807a032c7cf926007968ca0dcd16ace0f1dc97a84d7a not found: ID does not exist" containerID="ed49f712640c3ec14f04807a032c7cf926007968ca0dcd16ace0f1dc97a84d7a" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.395211 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed49f712640c3ec14f04807a032c7cf926007968ca0dcd16ace0f1dc97a84d7a"} err="failed to get container status \"ed49f712640c3ec14f04807a032c7cf926007968ca0dcd16ace0f1dc97a84d7a\": rpc error: code = NotFound desc = could not find container \"ed49f712640c3ec14f04807a032c7cf926007968ca0dcd16ace0f1dc97a84d7a\": container with ID starting with ed49f712640c3ec14f04807a032c7cf926007968ca0dcd16ace0f1dc97a84d7a not found: ID does not exist" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.395223 4810 scope.go:117] "RemoveContainer" containerID="0920962c564ec41c800c30de23663892437885ded40e3db2acbc5ff4420710d7" Dec 01 15:00:16 crc kubenswrapper[4810]: E1201 15:00:16.395545 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0920962c564ec41c800c30de23663892437885ded40e3db2acbc5ff4420710d7\": container with ID starting with 0920962c564ec41c800c30de23663892437885ded40e3db2acbc5ff4420710d7 not found: ID does not exist" containerID="0920962c564ec41c800c30de23663892437885ded40e3db2acbc5ff4420710d7" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.395578 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0920962c564ec41c800c30de23663892437885ded40e3db2acbc5ff4420710d7"} err="failed to get container status \"0920962c564ec41c800c30de23663892437885ded40e3db2acbc5ff4420710d7\": rpc error: code = NotFound desc = could not find container \"0920962c564ec41c800c30de23663892437885ded40e3db2acbc5ff4420710d7\": container with ID starting with 0920962c564ec41c800c30de23663892437885ded40e3db2acbc5ff4420710d7 not found: ID does not exist" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.395601 4810 scope.go:117] "RemoveContainer" containerID="7c0fca921a04e4ac4e65934e2a8f95df9f334f4836f26bf901cf250082061924" Dec 01 15:00:16 crc kubenswrapper[4810]: E1201 15:00:16.395842 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c0fca921a04e4ac4e65934e2a8f95df9f334f4836f26bf901cf250082061924\": container with ID starting with 7c0fca921a04e4ac4e65934e2a8f95df9f334f4836f26bf901cf250082061924 not found: ID does not exist" containerID="7c0fca921a04e4ac4e65934e2a8f95df9f334f4836f26bf901cf250082061924" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.395879 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c0fca921a04e4ac4e65934e2a8f95df9f334f4836f26bf901cf250082061924"} err="failed to get container status \"7c0fca921a04e4ac4e65934e2a8f95df9f334f4836f26bf901cf250082061924\": rpc error: code = NotFound desc = could not find container \"7c0fca921a04e4ac4e65934e2a8f95df9f334f4836f26bf901cf250082061924\": container with ID starting with 7c0fca921a04e4ac4e65934e2a8f95df9f334f4836f26bf901cf250082061924 not found: ID does not exist" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.395895 4810 scope.go:117] "RemoveContainer" containerID="7c19582265bd684bc9384473ffe961c1afdbc4a8cbb8d269b1882eb4dab174dc" Dec 01 15:00:16 crc kubenswrapper[4810]: E1201 15:00:16.396122 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c19582265bd684bc9384473ffe961c1afdbc4a8cbb8d269b1882eb4dab174dc\": container with ID starting with 7c19582265bd684bc9384473ffe961c1afdbc4a8cbb8d269b1882eb4dab174dc not found: ID does not exist" containerID="7c19582265bd684bc9384473ffe961c1afdbc4a8cbb8d269b1882eb4dab174dc" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.396148 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c19582265bd684bc9384473ffe961c1afdbc4a8cbb8d269b1882eb4dab174dc"} err="failed to get container status \"7c19582265bd684bc9384473ffe961c1afdbc4a8cbb8d269b1882eb4dab174dc\": rpc error: code = NotFound desc = could not find container \"7c19582265bd684bc9384473ffe961c1afdbc4a8cbb8d269b1882eb4dab174dc\": container with ID starting with 7c19582265bd684bc9384473ffe961c1afdbc4a8cbb8d269b1882eb4dab174dc not found: ID does not exist" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.396166 4810 scope.go:117] "RemoveContainer" containerID="184d4e1ea01b4832c5828fa2fe97577fb38a0e14ab4fe7d23a5339c79d094ef7" Dec 01 15:00:16 crc kubenswrapper[4810]: E1201 15:00:16.396367 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"184d4e1ea01b4832c5828fa2fe97577fb38a0e14ab4fe7d23a5339c79d094ef7\": container with ID starting with 184d4e1ea01b4832c5828fa2fe97577fb38a0e14ab4fe7d23a5339c79d094ef7 not found: ID does not exist" containerID="184d4e1ea01b4832c5828fa2fe97577fb38a0e14ab4fe7d23a5339c79d094ef7" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.396426 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"184d4e1ea01b4832c5828fa2fe97577fb38a0e14ab4fe7d23a5339c79d094ef7"} err="failed to get container status \"184d4e1ea01b4832c5828fa2fe97577fb38a0e14ab4fe7d23a5339c79d094ef7\": rpc error: code = NotFound desc = could not find container \"184d4e1ea01b4832c5828fa2fe97577fb38a0e14ab4fe7d23a5339c79d094ef7\": container with ID starting with 184d4e1ea01b4832c5828fa2fe97577fb38a0e14ab4fe7d23a5339c79d094ef7 not found: ID does not exist" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.396443 4810 scope.go:117] "RemoveContainer" containerID="7a9772a6d14d26dd0282c93cad34c0f2f79a0710fc08b93b93ed46c435efb103" Dec 01 15:00:16 crc kubenswrapper[4810]: E1201 15:00:16.396736 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a9772a6d14d26dd0282c93cad34c0f2f79a0710fc08b93b93ed46c435efb103\": container with ID starting with 7a9772a6d14d26dd0282c93cad34c0f2f79a0710fc08b93b93ed46c435efb103 not found: ID does not exist" containerID="7a9772a6d14d26dd0282c93cad34c0f2f79a0710fc08b93b93ed46c435efb103" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.396772 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a9772a6d14d26dd0282c93cad34c0f2f79a0710fc08b93b93ed46c435efb103"} err="failed to get container status \"7a9772a6d14d26dd0282c93cad34c0f2f79a0710fc08b93b93ed46c435efb103\": rpc error: code = NotFound desc = could not find container \"7a9772a6d14d26dd0282c93cad34c0f2f79a0710fc08b93b93ed46c435efb103\": container with ID starting with 7a9772a6d14d26dd0282c93cad34c0f2f79a0710fc08b93b93ed46c435efb103 not found: ID does not exist" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.396793 4810 scope.go:117] "RemoveContainer" containerID="8124aab34c8595c54770f6fe4ffe9d15bf793940c0b2cdffcb037b35b9a1e5ce" Dec 01 15:00:16 crc kubenswrapper[4810]: E1201 15:00:16.397088 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8124aab34c8595c54770f6fe4ffe9d15bf793940c0b2cdffcb037b35b9a1e5ce\": container with ID starting with 8124aab34c8595c54770f6fe4ffe9d15bf793940c0b2cdffcb037b35b9a1e5ce not found: ID does not exist" containerID="8124aab34c8595c54770f6fe4ffe9d15bf793940c0b2cdffcb037b35b9a1e5ce" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.397107 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8124aab34c8595c54770f6fe4ffe9d15bf793940c0b2cdffcb037b35b9a1e5ce"} err="failed to get container status \"8124aab34c8595c54770f6fe4ffe9d15bf793940c0b2cdffcb037b35b9a1e5ce\": rpc error: code = NotFound desc = could not find container \"8124aab34c8595c54770f6fe4ffe9d15bf793940c0b2cdffcb037b35b9a1e5ce\": container with ID starting with 8124aab34c8595c54770f6fe4ffe9d15bf793940c0b2cdffcb037b35b9a1e5ce not found: ID does not exist" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.397119 4810 scope.go:117] "RemoveContainer" containerID="7b27d0e70deed824a6d31268fa11d712beb18bff269cd2c61fb15665851547b6" Dec 01 15:00:16 crc kubenswrapper[4810]: E1201 15:00:16.397368 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b27d0e70deed824a6d31268fa11d712beb18bff269cd2c61fb15665851547b6\": container with ID starting with 7b27d0e70deed824a6d31268fa11d712beb18bff269cd2c61fb15665851547b6 not found: ID does not exist" containerID="7b27d0e70deed824a6d31268fa11d712beb18bff269cd2c61fb15665851547b6" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.397400 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b27d0e70deed824a6d31268fa11d712beb18bff269cd2c61fb15665851547b6"} err="failed to get container status \"7b27d0e70deed824a6d31268fa11d712beb18bff269cd2c61fb15665851547b6\": rpc error: code = NotFound desc = could not find container \"7b27d0e70deed824a6d31268fa11d712beb18bff269cd2c61fb15665851547b6\": container with ID starting with 7b27d0e70deed824a6d31268fa11d712beb18bff269cd2c61fb15665851547b6 not found: ID does not exist" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.397421 4810 scope.go:117] "RemoveContainer" containerID="cc265ca64611ebb2ffafcd64f73b893f0707c1a67a0e439ab6b7c8ef0dec3227" Dec 01 15:00:16 crc kubenswrapper[4810]: E1201 15:00:16.397691 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc265ca64611ebb2ffafcd64f73b893f0707c1a67a0e439ab6b7c8ef0dec3227\": container with ID starting with cc265ca64611ebb2ffafcd64f73b893f0707c1a67a0e439ab6b7c8ef0dec3227 not found: ID does not exist" containerID="cc265ca64611ebb2ffafcd64f73b893f0707c1a67a0e439ab6b7c8ef0dec3227" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.397711 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc265ca64611ebb2ffafcd64f73b893f0707c1a67a0e439ab6b7c8ef0dec3227"} err="failed to get container status \"cc265ca64611ebb2ffafcd64f73b893f0707c1a67a0e439ab6b7c8ef0dec3227\": rpc error: code = NotFound desc = could not find container \"cc265ca64611ebb2ffafcd64f73b893f0707c1a67a0e439ab6b7c8ef0dec3227\": container with ID starting with cc265ca64611ebb2ffafcd64f73b893f0707c1a67a0e439ab6b7c8ef0dec3227 not found: ID does not exist" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.456539 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c699a0a-f268-419b-a829-a9ebaa52b322-scripts\") pod \"9c699a0a-f268-419b-a829-a9ebaa52b322\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.456647 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-etc-ovs\") pod \"9c699a0a-f268-419b-a829-a9ebaa52b322\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.456698 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-var-lib\") pod \"9c699a0a-f268-419b-a829-a9ebaa52b322\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.456761 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-var-run\") pod \"9c699a0a-f268-419b-a829-a9ebaa52b322\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.456772 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "9c699a0a-f268-419b-a829-a9ebaa52b322" (UID: "9c699a0a-f268-419b-a829-a9ebaa52b322"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.456790 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dm7m\" (UniqueName: \"kubernetes.io/projected/9c699a0a-f268-419b-a829-a9ebaa52b322-kube-api-access-7dm7m\") pod \"9c699a0a-f268-419b-a829-a9ebaa52b322\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.456840 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-var-log\") pod \"9c699a0a-f268-419b-a829-a9ebaa52b322\" (UID: \"9c699a0a-f268-419b-a829-a9ebaa52b322\") " Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.456860 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-var-lib" (OuterVolumeSpecName: "var-lib") pod "9c699a0a-f268-419b-a829-a9ebaa52b322" (UID: "9c699a0a-f268-419b-a829-a9ebaa52b322"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.456896 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-var-run" (OuterVolumeSpecName: "var-run") pod "9c699a0a-f268-419b-a829-a9ebaa52b322" (UID: "9c699a0a-f268-419b-a829-a9ebaa52b322"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.456980 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-var-log" (OuterVolumeSpecName: "var-log") pod "9c699a0a-f268-419b-a829-a9ebaa52b322" (UID: "9c699a0a-f268-419b-a829-a9ebaa52b322"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.457246 4810 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-var-run\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.457268 4810 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-var-log\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.457279 4810 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-etc-ovs\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.457289 4810 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9c699a0a-f268-419b-a829-a9ebaa52b322-var-lib\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.457731 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c699a0a-f268-419b-a829-a9ebaa52b322-scripts" (OuterVolumeSpecName: "scripts") pod "9c699a0a-f268-419b-a829-a9ebaa52b322" (UID: "9c699a0a-f268-419b-a829-a9ebaa52b322"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.461670 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c699a0a-f268-419b-a829-a9ebaa52b322-kube-api-access-7dm7m" (OuterVolumeSpecName: "kube-api-access-7dm7m") pod "9c699a0a-f268-419b-a829-a9ebaa52b322" (UID: "9c699a0a-f268-419b-a829-a9ebaa52b322"). InnerVolumeSpecName "kube-api-access-7dm7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.508464 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f455986-23c9-4088-9829-51c74e5b7770" path="/var/lib/kubelet/pods/1f455986-23c9-4088-9829-51c74e5b7770/volumes" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.558723 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dm7m\" (UniqueName: \"kubernetes.io/projected/9c699a0a-f268-419b-a829-a9ebaa52b322-kube-api-access-7dm7m\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.558758 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c699a0a-f268-419b-a829-a9ebaa52b322-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:16 crc kubenswrapper[4810]: I1201 15:00:16.816327 4810 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podfd2094a7-9a7b-417b-b4ea-b1da8b69cac6"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podfd2094a7-9a7b-417b-b4ea-b1da8b69cac6] : Timed out while waiting for systemd to remove kubepods-besteffort-podfd2094a7_9a7b_417b_b4ea_b1da8b69cac6.slice" Dec 01 15:00:16 crc kubenswrapper[4810]: E1201 15:00:16.816377 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort podfd2094a7-9a7b-417b-b4ea-b1da8b69cac6] : unable to destroy cgroup paths for cgroup [kubepods besteffort podfd2094a7-9a7b-417b-b4ea-b1da8b69cac6] : Timed out while waiting for systemd to remove kubepods-besteffort-podfd2094a7_9a7b_417b_b4ea_b1da8b69cac6.slice" pod="openstack/ovn-controller-metrics-6lqnm" podUID="fd2094a7-9a7b-417b-b4ea-b1da8b69cac6" Dec 01 15:00:17 crc kubenswrapper[4810]: I1201 15:00:17.042500 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-x82dv_9c699a0a-f268-419b-a829-a9ebaa52b322/ovs-vswitchd/0.log" Dec 01 15:00:17 crc kubenswrapper[4810]: I1201 15:00:17.044064 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-6lqnm" Dec 01 15:00:17 crc kubenswrapper[4810]: I1201 15:00:17.044114 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-x82dv" event={"ID":"9c699a0a-f268-419b-a829-a9ebaa52b322","Type":"ContainerDied","Data":"ee51a4871d387f6ab294b0535c53206dcd1bf44b9ed26b71d9760d06883f5b7f"} Dec 01 15:00:17 crc kubenswrapper[4810]: I1201 15:00:17.044185 4810 scope.go:117] "RemoveContainer" containerID="74929c3756628b07ae5a01f02851d6343e834469bd57233ad4699fd520b83fb6" Dec 01 15:00:17 crc kubenswrapper[4810]: I1201 15:00:17.044445 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-x82dv" Dec 01 15:00:17 crc kubenswrapper[4810]: I1201 15:00:17.087277 4810 scope.go:117] "RemoveContainer" containerID="0105bee3b90c810edb14fcce980c51fa9a5be9e0e826388387b035db74313a22" Dec 01 15:00:17 crc kubenswrapper[4810]: I1201 15:00:17.088626 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-x82dv"] Dec 01 15:00:17 crc kubenswrapper[4810]: I1201 15:00:17.097348 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-x82dv"] Dec 01 15:00:17 crc kubenswrapper[4810]: I1201 15:00:17.106353 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-6lqnm"] Dec 01 15:00:17 crc kubenswrapper[4810]: I1201 15:00:17.106429 4810 scope.go:117] "RemoveContainer" containerID="0dd2c680c8d22899f94d6de624293194c03d0629c22fb4e022cc58fae7b6f0aa" Dec 01 15:00:17 crc kubenswrapper[4810]: I1201 15:00:17.113322 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-6lqnm"] Dec 01 15:00:18 crc kubenswrapper[4810]: I1201 15:00:18.502403 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" path="/var/lib/kubelet/pods/9c699a0a-f268-419b-a829-a9ebaa52b322/volumes" Dec 01 15:00:18 crc kubenswrapper[4810]: I1201 15:00:18.503882 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd2094a7-9a7b-417b-b4ea-b1da8b69cac6" path="/var/lib/kubelet/pods/fd2094a7-9a7b-417b-b4ea-b1da8b69cac6/volumes" Dec 01 15:00:20 crc kubenswrapper[4810]: I1201 15:00:20.156965 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-6c5f549c7d-52fkz" podUID="31b83b20-83c4-4e06-8eb0-1e1d83d4b160" containerName="placement-api" probeResult="failure" output="Get \"https://10.217.0.155:8778/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 01 15:00:20 crc kubenswrapper[4810]: I1201 15:00:20.161769 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-6c5f549c7d-52fkz" podUID="31b83b20-83c4-4e06-8eb0-1e1d83d4b160" containerName="placement-log" probeResult="failure" output="Get \"https://10.217.0.155:8778/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 01 15:00:20 crc kubenswrapper[4810]: E1201 15:00:20.870267 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 15:00:20 crc kubenswrapper[4810]: E1201 15:00:20.870376 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts podName:476b71ac-1050-46a6-b51c-168e75b1a350 nodeName:}" failed. No retries permitted until 2025-12-01 15:00:52.870351416 +0000 UTC m=+1618.633861059 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts") pod "novaapic0fd-account-delete-w7ttc" (UID: "476b71ac-1050-46a6-b51c-168e75b1a350") : configmap "openstack-scripts" not found Dec 01 15:00:20 crc kubenswrapper[4810]: E1201 15:00:20.972576 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 15:00:20 crc kubenswrapper[4810]: E1201 15:00:20.972623 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 15:00:20 crc kubenswrapper[4810]: E1201 15:00:20.972667 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts podName:f7b06553-ec05-4c6f-82c0-ef0b4dd3005a nodeName:}" failed. No retries permitted until 2025-12-01 15:00:52.972648276 +0000 UTC m=+1618.736157869 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts") pod "neutronc91e-account-delete-t585q" (UID: "f7b06553-ec05-4c6f-82c0-ef0b4dd3005a") : configmap "openstack-scripts" not found Dec 01 15:00:20 crc kubenswrapper[4810]: E1201 15:00:20.972689 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts podName:bf15ccdf-1e4e-44ab-8ccb-819502935a7a nodeName:}" failed. No retries permitted until 2025-12-01 15:00:52.972681637 +0000 UTC m=+1618.736191240 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts") pod "cinder17c6-account-delete-qktk6" (UID: "bf15ccdf-1e4e-44ab-8ccb-819502935a7a") : configmap "openstack-scripts" not found Dec 01 15:00:21 crc kubenswrapper[4810]: E1201 15:00:21.872061 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7b06553_ec05_4c6f_82c0_ef0b4dd3005a.slice/crio-conmon-2ba07f4b84a146833c2e8b31369eaa1e961a68fc47ee9b4c60c092208482cff9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ef9fa82_f302_47e9_bed3_b02f34a2340f.slice/crio-conmon-dfcba994df93f276679f0b982a3fda4b0ba3b97518b8e6d0c24012c34d33e149.scope\": RecentStats: unable to find data in memory cache]" Dec 01 15:00:22 crc kubenswrapper[4810]: E1201 15:00:22.089997 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 15:00:22 crc kubenswrapper[4810]: E1201 15:00:22.090283 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts podName:4f7b5a25-eed0-4122-a036-093ac7adca22 nodeName:}" failed. No retries permitted until 2025-12-01 15:00:54.090264783 +0000 UTC m=+1619.853774386 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts") pod "novacell096e4-account-delete-qqlg4" (UID: "4f7b5a25-eed0-4122-a036-093ac7adca22") : configmap "openstack-scripts" not found Dec 01 15:00:22 crc kubenswrapper[4810]: E1201 15:00:22.089933 4810 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 01 15:00:22 crc kubenswrapper[4810]: E1201 15:00:22.090394 4810 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts podName:1ef9fa82-f302-47e9-bed3-b02f34a2340f nodeName:}" failed. No retries permitted until 2025-12-01 15:00:54.090355435 +0000 UTC m=+1619.853865078 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts") pod "barbicanff71-account-delete-2fnl8" (UID: "1ef9fa82-f302-47e9-bed3-b02f34a2340f") : configmap "openstack-scripts" not found Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.092338 4810 generic.go:334] "Generic (PLEG): container finished" podID="1ef9fa82-f302-47e9-bed3-b02f34a2340f" containerID="dfcba994df93f276679f0b982a3fda4b0ba3b97518b8e6d0c24012c34d33e149" exitCode=137 Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.092396 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanff71-account-delete-2fnl8" event={"ID":"1ef9fa82-f302-47e9-bed3-b02f34a2340f","Type":"ContainerDied","Data":"dfcba994df93f276679f0b982a3fda4b0ba3b97518b8e6d0c24012c34d33e149"} Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.092420 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanff71-account-delete-2fnl8" event={"ID":"1ef9fa82-f302-47e9-bed3-b02f34a2340f","Type":"ContainerDied","Data":"c8a7aa776dc68e3946de920183358dd2dab73021442a06e65e4afc5eae2312fe"} Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.092432 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8a7aa776dc68e3946de920183358dd2dab73021442a06e65e4afc5eae2312fe" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.095059 4810 generic.go:334] "Generic (PLEG): container finished" podID="476b71ac-1050-46a6-b51c-168e75b1a350" containerID="691a88e3e213a94e7564d8b43ac6e1b2e48afa2314fc5a29d364c71768f366fb" exitCode=137 Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.095102 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapic0fd-account-delete-w7ttc" event={"ID":"476b71ac-1050-46a6-b51c-168e75b1a350","Type":"ContainerDied","Data":"691a88e3e213a94e7564d8b43ac6e1b2e48afa2314fc5a29d364c71768f366fb"} Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.096907 4810 generic.go:334] "Generic (PLEG): container finished" podID="f7b06553-ec05-4c6f-82c0-ef0b4dd3005a" containerID="2ba07f4b84a146833c2e8b31369eaa1e961a68fc47ee9b4c60c092208482cff9" exitCode=137 Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.096953 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronc91e-account-delete-t585q" event={"ID":"f7b06553-ec05-4c6f-82c0-ef0b4dd3005a","Type":"ContainerDied","Data":"2ba07f4b84a146833c2e8b31369eaa1e961a68fc47ee9b4c60c092208482cff9"} Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.096970 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutronc91e-account-delete-t585q" event={"ID":"f7b06553-ec05-4c6f-82c0-ef0b4dd3005a","Type":"ContainerDied","Data":"6d62c108069647c18de078a2ba918fa95f1dc3fb75187b7cfcc958cebe7a3323"} Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.096980 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d62c108069647c18de078a2ba918fa95f1dc3fb75187b7cfcc958cebe7a3323" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.099624 4810 generic.go:334] "Generic (PLEG): container finished" podID="bf15ccdf-1e4e-44ab-8ccb-819502935a7a" containerID="afb9f73f83766bd0ed43d8d077992ad62bc5c10098a5767cbf7c3f030466f40f" exitCode=137 Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.099687 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder17c6-account-delete-qktk6" event={"ID":"bf15ccdf-1e4e-44ab-8ccb-819502935a7a","Type":"ContainerDied","Data":"afb9f73f83766bd0ed43d8d077992ad62bc5c10098a5767cbf7c3f030466f40f"} Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.135882 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutronc91e-account-delete-t585q" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.143452 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanff71-account-delete-2fnl8" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.156171 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic0fd-account-delete-w7ttc" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.190923 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8cjq\" (UniqueName: \"kubernetes.io/projected/1ef9fa82-f302-47e9-bed3-b02f34a2340f-kube-api-access-l8cjq\") pod \"1ef9fa82-f302-47e9-bed3-b02f34a2340f\" (UID: \"1ef9fa82-f302-47e9-bed3-b02f34a2340f\") " Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.191398 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts\") pod \"1ef9fa82-f302-47e9-bed3-b02f34a2340f\" (UID: \"1ef9fa82-f302-47e9-bed3-b02f34a2340f\") " Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.191614 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qgd7\" (UniqueName: \"kubernetes.io/projected/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-kube-api-access-6qgd7\") pod \"f7b06553-ec05-4c6f-82c0-ef0b4dd3005a\" (UID: \"f7b06553-ec05-4c6f-82c0-ef0b4dd3005a\") " Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.191828 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts\") pod \"f7b06553-ec05-4c6f-82c0-ef0b4dd3005a\" (UID: \"f7b06553-ec05-4c6f-82c0-ef0b4dd3005a\") " Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.192059 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts\") pod \"476b71ac-1050-46a6-b51c-168e75b1a350\" (UID: \"476b71ac-1050-46a6-b51c-168e75b1a350\") " Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.192225 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x66rb\" (UniqueName: \"kubernetes.io/projected/476b71ac-1050-46a6-b51c-168e75b1a350-kube-api-access-x66rb\") pod \"476b71ac-1050-46a6-b51c-168e75b1a350\" (UID: \"476b71ac-1050-46a6-b51c-168e75b1a350\") " Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.192716 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f7b06553-ec05-4c6f-82c0-ef0b4dd3005a" (UID: "f7b06553-ec05-4c6f-82c0-ef0b4dd3005a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.192877 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "476b71ac-1050-46a6-b51c-168e75b1a350" (UID: "476b71ac-1050-46a6-b51c-168e75b1a350"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.193203 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.193329 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/476b71ac-1050-46a6-b51c-168e75b1a350-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.196410 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1ef9fa82-f302-47e9-bed3-b02f34a2340f" (UID: "1ef9fa82-f302-47e9-bed3-b02f34a2340f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.198281 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/476b71ac-1050-46a6-b51c-168e75b1a350-kube-api-access-x66rb" (OuterVolumeSpecName: "kube-api-access-x66rb") pod "476b71ac-1050-46a6-b51c-168e75b1a350" (UID: "476b71ac-1050-46a6-b51c-168e75b1a350"). InnerVolumeSpecName "kube-api-access-x66rb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.198357 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-kube-api-access-6qgd7" (OuterVolumeSpecName: "kube-api-access-6qgd7") pod "f7b06553-ec05-4c6f-82c0-ef0b4dd3005a" (UID: "f7b06553-ec05-4c6f-82c0-ef0b4dd3005a"). InnerVolumeSpecName "kube-api-access-6qgd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.198441 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ef9fa82-f302-47e9-bed3-b02f34a2340f-kube-api-access-l8cjq" (OuterVolumeSpecName: "kube-api-access-l8cjq") pod "1ef9fa82-f302-47e9-bed3-b02f34a2340f" (UID: "1ef9fa82-f302-47e9-bed3-b02f34a2340f"). InnerVolumeSpecName "kube-api-access-l8cjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.295577 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8cjq\" (UniqueName: \"kubernetes.io/projected/1ef9fa82-f302-47e9-bed3-b02f34a2340f-kube-api-access-l8cjq\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.295619 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ef9fa82-f302-47e9-bed3-b02f34a2340f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.295630 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qgd7\" (UniqueName: \"kubernetes.io/projected/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a-kube-api-access-6qgd7\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.295641 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x66rb\" (UniqueName: \"kubernetes.io/projected/476b71ac-1050-46a6-b51c-168e75b1a350-kube-api-access-x66rb\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.432236 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder17c6-account-delete-qktk6" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.498523 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7274\" (UniqueName: \"kubernetes.io/projected/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-kube-api-access-n7274\") pod \"bf15ccdf-1e4e-44ab-8ccb-819502935a7a\" (UID: \"bf15ccdf-1e4e-44ab-8ccb-819502935a7a\") " Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.498590 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts\") pod \"bf15ccdf-1e4e-44ab-8ccb-819502935a7a\" (UID: \"bf15ccdf-1e4e-44ab-8ccb-819502935a7a\") " Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.499282 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bf15ccdf-1e4e-44ab-8ccb-819502935a7a" (UID: "bf15ccdf-1e4e-44ab-8ccb-819502935a7a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.502003 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-kube-api-access-n7274" (OuterVolumeSpecName: "kube-api-access-n7274") pod "bf15ccdf-1e4e-44ab-8ccb-819502935a7a" (UID: "bf15ccdf-1e4e-44ab-8ccb-819502935a7a"). InnerVolumeSpecName "kube-api-access-n7274". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.600102 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7274\" (UniqueName: \"kubernetes.io/projected/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-kube-api-access-n7274\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:22 crc kubenswrapper[4810]: I1201 15:00:22.600132 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf15ccdf-1e4e-44ab-8ccb-819502935a7a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.013515 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell096e4-account-delete-qqlg4" Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.105124 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bc8s4\" (UniqueName: \"kubernetes.io/projected/4f7b5a25-eed0-4122-a036-093ac7adca22-kube-api-access-bc8s4\") pod \"4f7b5a25-eed0-4122-a036-093ac7adca22\" (UID: \"4f7b5a25-eed0-4122-a036-093ac7adca22\") " Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.105390 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts\") pod \"4f7b5a25-eed0-4122-a036-093ac7adca22\" (UID: \"4f7b5a25-eed0-4122-a036-093ac7adca22\") " Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.106144 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4f7b5a25-eed0-4122-a036-093ac7adca22" (UID: "4f7b5a25-eed0-4122-a036-093ac7adca22"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.108731 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f7b5a25-eed0-4122-a036-093ac7adca22-kube-api-access-bc8s4" (OuterVolumeSpecName: "kube-api-access-bc8s4") pod "4f7b5a25-eed0-4122-a036-093ac7adca22" (UID: "4f7b5a25-eed0-4122-a036-093ac7adca22"). InnerVolumeSpecName "kube-api-access-bc8s4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.109776 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder17c6-account-delete-qktk6" event={"ID":"bf15ccdf-1e4e-44ab-8ccb-819502935a7a","Type":"ContainerDied","Data":"0a07bd9197943b1a845068c9151c0945af10f687ff4d4b2b37177546b1a72c14"} Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.109830 4810 scope.go:117] "RemoveContainer" containerID="afb9f73f83766bd0ed43d8d077992ad62bc5c10098a5767cbf7c3f030466f40f" Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.109996 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder17c6-account-delete-qktk6" Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.117729 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapic0fd-account-delete-w7ttc" event={"ID":"476b71ac-1050-46a6-b51c-168e75b1a350","Type":"ContainerDied","Data":"6e9b6324957395c9efa53642231b0756aaf310606b0fbc1e11aee7cd643bdc77"} Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.117808 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic0fd-account-delete-w7ttc" Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.123204 4810 generic.go:334] "Generic (PLEG): container finished" podID="4f7b5a25-eed0-4122-a036-093ac7adca22" containerID="ba6bbaa4e5865842e4d015f8e09df7c3af89eab522256f5f54fc25471d6b214f" exitCode=137 Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.123240 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell096e4-account-delete-qqlg4" event={"ID":"4f7b5a25-eed0-4122-a036-093ac7adca22","Type":"ContainerDied","Data":"ba6bbaa4e5865842e4d015f8e09df7c3af89eab522256f5f54fc25471d6b214f"} Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.123277 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell096e4-account-delete-qqlg4" Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.123298 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell096e4-account-delete-qqlg4" event={"ID":"4f7b5a25-eed0-4122-a036-093ac7adca22","Type":"ContainerDied","Data":"b49f17b2795a2340ebd876f7a3c218b2cc7a3ea65b59733f36cbea047a113f55"} Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.123310 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutronc91e-account-delete-t585q" Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.123285 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanff71-account-delete-2fnl8" Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.149558 4810 scope.go:117] "RemoveContainer" containerID="691a88e3e213a94e7564d8b43ac6e1b2e48afa2314fc5a29d364c71768f366fb" Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.168483 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicanff71-account-delete-2fnl8"] Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.180163 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbicanff71-account-delete-2fnl8"] Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.188738 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder17c6-account-delete-qktk6"] Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.190435 4810 scope.go:117] "RemoveContainer" containerID="ba6bbaa4e5865842e4d015f8e09df7c3af89eab522256f5f54fc25471d6b214f" Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.203370 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder17c6-account-delete-qktk6"] Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.207433 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f7b5a25-eed0-4122-a036-093ac7adca22-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.207486 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bc8s4\" (UniqueName: \"kubernetes.io/projected/4f7b5a25-eed0-4122-a036-093ac7adca22-kube-api-access-bc8s4\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.215121 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell096e4-account-delete-qqlg4"] Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.223467 4810 scope.go:117] "RemoveContainer" containerID="ba6bbaa4e5865842e4d015f8e09df7c3af89eab522256f5f54fc25471d6b214f" Dec 01 15:00:23 crc kubenswrapper[4810]: E1201 15:00:23.223872 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba6bbaa4e5865842e4d015f8e09df7c3af89eab522256f5f54fc25471d6b214f\": container with ID starting with ba6bbaa4e5865842e4d015f8e09df7c3af89eab522256f5f54fc25471d6b214f not found: ID does not exist" containerID="ba6bbaa4e5865842e4d015f8e09df7c3af89eab522256f5f54fc25471d6b214f" Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.223901 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba6bbaa4e5865842e4d015f8e09df7c3af89eab522256f5f54fc25471d6b214f"} err="failed to get container status \"ba6bbaa4e5865842e4d015f8e09df7c3af89eab522256f5f54fc25471d6b214f\": rpc error: code = NotFound desc = could not find container \"ba6bbaa4e5865842e4d015f8e09df7c3af89eab522256f5f54fc25471d6b214f\": container with ID starting with ba6bbaa4e5865842e4d015f8e09df7c3af89eab522256f5f54fc25471d6b214f not found: ID does not exist" Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.224678 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell096e4-account-delete-qqlg4"] Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.231785 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapic0fd-account-delete-w7ttc"] Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.238244 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapic0fd-account-delete-w7ttc"] Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.245061 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutronc91e-account-delete-t585q"] Dec 01 15:00:23 crc kubenswrapper[4810]: I1201 15:00:23.251068 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutronc91e-account-delete-t585q"] Dec 01 15:00:24 crc kubenswrapper[4810]: I1201 15:00:24.500310 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ef9fa82-f302-47e9-bed3-b02f34a2340f" path="/var/lib/kubelet/pods/1ef9fa82-f302-47e9-bed3-b02f34a2340f/volumes" Dec 01 15:00:24 crc kubenswrapper[4810]: I1201 15:00:24.501025 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="476b71ac-1050-46a6-b51c-168e75b1a350" path="/var/lib/kubelet/pods/476b71ac-1050-46a6-b51c-168e75b1a350/volumes" Dec 01 15:00:24 crc kubenswrapper[4810]: I1201 15:00:24.501985 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f7b5a25-eed0-4122-a036-093ac7adca22" path="/var/lib/kubelet/pods/4f7b5a25-eed0-4122-a036-093ac7adca22/volumes" Dec 01 15:00:24 crc kubenswrapper[4810]: I1201 15:00:24.502749 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf15ccdf-1e4e-44ab-8ccb-819502935a7a" path="/var/lib/kubelet/pods/bf15ccdf-1e4e-44ab-8ccb-819502935a7a/volumes" Dec 01 15:00:24 crc kubenswrapper[4810]: I1201 15:00:24.504194 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7b06553-ec05-4c6f-82c0-ef0b4dd3005a" path="/var/lib/kubelet/pods/f7b06553-ec05-4c6f-82c0-ef0b4dd3005a/volumes" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.712177 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-82xff"] Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713002 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba5f8a69-22cf-4196-bc89-ffa60eda9303" containerName="neutron-api" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713016 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba5f8a69-22cf-4196-bc89-ffa60eda9303" containerName="neutron-api" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713028 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-updater" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713034 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-updater" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713044 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-replicator" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713053 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-replicator" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713066 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="swift-recon-cron" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713072 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="swift-recon-cron" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713085 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="container-replicator" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713092 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="container-replicator" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713106 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7b06553-ec05-4c6f-82c0-ef0b4dd3005a" containerName="mariadb-account-delete" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713113 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7b06553-ec05-4c6f-82c0-ef0b4dd3005a" containerName="mariadb-account-delete" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713125 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="account-replicator" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713132 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="account-replicator" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713143 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-expirer" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713150 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-expirer" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713161 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf15ccdf-1e4e-44ab-8ccb-819502935a7a" containerName="mariadb-account-delete" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713167 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf15ccdf-1e4e-44ab-8ccb-819502935a7a" containerName="mariadb-account-delete" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713174 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovs-vswitchd" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713180 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovs-vswitchd" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713187 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="476b71ac-1050-46a6-b51c-168e75b1a350" containerName="mariadb-account-delete" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713193 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="476b71ac-1050-46a6-b51c-168e75b1a350" containerName="mariadb-account-delete" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713202 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="container-auditor" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713207 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="container-auditor" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713213 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="account-server" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713219 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="account-server" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713225 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-server" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713230 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-server" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713241 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ef9fa82-f302-47e9-bed3-b02f34a2340f" containerName="mariadb-account-delete" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713247 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ef9fa82-f302-47e9-bed3-b02f34a2340f" containerName="mariadb-account-delete" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713257 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec7ca8a4-04cb-4b40-be20-6f2d872f8b25" containerName="collect-profiles" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713263 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec7ca8a4-04cb-4b40-be20-6f2d872f8b25" containerName="collect-profiles" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713270 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="container-server" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713276 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="container-server" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713289 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovsdb-server" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713295 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovsdb-server" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713304 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="account-auditor" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713310 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="account-auditor" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713340 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f7b5a25-eed0-4122-a036-093ac7adca22" containerName="mariadb-account-delete" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713347 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f7b5a25-eed0-4122-a036-093ac7adca22" containerName="mariadb-account-delete" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713354 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="rsync" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713360 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="rsync" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713370 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-auditor" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713376 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-auditor" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713383 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovsdb-server-init" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713389 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovsdb-server-init" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713397 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="container-updater" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713407 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="container-updater" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713421 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="account-reaper" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713428 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="account-reaper" Dec 01 15:00:29 crc kubenswrapper[4810]: E1201 15:00:29.713440 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba5f8a69-22cf-4196-bc89-ffa60eda9303" containerName="neutron-httpd" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713448 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba5f8a69-22cf-4196-bc89-ffa60eda9303" containerName="neutron-httpd" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713642 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="container-updater" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713661 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ef9fa82-f302-47e9-bed3-b02f34a2340f" containerName="mariadb-account-delete" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713673 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="container-auditor" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713686 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-server" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713698 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba5f8a69-22cf-4196-bc89-ffa60eda9303" containerName="neutron-api" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713709 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-expirer" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713716 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="account-replicator" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713724 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovsdb-server" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713734 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf15ccdf-1e4e-44ab-8ccb-819502935a7a" containerName="mariadb-account-delete" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713743 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c699a0a-f268-419b-a829-a9ebaa52b322" containerName="ovs-vswitchd" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713752 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-updater" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713762 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f7b5a25-eed0-4122-a036-093ac7adca22" containerName="mariadb-account-delete" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713776 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="account-auditor" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713786 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="container-server" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713793 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba5f8a69-22cf-4196-bc89-ffa60eda9303" containerName="neutron-httpd" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713803 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="container-replicator" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713816 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="rsync" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713827 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="account-server" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713837 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="account-reaper" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713845 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7b06553-ec05-4c6f-82c0-ef0b4dd3005a" containerName="mariadb-account-delete" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713856 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-auditor" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713867 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec7ca8a4-04cb-4b40-be20-6f2d872f8b25" containerName="collect-profiles" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713886 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="swift-recon-cron" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713896 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f455986-23c9-4088-9829-51c74e5b7770" containerName="object-replicator" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.713906 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="476b71ac-1050-46a6-b51c-168e75b1a350" containerName="mariadb-account-delete" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.715069 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-82xff" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.734787 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-82xff"] Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.898093 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5a7c812-70e4-4e28-9005-05411392f7bd-catalog-content\") pod \"community-operators-82xff\" (UID: \"a5a7c812-70e4-4e28-9005-05411392f7bd\") " pod="openshift-marketplace/community-operators-82xff" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.898159 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5a7c812-70e4-4e28-9005-05411392f7bd-utilities\") pod \"community-operators-82xff\" (UID: \"a5a7c812-70e4-4e28-9005-05411392f7bd\") " pod="openshift-marketplace/community-operators-82xff" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.898199 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljb82\" (UniqueName: \"kubernetes.io/projected/a5a7c812-70e4-4e28-9005-05411392f7bd-kube-api-access-ljb82\") pod \"community-operators-82xff\" (UID: \"a5a7c812-70e4-4e28-9005-05411392f7bd\") " pod="openshift-marketplace/community-operators-82xff" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.999439 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljb82\" (UniqueName: \"kubernetes.io/projected/a5a7c812-70e4-4e28-9005-05411392f7bd-kube-api-access-ljb82\") pod \"community-operators-82xff\" (UID: \"a5a7c812-70e4-4e28-9005-05411392f7bd\") " pod="openshift-marketplace/community-operators-82xff" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.999545 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5a7c812-70e4-4e28-9005-05411392f7bd-catalog-content\") pod \"community-operators-82xff\" (UID: \"a5a7c812-70e4-4e28-9005-05411392f7bd\") " pod="openshift-marketplace/community-operators-82xff" Dec 01 15:00:29 crc kubenswrapper[4810]: I1201 15:00:29.999591 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5a7c812-70e4-4e28-9005-05411392f7bd-utilities\") pod \"community-operators-82xff\" (UID: \"a5a7c812-70e4-4e28-9005-05411392f7bd\") " pod="openshift-marketplace/community-operators-82xff" Dec 01 15:00:30 crc kubenswrapper[4810]: I1201 15:00:30.000135 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5a7c812-70e4-4e28-9005-05411392f7bd-utilities\") pod \"community-operators-82xff\" (UID: \"a5a7c812-70e4-4e28-9005-05411392f7bd\") " pod="openshift-marketplace/community-operators-82xff" Dec 01 15:00:30 crc kubenswrapper[4810]: I1201 15:00:30.000171 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5a7c812-70e4-4e28-9005-05411392f7bd-catalog-content\") pod \"community-operators-82xff\" (UID: \"a5a7c812-70e4-4e28-9005-05411392f7bd\") " pod="openshift-marketplace/community-operators-82xff" Dec 01 15:00:30 crc kubenswrapper[4810]: I1201 15:00:30.027836 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljb82\" (UniqueName: \"kubernetes.io/projected/a5a7c812-70e4-4e28-9005-05411392f7bd-kube-api-access-ljb82\") pod \"community-operators-82xff\" (UID: \"a5a7c812-70e4-4e28-9005-05411392f7bd\") " pod="openshift-marketplace/community-operators-82xff" Dec 01 15:00:30 crc kubenswrapper[4810]: I1201 15:00:30.076813 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-82xff" Dec 01 15:00:30 crc kubenswrapper[4810]: I1201 15:00:30.546950 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-82xff"] Dec 01 15:00:31 crc kubenswrapper[4810]: I1201 15:00:31.190808 4810 generic.go:334] "Generic (PLEG): container finished" podID="a5a7c812-70e4-4e28-9005-05411392f7bd" containerID="de0974f91f1ac25ac4d69444f095b0a2e9b683c87668c8344dc3de84dc830f34" exitCode=0 Dec 01 15:00:31 crc kubenswrapper[4810]: I1201 15:00:31.190902 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82xff" event={"ID":"a5a7c812-70e4-4e28-9005-05411392f7bd","Type":"ContainerDied","Data":"de0974f91f1ac25ac4d69444f095b0a2e9b683c87668c8344dc3de84dc830f34"} Dec 01 15:00:31 crc kubenswrapper[4810]: I1201 15:00:31.191108 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82xff" event={"ID":"a5a7c812-70e4-4e28-9005-05411392f7bd","Type":"ContainerStarted","Data":"7493c891a90dc109a3e5d8add7b351ad12ef5abaa074280457712ef4c2965901"} Dec 01 15:00:34 crc kubenswrapper[4810]: I1201 15:00:34.216993 4810 generic.go:334] "Generic (PLEG): container finished" podID="a5a7c812-70e4-4e28-9005-05411392f7bd" containerID="6eb52e35afd975a271f5760f7b3e7b152e5ec62ef48a0a67458712c0e268a301" exitCode=0 Dec 01 15:00:34 crc kubenswrapper[4810]: I1201 15:00:34.217202 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82xff" event={"ID":"a5a7c812-70e4-4e28-9005-05411392f7bd","Type":"ContainerDied","Data":"6eb52e35afd975a271f5760f7b3e7b152e5ec62ef48a0a67458712c0e268a301"} Dec 01 15:00:36 crc kubenswrapper[4810]: I1201 15:00:36.233924 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82xff" event={"ID":"a5a7c812-70e4-4e28-9005-05411392f7bd","Type":"ContainerStarted","Data":"accd14c1b25fd31b149748588a4863d00f6e6607e66071479af23ce31453aa9e"} Dec 01 15:00:36 crc kubenswrapper[4810]: I1201 15:00:36.252674 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-82xff" podStartSLOduration=3.213244699 podStartE2EDuration="7.252655255s" podCreationTimestamp="2025-12-01 15:00:29 +0000 UTC" firstStartedPulling="2025-12-01 15:00:31.192252567 +0000 UTC m=+1596.955762170" lastFinishedPulling="2025-12-01 15:00:35.231663123 +0000 UTC m=+1600.995172726" observedRunningTime="2025-12-01 15:00:36.251078332 +0000 UTC m=+1602.014587935" watchObservedRunningTime="2025-12-01 15:00:36.252655255 +0000 UTC m=+1602.016164858" Dec 01 15:00:40 crc kubenswrapper[4810]: I1201 15:00:40.076948 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-82xff" Dec 01 15:00:40 crc kubenswrapper[4810]: I1201 15:00:40.077604 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-82xff" Dec 01 15:00:40 crc kubenswrapper[4810]: I1201 15:00:40.126139 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-82xff" Dec 01 15:00:40 crc kubenswrapper[4810]: I1201 15:00:40.303849 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-82xff" Dec 01 15:00:40 crc kubenswrapper[4810]: I1201 15:00:40.355939 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-82xff"] Dec 01 15:00:42 crc kubenswrapper[4810]: I1201 15:00:42.281448 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-82xff" podUID="a5a7c812-70e4-4e28-9005-05411392f7bd" containerName="registry-server" containerID="cri-o://accd14c1b25fd31b149748588a4863d00f6e6607e66071479af23ce31453aa9e" gracePeriod=2 Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.164461 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-82xff" Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.287123 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5a7c812-70e4-4e28-9005-05411392f7bd-utilities\") pod \"a5a7c812-70e4-4e28-9005-05411392f7bd\" (UID: \"a5a7c812-70e4-4e28-9005-05411392f7bd\") " Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.287330 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljb82\" (UniqueName: \"kubernetes.io/projected/a5a7c812-70e4-4e28-9005-05411392f7bd-kube-api-access-ljb82\") pod \"a5a7c812-70e4-4e28-9005-05411392f7bd\" (UID: \"a5a7c812-70e4-4e28-9005-05411392f7bd\") " Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.287422 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5a7c812-70e4-4e28-9005-05411392f7bd-catalog-content\") pod \"a5a7c812-70e4-4e28-9005-05411392f7bd\" (UID: \"a5a7c812-70e4-4e28-9005-05411392f7bd\") " Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.288721 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5a7c812-70e4-4e28-9005-05411392f7bd-utilities" (OuterVolumeSpecName: "utilities") pod "a5a7c812-70e4-4e28-9005-05411392f7bd" (UID: "a5a7c812-70e4-4e28-9005-05411392f7bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.296830 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5a7c812-70e4-4e28-9005-05411392f7bd-kube-api-access-ljb82" (OuterVolumeSpecName: "kube-api-access-ljb82") pod "a5a7c812-70e4-4e28-9005-05411392f7bd" (UID: "a5a7c812-70e4-4e28-9005-05411392f7bd"). InnerVolumeSpecName "kube-api-access-ljb82". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.310712 4810 generic.go:334] "Generic (PLEG): container finished" podID="a5a7c812-70e4-4e28-9005-05411392f7bd" containerID="accd14c1b25fd31b149748588a4863d00f6e6607e66071479af23ce31453aa9e" exitCode=0 Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.310763 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82xff" event={"ID":"a5a7c812-70e4-4e28-9005-05411392f7bd","Type":"ContainerDied","Data":"accd14c1b25fd31b149748588a4863d00f6e6607e66071479af23ce31453aa9e"} Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.310791 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-82xff" event={"ID":"a5a7c812-70e4-4e28-9005-05411392f7bd","Type":"ContainerDied","Data":"7493c891a90dc109a3e5d8add7b351ad12ef5abaa074280457712ef4c2965901"} Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.310816 4810 scope.go:117] "RemoveContainer" containerID="accd14c1b25fd31b149748588a4863d00f6e6607e66071479af23ce31453aa9e" Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.310993 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-82xff" Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.380678 4810 scope.go:117] "RemoveContainer" containerID="6eb52e35afd975a271f5760f7b3e7b152e5ec62ef48a0a67458712c0e268a301" Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.391498 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5a7c812-70e4-4e28-9005-05411392f7bd-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.391522 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljb82\" (UniqueName: \"kubernetes.io/projected/a5a7c812-70e4-4e28-9005-05411392f7bd-kube-api-access-ljb82\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.406727 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5a7c812-70e4-4e28-9005-05411392f7bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5a7c812-70e4-4e28-9005-05411392f7bd" (UID: "a5a7c812-70e4-4e28-9005-05411392f7bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.436661 4810 scope.go:117] "RemoveContainer" containerID="de0974f91f1ac25ac4d69444f095b0a2e9b683c87668c8344dc3de84dc830f34" Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.492279 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5a7c812-70e4-4e28-9005-05411392f7bd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.496235 4810 scope.go:117] "RemoveContainer" containerID="accd14c1b25fd31b149748588a4863d00f6e6607e66071479af23ce31453aa9e" Dec 01 15:00:43 crc kubenswrapper[4810]: E1201 15:00:43.499171 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"accd14c1b25fd31b149748588a4863d00f6e6607e66071479af23ce31453aa9e\": container with ID starting with accd14c1b25fd31b149748588a4863d00f6e6607e66071479af23ce31453aa9e not found: ID does not exist" containerID="accd14c1b25fd31b149748588a4863d00f6e6607e66071479af23ce31453aa9e" Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.499206 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"accd14c1b25fd31b149748588a4863d00f6e6607e66071479af23ce31453aa9e"} err="failed to get container status \"accd14c1b25fd31b149748588a4863d00f6e6607e66071479af23ce31453aa9e\": rpc error: code = NotFound desc = could not find container \"accd14c1b25fd31b149748588a4863d00f6e6607e66071479af23ce31453aa9e\": container with ID starting with accd14c1b25fd31b149748588a4863d00f6e6607e66071479af23ce31453aa9e not found: ID does not exist" Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.499229 4810 scope.go:117] "RemoveContainer" containerID="6eb52e35afd975a271f5760f7b3e7b152e5ec62ef48a0a67458712c0e268a301" Dec 01 15:00:43 crc kubenswrapper[4810]: E1201 15:00:43.499461 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6eb52e35afd975a271f5760f7b3e7b152e5ec62ef48a0a67458712c0e268a301\": container with ID starting with 6eb52e35afd975a271f5760f7b3e7b152e5ec62ef48a0a67458712c0e268a301 not found: ID does not exist" containerID="6eb52e35afd975a271f5760f7b3e7b152e5ec62ef48a0a67458712c0e268a301" Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.499496 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6eb52e35afd975a271f5760f7b3e7b152e5ec62ef48a0a67458712c0e268a301"} err="failed to get container status \"6eb52e35afd975a271f5760f7b3e7b152e5ec62ef48a0a67458712c0e268a301\": rpc error: code = NotFound desc = could not find container \"6eb52e35afd975a271f5760f7b3e7b152e5ec62ef48a0a67458712c0e268a301\": container with ID starting with 6eb52e35afd975a271f5760f7b3e7b152e5ec62ef48a0a67458712c0e268a301 not found: ID does not exist" Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.499509 4810 scope.go:117] "RemoveContainer" containerID="de0974f91f1ac25ac4d69444f095b0a2e9b683c87668c8344dc3de84dc830f34" Dec 01 15:00:43 crc kubenswrapper[4810]: E1201 15:00:43.499981 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de0974f91f1ac25ac4d69444f095b0a2e9b683c87668c8344dc3de84dc830f34\": container with ID starting with de0974f91f1ac25ac4d69444f095b0a2e9b683c87668c8344dc3de84dc830f34 not found: ID does not exist" containerID="de0974f91f1ac25ac4d69444f095b0a2e9b683c87668c8344dc3de84dc830f34" Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.500001 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de0974f91f1ac25ac4d69444f095b0a2e9b683c87668c8344dc3de84dc830f34"} err="failed to get container status \"de0974f91f1ac25ac4d69444f095b0a2e9b683c87668c8344dc3de84dc830f34\": rpc error: code = NotFound desc = could not find container \"de0974f91f1ac25ac4d69444f095b0a2e9b683c87668c8344dc3de84dc830f34\": container with ID starting with de0974f91f1ac25ac4d69444f095b0a2e9b683c87668c8344dc3de84dc830f34 not found: ID does not exist" Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.637483 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-82xff"] Dec 01 15:00:43 crc kubenswrapper[4810]: I1201 15:00:43.642881 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-82xff"] Dec 01 15:00:44 crc kubenswrapper[4810]: I1201 15:00:44.500827 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5a7c812-70e4-4e28-9005-05411392f7bd" path="/var/lib/kubelet/pods/a5a7c812-70e4-4e28-9005-05411392f7bd/volumes" Dec 01 15:01:01 crc kubenswrapper[4810]: I1201 15:01:01.247887 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dchkm"] Dec 01 15:01:01 crc kubenswrapper[4810]: E1201 15:01:01.249057 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5a7c812-70e4-4e28-9005-05411392f7bd" containerName="extract-content" Dec 01 15:01:01 crc kubenswrapper[4810]: I1201 15:01:01.249082 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5a7c812-70e4-4e28-9005-05411392f7bd" containerName="extract-content" Dec 01 15:01:01 crc kubenswrapper[4810]: E1201 15:01:01.249101 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5a7c812-70e4-4e28-9005-05411392f7bd" containerName="extract-utilities" Dec 01 15:01:01 crc kubenswrapper[4810]: I1201 15:01:01.249110 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5a7c812-70e4-4e28-9005-05411392f7bd" containerName="extract-utilities" Dec 01 15:01:01 crc kubenswrapper[4810]: E1201 15:01:01.249145 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5a7c812-70e4-4e28-9005-05411392f7bd" containerName="registry-server" Dec 01 15:01:01 crc kubenswrapper[4810]: I1201 15:01:01.249153 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5a7c812-70e4-4e28-9005-05411392f7bd" containerName="registry-server" Dec 01 15:01:01 crc kubenswrapper[4810]: I1201 15:01:01.249317 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5a7c812-70e4-4e28-9005-05411392f7bd" containerName="registry-server" Dec 01 15:01:01 crc kubenswrapper[4810]: I1201 15:01:01.250668 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dchkm" Dec 01 15:01:01 crc kubenswrapper[4810]: I1201 15:01:01.262828 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dchkm"] Dec 01 15:01:01 crc kubenswrapper[4810]: I1201 15:01:01.312542 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b54b60a-4f86-4c96-a7b5-2598e5284158-catalog-content\") pod \"certified-operators-dchkm\" (UID: \"5b54b60a-4f86-4c96-a7b5-2598e5284158\") " pod="openshift-marketplace/certified-operators-dchkm" Dec 01 15:01:01 crc kubenswrapper[4810]: I1201 15:01:01.312623 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b54b60a-4f86-4c96-a7b5-2598e5284158-utilities\") pod \"certified-operators-dchkm\" (UID: \"5b54b60a-4f86-4c96-a7b5-2598e5284158\") " pod="openshift-marketplace/certified-operators-dchkm" Dec 01 15:01:01 crc kubenswrapper[4810]: I1201 15:01:01.312930 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8z8l\" (UniqueName: \"kubernetes.io/projected/5b54b60a-4f86-4c96-a7b5-2598e5284158-kube-api-access-k8z8l\") pod \"certified-operators-dchkm\" (UID: \"5b54b60a-4f86-4c96-a7b5-2598e5284158\") " pod="openshift-marketplace/certified-operators-dchkm" Dec 01 15:01:01 crc kubenswrapper[4810]: I1201 15:01:01.413630 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8z8l\" (UniqueName: \"kubernetes.io/projected/5b54b60a-4f86-4c96-a7b5-2598e5284158-kube-api-access-k8z8l\") pod \"certified-operators-dchkm\" (UID: \"5b54b60a-4f86-4c96-a7b5-2598e5284158\") " pod="openshift-marketplace/certified-operators-dchkm" Dec 01 15:01:01 crc kubenswrapper[4810]: I1201 15:01:01.413696 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b54b60a-4f86-4c96-a7b5-2598e5284158-catalog-content\") pod \"certified-operators-dchkm\" (UID: \"5b54b60a-4f86-4c96-a7b5-2598e5284158\") " pod="openshift-marketplace/certified-operators-dchkm" Dec 01 15:01:01 crc kubenswrapper[4810]: I1201 15:01:01.413737 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b54b60a-4f86-4c96-a7b5-2598e5284158-utilities\") pod \"certified-operators-dchkm\" (UID: \"5b54b60a-4f86-4c96-a7b5-2598e5284158\") " pod="openshift-marketplace/certified-operators-dchkm" Dec 01 15:01:01 crc kubenswrapper[4810]: I1201 15:01:01.414314 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b54b60a-4f86-4c96-a7b5-2598e5284158-utilities\") pod \"certified-operators-dchkm\" (UID: \"5b54b60a-4f86-4c96-a7b5-2598e5284158\") " pod="openshift-marketplace/certified-operators-dchkm" Dec 01 15:01:01 crc kubenswrapper[4810]: I1201 15:01:01.414314 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b54b60a-4f86-4c96-a7b5-2598e5284158-catalog-content\") pod \"certified-operators-dchkm\" (UID: \"5b54b60a-4f86-4c96-a7b5-2598e5284158\") " pod="openshift-marketplace/certified-operators-dchkm" Dec 01 15:01:01 crc kubenswrapper[4810]: I1201 15:01:01.437334 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8z8l\" (UniqueName: \"kubernetes.io/projected/5b54b60a-4f86-4c96-a7b5-2598e5284158-kube-api-access-k8z8l\") pod \"certified-operators-dchkm\" (UID: \"5b54b60a-4f86-4c96-a7b5-2598e5284158\") " pod="openshift-marketplace/certified-operators-dchkm" Dec 01 15:01:01 crc kubenswrapper[4810]: I1201 15:01:01.582240 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dchkm" Dec 01 15:01:01 crc kubenswrapper[4810]: I1201 15:01:01.891267 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dchkm"] Dec 01 15:01:02 crc kubenswrapper[4810]: I1201 15:01:02.467457 4810 generic.go:334] "Generic (PLEG): container finished" podID="5b54b60a-4f86-4c96-a7b5-2598e5284158" containerID="9c2747f4be35558f4f3790f10cb55aa3a0fb67206eb2a992199b24ddd604cf0c" exitCode=0 Dec 01 15:01:02 crc kubenswrapper[4810]: I1201 15:01:02.467568 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dchkm" event={"ID":"5b54b60a-4f86-4c96-a7b5-2598e5284158","Type":"ContainerDied","Data":"9c2747f4be35558f4f3790f10cb55aa3a0fb67206eb2a992199b24ddd604cf0c"} Dec 01 15:01:02 crc kubenswrapper[4810]: I1201 15:01:02.467830 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dchkm" event={"ID":"5b54b60a-4f86-4c96-a7b5-2598e5284158","Type":"ContainerStarted","Data":"315761bc8c137cd9d8d8cc1b97951906782c87fc9293606e4287e3667faea116"} Dec 01 15:01:02 crc kubenswrapper[4810]: I1201 15:01:02.972643 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:01:02 crc kubenswrapper[4810]: I1201 15:01:02.972725 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:01:04 crc kubenswrapper[4810]: I1201 15:01:04.488029 4810 generic.go:334] "Generic (PLEG): container finished" podID="5b54b60a-4f86-4c96-a7b5-2598e5284158" containerID="684adb74376a9a6868489066381187ef58a5de662ef42a7a45ac0d5aa6bc1f07" exitCode=0 Dec 01 15:01:04 crc kubenswrapper[4810]: I1201 15:01:04.488108 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dchkm" event={"ID":"5b54b60a-4f86-4c96-a7b5-2598e5284158","Type":"ContainerDied","Data":"684adb74376a9a6868489066381187ef58a5de662ef42a7a45ac0d5aa6bc1f07"} Dec 01 15:01:05 crc kubenswrapper[4810]: I1201 15:01:05.721680 4810 scope.go:117] "RemoveContainer" containerID="9870c1b8a5276fda6d94cdc32f4bc57664dfbe1d025f8da26d8eb70045b38620" Dec 01 15:01:05 crc kubenswrapper[4810]: I1201 15:01:05.740673 4810 scope.go:117] "RemoveContainer" containerID="6f00749a221c1f7e6ab22d685c419d81ad12d39e73efcda6cf4bc60436e158e9" Dec 01 15:01:05 crc kubenswrapper[4810]: I1201 15:01:05.778756 4810 scope.go:117] "RemoveContainer" containerID="8ebd6744e6dfcf18f3444cf3cf726b084641649414a0ce03fbfd64313fce06e2" Dec 01 15:01:05 crc kubenswrapper[4810]: I1201 15:01:05.812447 4810 scope.go:117] "RemoveContainer" containerID="1d0d48e84f1be3cb4796f921b940c2af50fbf54d7bc7bc24d59d735872cb1f8c" Dec 01 15:01:05 crc kubenswrapper[4810]: I1201 15:01:05.857953 4810 scope.go:117] "RemoveContainer" containerID="6a11bca1051282741fc502516c5d8575301311334a580a73092fab993519d600" Dec 01 15:01:05 crc kubenswrapper[4810]: I1201 15:01:05.890014 4810 scope.go:117] "RemoveContainer" containerID="200b93a61e8f7c29987fcdbc9f71328c297d619526d29ff4810c1a71a69f6e6c" Dec 01 15:01:05 crc kubenswrapper[4810]: I1201 15:01:05.913097 4810 scope.go:117] "RemoveContainer" containerID="7995ec5ad10e6bed11e113536d305833e1e514a42699e70c76cc0ce45be3bbc2" Dec 01 15:01:05 crc kubenswrapper[4810]: I1201 15:01:05.929217 4810 scope.go:117] "RemoveContainer" containerID="4733dd6747571f5dcdd5fb240a5a3752d303ca44a559a0da9797ccf77d32d51b" Dec 01 15:01:05 crc kubenswrapper[4810]: I1201 15:01:05.947326 4810 scope.go:117] "RemoveContainer" containerID="20908c4e323538658a97e4515a4c9cfc48a3249e1229339681a663eb84fee0df" Dec 01 15:01:05 crc kubenswrapper[4810]: I1201 15:01:05.964802 4810 scope.go:117] "RemoveContainer" containerID="87b827b5dea17db7f9051c490e22381eb54e6dc308ffffd4d8f479968b5b8aef" Dec 01 15:01:05 crc kubenswrapper[4810]: I1201 15:01:05.991329 4810 scope.go:117] "RemoveContainer" containerID="dbf605ce59fdfc665d554ecf7dd8225a9a123f40c558fd3683c4706be9b2e0a8" Dec 01 15:01:06 crc kubenswrapper[4810]: I1201 15:01:06.020388 4810 scope.go:117] "RemoveContainer" containerID="baa21b2e3fe7443b704b8d1a9ef87496a23321efcab3a8630d626d9398fdef10" Dec 01 15:01:06 crc kubenswrapper[4810]: I1201 15:01:06.040966 4810 scope.go:117] "RemoveContainer" containerID="8483ad5843c1db556402790a55a9fc7d801088aacf9c6a0cb8f53cc0e5bd13b9" Dec 01 15:01:06 crc kubenswrapper[4810]: I1201 15:01:06.057066 4810 scope.go:117] "RemoveContainer" containerID="4b9246edd94b5a8cf6a041a9b8421b6220f01030cf4806b8da8546eafef406a4" Dec 01 15:01:06 crc kubenswrapper[4810]: I1201 15:01:06.081945 4810 scope.go:117] "RemoveContainer" containerID="f21f812c44b7972a5f6ecdd8dbe2ffd46fb63a8fa481dd14afa8b978ba21edfb" Dec 01 15:01:06 crc kubenswrapper[4810]: I1201 15:01:06.121954 4810 scope.go:117] "RemoveContainer" containerID="9e960dff62c094d4b5bb80a9748389f999d9e2812e3fcfb9bc2930885059fd81" Dec 01 15:01:06 crc kubenswrapper[4810]: I1201 15:01:06.145973 4810 scope.go:117] "RemoveContainer" containerID="8fbfb2930b964fb785ed1a3925eda27b84fe588e38fc85afb7e23742343427af" Dec 01 15:01:06 crc kubenswrapper[4810]: I1201 15:01:06.514905 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dchkm" event={"ID":"5b54b60a-4f86-4c96-a7b5-2598e5284158","Type":"ContainerStarted","Data":"e06f56e0fa891a18a2b7bc5c7e9e4652a8f1939e332bda3db221aad4b1977a2b"} Dec 01 15:01:06 crc kubenswrapper[4810]: I1201 15:01:06.554409 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dchkm" podStartSLOduration=1.941580232 podStartE2EDuration="5.554386976s" podCreationTimestamp="2025-12-01 15:01:01 +0000 UTC" firstStartedPulling="2025-12-01 15:01:02.469321359 +0000 UTC m=+1628.232830962" lastFinishedPulling="2025-12-01 15:01:06.082128103 +0000 UTC m=+1631.845637706" observedRunningTime="2025-12-01 15:01:06.551396904 +0000 UTC m=+1632.314906507" watchObservedRunningTime="2025-12-01 15:01:06.554386976 +0000 UTC m=+1632.317896579" Dec 01 15:01:11 crc kubenswrapper[4810]: I1201 15:01:11.583015 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dchkm" Dec 01 15:01:11 crc kubenswrapper[4810]: I1201 15:01:11.583576 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dchkm" Dec 01 15:01:11 crc kubenswrapper[4810]: I1201 15:01:11.623095 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dchkm" Dec 01 15:01:12 crc kubenswrapper[4810]: I1201 15:01:12.612288 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dchkm" Dec 01 15:01:12 crc kubenswrapper[4810]: I1201 15:01:12.654249 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dchkm"] Dec 01 15:01:14 crc kubenswrapper[4810]: I1201 15:01:14.582172 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dchkm" podUID="5b54b60a-4f86-4c96-a7b5-2598e5284158" containerName="registry-server" containerID="cri-o://e06f56e0fa891a18a2b7bc5c7e9e4652a8f1939e332bda3db221aad4b1977a2b" gracePeriod=2 Dec 01 15:01:15 crc kubenswrapper[4810]: I1201 15:01:15.590155 4810 generic.go:334] "Generic (PLEG): container finished" podID="5b54b60a-4f86-4c96-a7b5-2598e5284158" containerID="e06f56e0fa891a18a2b7bc5c7e9e4652a8f1939e332bda3db221aad4b1977a2b" exitCode=0 Dec 01 15:01:15 crc kubenswrapper[4810]: I1201 15:01:15.590196 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dchkm" event={"ID":"5b54b60a-4f86-4c96-a7b5-2598e5284158","Type":"ContainerDied","Data":"e06f56e0fa891a18a2b7bc5c7e9e4652a8f1939e332bda3db221aad4b1977a2b"} Dec 01 15:01:16 crc kubenswrapper[4810]: I1201 15:01:16.134416 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dchkm" Dec 01 15:01:16 crc kubenswrapper[4810]: I1201 15:01:16.241519 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b54b60a-4f86-4c96-a7b5-2598e5284158-utilities\") pod \"5b54b60a-4f86-4c96-a7b5-2598e5284158\" (UID: \"5b54b60a-4f86-4c96-a7b5-2598e5284158\") " Dec 01 15:01:16 crc kubenswrapper[4810]: I1201 15:01:16.241711 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8z8l\" (UniqueName: \"kubernetes.io/projected/5b54b60a-4f86-4c96-a7b5-2598e5284158-kube-api-access-k8z8l\") pod \"5b54b60a-4f86-4c96-a7b5-2598e5284158\" (UID: \"5b54b60a-4f86-4c96-a7b5-2598e5284158\") " Dec 01 15:01:16 crc kubenswrapper[4810]: I1201 15:01:16.241744 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b54b60a-4f86-4c96-a7b5-2598e5284158-catalog-content\") pod \"5b54b60a-4f86-4c96-a7b5-2598e5284158\" (UID: \"5b54b60a-4f86-4c96-a7b5-2598e5284158\") " Dec 01 15:01:16 crc kubenswrapper[4810]: I1201 15:01:16.242733 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b54b60a-4f86-4c96-a7b5-2598e5284158-utilities" (OuterVolumeSpecName: "utilities") pod "5b54b60a-4f86-4c96-a7b5-2598e5284158" (UID: "5b54b60a-4f86-4c96-a7b5-2598e5284158"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:01:16 crc kubenswrapper[4810]: I1201 15:01:16.246880 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b54b60a-4f86-4c96-a7b5-2598e5284158-kube-api-access-k8z8l" (OuterVolumeSpecName: "kube-api-access-k8z8l") pod "5b54b60a-4f86-4c96-a7b5-2598e5284158" (UID: "5b54b60a-4f86-4c96-a7b5-2598e5284158"). InnerVolumeSpecName "kube-api-access-k8z8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:01:16 crc kubenswrapper[4810]: I1201 15:01:16.291890 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b54b60a-4f86-4c96-a7b5-2598e5284158-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b54b60a-4f86-4c96-a7b5-2598e5284158" (UID: "5b54b60a-4f86-4c96-a7b5-2598e5284158"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:01:16 crc kubenswrapper[4810]: I1201 15:01:16.342810 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b54b60a-4f86-4c96-a7b5-2598e5284158-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:01:16 crc kubenswrapper[4810]: I1201 15:01:16.342840 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b54b60a-4f86-4c96-a7b5-2598e5284158-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:01:16 crc kubenswrapper[4810]: I1201 15:01:16.342852 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8z8l\" (UniqueName: \"kubernetes.io/projected/5b54b60a-4f86-4c96-a7b5-2598e5284158-kube-api-access-k8z8l\") on node \"crc\" DevicePath \"\"" Dec 01 15:01:16 crc kubenswrapper[4810]: I1201 15:01:16.601757 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dchkm" event={"ID":"5b54b60a-4f86-4c96-a7b5-2598e5284158","Type":"ContainerDied","Data":"315761bc8c137cd9d8d8cc1b97951906782c87fc9293606e4287e3667faea116"} Dec 01 15:01:16 crc kubenswrapper[4810]: I1201 15:01:16.601802 4810 scope.go:117] "RemoveContainer" containerID="e06f56e0fa891a18a2b7bc5c7e9e4652a8f1939e332bda3db221aad4b1977a2b" Dec 01 15:01:16 crc kubenswrapper[4810]: I1201 15:01:16.601828 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dchkm" Dec 01 15:01:16 crc kubenswrapper[4810]: I1201 15:01:16.626420 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dchkm"] Dec 01 15:01:16 crc kubenswrapper[4810]: I1201 15:01:16.631372 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dchkm"] Dec 01 15:01:16 crc kubenswrapper[4810]: I1201 15:01:16.631386 4810 scope.go:117] "RemoveContainer" containerID="684adb74376a9a6868489066381187ef58a5de662ef42a7a45ac0d5aa6bc1f07" Dec 01 15:01:16 crc kubenswrapper[4810]: I1201 15:01:16.650435 4810 scope.go:117] "RemoveContainer" containerID="9c2747f4be35558f4f3790f10cb55aa3a0fb67206eb2a992199b24ddd604cf0c" Dec 01 15:01:18 crc kubenswrapper[4810]: I1201 15:01:18.500285 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b54b60a-4f86-4c96-a7b5-2598e5284158" path="/var/lib/kubelet/pods/5b54b60a-4f86-4c96-a7b5-2598e5284158/volumes" Dec 01 15:01:32 crc kubenswrapper[4810]: I1201 15:01:32.972351 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:01:32 crc kubenswrapper[4810]: I1201 15:01:32.973130 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:02:02 crc kubenswrapper[4810]: I1201 15:02:02.972162 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:02:02 crc kubenswrapper[4810]: I1201 15:02:02.974660 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:02:02 crc kubenswrapper[4810]: I1201 15:02:02.974767 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 15:02:02 crc kubenswrapper[4810]: I1201 15:02:02.975520 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 15:02:02 crc kubenswrapper[4810]: I1201 15:02:02.975576 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" gracePeriod=600 Dec 01 15:02:03 crc kubenswrapper[4810]: E1201 15:02:03.100508 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:02:03 crc kubenswrapper[4810]: I1201 15:02:03.996566 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" exitCode=0 Dec 01 15:02:03 crc kubenswrapper[4810]: I1201 15:02:03.996613 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef"} Dec 01 15:02:03 crc kubenswrapper[4810]: I1201 15:02:03.996645 4810 scope.go:117] "RemoveContainer" containerID="5eb750c936738b66d65230eb7997d67c73bd78542b11814d460ae2c7e8d2d0eb" Dec 01 15:02:03 crc kubenswrapper[4810]: I1201 15:02:03.997323 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:02:03 crc kubenswrapper[4810]: E1201 15:02:03.997754 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:02:06 crc kubenswrapper[4810]: I1201 15:02:06.453206 4810 scope.go:117] "RemoveContainer" containerID="ca5bbd70522206ebd37cdd1720b8e0f4acb738863472b35a2d0739bbc6d28b23" Dec 01 15:02:06 crc kubenswrapper[4810]: I1201 15:02:06.476303 4810 scope.go:117] "RemoveContainer" containerID="28c16aba11f39ca65575b319d525083a1ad5c9dbebbd5e886ea98a39f8b9c23c" Dec 01 15:02:06 crc kubenswrapper[4810]: I1201 15:02:06.504797 4810 scope.go:117] "RemoveContainer" containerID="d9a642dd5e5161d1c3547e4c6fe4ee983dcde299f557d63adfeffa793f5b30c3" Dec 01 15:02:06 crc kubenswrapper[4810]: I1201 15:02:06.544275 4810 scope.go:117] "RemoveContainer" containerID="a40c19049624e61426f700a570719d59ec7a3b73107e4fbc2d3c42482e3793b0" Dec 01 15:02:06 crc kubenswrapper[4810]: I1201 15:02:06.570589 4810 scope.go:117] "RemoveContainer" containerID="375d1e18c5849fe9602a5edfd221a72767570f1b703cd433970ef2374a9b748b" Dec 01 15:02:06 crc kubenswrapper[4810]: I1201 15:02:06.605696 4810 scope.go:117] "RemoveContainer" containerID="877b17a1615addffb19f186b3b5d4ca546eeda45758f240cdd33f5e0fe8a09dc" Dec 01 15:02:06 crc kubenswrapper[4810]: I1201 15:02:06.631209 4810 scope.go:117] "RemoveContainer" containerID="5a5b7334cdb5668aa597f83c43e1f06755c33b97ba720234d356c46686f469f6" Dec 01 15:02:06 crc kubenswrapper[4810]: I1201 15:02:06.651344 4810 scope.go:117] "RemoveContainer" containerID="e16c83319d033bff450700a0e11c3bd1ea8dc45e44e6d30b69b942faeaa93a51" Dec 01 15:02:06 crc kubenswrapper[4810]: I1201 15:02:06.685521 4810 scope.go:117] "RemoveContainer" containerID="e06e409cdd63d5c7f2fe85cf645dd9d82ceaa580d3ea7ea520c851fddaa5d10d" Dec 01 15:02:17 crc kubenswrapper[4810]: I1201 15:02:17.492056 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:02:17 crc kubenswrapper[4810]: E1201 15:02:17.492866 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:02:30 crc kubenswrapper[4810]: I1201 15:02:30.491439 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:02:30 crc kubenswrapper[4810]: E1201 15:02:30.492252 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:02:42 crc kubenswrapper[4810]: I1201 15:02:42.492388 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:02:42 crc kubenswrapper[4810]: E1201 15:02:42.493540 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:02:53 crc kubenswrapper[4810]: I1201 15:02:53.491692 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:02:53 crc kubenswrapper[4810]: E1201 15:02:53.494386 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:03:06 crc kubenswrapper[4810]: I1201 15:03:06.829878 4810 scope.go:117] "RemoveContainer" containerID="a093fc007de4e4dc8464d53799a70b12b4165e5b97136fd03d12179dfe40411d" Dec 01 15:03:06 crc kubenswrapper[4810]: I1201 15:03:06.856040 4810 scope.go:117] "RemoveContainer" containerID="4fa5bae08b56685a2931d6610c251e1527a50f7f4d4f0e483abd703d7161e032" Dec 01 15:03:06 crc kubenswrapper[4810]: I1201 15:03:06.891115 4810 scope.go:117] "RemoveContainer" containerID="a6874557a40d912608216311cf27091cc14caad2e17413b248e220dba30dee7e" Dec 01 15:03:06 crc kubenswrapper[4810]: I1201 15:03:06.915181 4810 scope.go:117] "RemoveContainer" containerID="06071e1b921f7719bea9f6badcd98338a4b9dc1695fc5c83f76c3f50c3c6483d" Dec 01 15:03:06 crc kubenswrapper[4810]: I1201 15:03:06.939116 4810 scope.go:117] "RemoveContainer" containerID="0d35ce8009f60e73fe688ef64e665403a31395b2f3b0dca76964cc0ed35f017b" Dec 01 15:03:06 crc kubenswrapper[4810]: I1201 15:03:06.976394 4810 scope.go:117] "RemoveContainer" containerID="63159094debcf674a5071ab3215f742252e19a2d2a7c8655f727042b942a3233" Dec 01 15:03:06 crc kubenswrapper[4810]: I1201 15:03:06.995565 4810 scope.go:117] "RemoveContainer" containerID="cd15449ff36f15ba79faeb70f743c8df8b5661e48cc14348d7fb12694847e714" Dec 01 15:03:08 crc kubenswrapper[4810]: I1201 15:03:08.491898 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:03:08 crc kubenswrapper[4810]: E1201 15:03:08.492901 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:03:23 crc kubenswrapper[4810]: I1201 15:03:23.490819 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:03:23 crc kubenswrapper[4810]: E1201 15:03:23.491634 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:03:37 crc kubenswrapper[4810]: I1201 15:03:37.491310 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:03:37 crc kubenswrapper[4810]: E1201 15:03:37.492333 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:03:51 crc kubenswrapper[4810]: I1201 15:03:51.490936 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:03:51 crc kubenswrapper[4810]: E1201 15:03:51.491716 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:04:04 crc kubenswrapper[4810]: I1201 15:04:04.494965 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:04:04 crc kubenswrapper[4810]: E1201 15:04:04.496752 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:04:07 crc kubenswrapper[4810]: I1201 15:04:07.123949 4810 scope.go:117] "RemoveContainer" containerID="bcb6d4eef8012d35e7a2581c37d3de047e5d6c95182ecb8fe94d92a03fdbccd4" Dec 01 15:04:07 crc kubenswrapper[4810]: I1201 15:04:07.169569 4810 scope.go:117] "RemoveContainer" containerID="9a9cb3e6a77baeced59069d7dc5d260e53122dd4e38d62c9a1293deed78602d5" Dec 01 15:04:18 crc kubenswrapper[4810]: I1201 15:04:18.491884 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:04:18 crc kubenswrapper[4810]: E1201 15:04:18.493244 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:04:31 crc kubenswrapper[4810]: I1201 15:04:31.491102 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:04:31 crc kubenswrapper[4810]: E1201 15:04:31.491968 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:04:45 crc kubenswrapper[4810]: I1201 15:04:45.491076 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:04:45 crc kubenswrapper[4810]: E1201 15:04:45.491748 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:04:56 crc kubenswrapper[4810]: I1201 15:04:56.490984 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:04:56 crc kubenswrapper[4810]: E1201 15:04:56.491816 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:05:07 crc kubenswrapper[4810]: I1201 15:05:07.238307 4810 scope.go:117] "RemoveContainer" containerID="d94da3e2e43536ab2044032575c579b392f9637451fe336030187a32c1883b3b" Dec 01 15:05:07 crc kubenswrapper[4810]: I1201 15:05:07.276916 4810 scope.go:117] "RemoveContainer" containerID="a15084f9f3b84020b6304d0dab9e5fe86bc99bdcb1432e5ed7e798ae56fba3fd" Dec 01 15:05:07 crc kubenswrapper[4810]: I1201 15:05:07.292105 4810 scope.go:117] "RemoveContainer" containerID="9f977eefdbefbf88215823985ab51d0b45d5ff521f8f4908a3d8f4baf7b8ea03" Dec 01 15:05:07 crc kubenswrapper[4810]: I1201 15:05:07.315657 4810 scope.go:117] "RemoveContainer" containerID="08dc9cd0ed85a4bf549a111095b9209661ba67da14fe12de2f749ed800959244" Dec 01 15:05:07 crc kubenswrapper[4810]: I1201 15:05:07.334671 4810 scope.go:117] "RemoveContainer" containerID="9f4287eaf131c1548b675beb18785d80dde067cda2f647db8236c725db0e0507" Dec 01 15:05:07 crc kubenswrapper[4810]: I1201 15:05:07.378609 4810 scope.go:117] "RemoveContainer" containerID="7f92191f00ad28697cca6ad61e70b41f045bb53297c7b28a21e7c29a04def64d" Dec 01 15:05:09 crc kubenswrapper[4810]: I1201 15:05:09.492539 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:05:09 crc kubenswrapper[4810]: E1201 15:05:09.493280 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:05:24 crc kubenswrapper[4810]: I1201 15:05:24.494817 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:05:24 crc kubenswrapper[4810]: E1201 15:05:24.495568 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:05:37 crc kubenswrapper[4810]: I1201 15:05:37.491034 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:05:37 crc kubenswrapper[4810]: E1201 15:05:37.491902 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:05:50 crc kubenswrapper[4810]: I1201 15:05:50.491987 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:05:50 crc kubenswrapper[4810]: E1201 15:05:50.492845 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:06:03 crc kubenswrapper[4810]: I1201 15:06:03.490903 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:06:03 crc kubenswrapper[4810]: E1201 15:06:03.491689 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:06:07 crc kubenswrapper[4810]: I1201 15:06:07.444330 4810 scope.go:117] "RemoveContainer" containerID="dfcba994df93f276679f0b982a3fda4b0ba3b97518b8e6d0c24012c34d33e149" Dec 01 15:06:07 crc kubenswrapper[4810]: I1201 15:06:07.473915 4810 scope.go:117] "RemoveContainer" containerID="2ba07f4b84a146833c2e8b31369eaa1e961a68fc47ee9b4c60c092208482cff9" Dec 01 15:06:07 crc kubenswrapper[4810]: I1201 15:06:07.492190 4810 scope.go:117] "RemoveContainer" containerID="a54201a119ae4e648033f69b4c2cda9c0c2342bcf83446696e259213e7c43dfe" Dec 01 15:06:07 crc kubenswrapper[4810]: I1201 15:06:07.515653 4810 scope.go:117] "RemoveContainer" containerID="fac78a09d1aaaaf2dc5fd71c00691039693074a0a7b1dee4f7775f2ca6566022" Dec 01 15:06:18 crc kubenswrapper[4810]: I1201 15:06:18.491580 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:06:18 crc kubenswrapper[4810]: E1201 15:06:18.492777 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:06:31 crc kubenswrapper[4810]: I1201 15:06:31.491455 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:06:31 crc kubenswrapper[4810]: E1201 15:06:31.493533 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:06:45 crc kubenswrapper[4810]: I1201 15:06:45.491041 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:06:45 crc kubenswrapper[4810]: E1201 15:06:45.491638 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:07:00 crc kubenswrapper[4810]: I1201 15:07:00.491211 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:07:00 crc kubenswrapper[4810]: E1201 15:07:00.492033 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:07:14 crc kubenswrapper[4810]: I1201 15:07:14.501083 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:07:15 crc kubenswrapper[4810]: I1201 15:07:15.448193 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"cdf338bcebff0a98fe1226fd221bbb927f85e462320f38cb1d0863a7b290222f"} Dec 01 15:08:47 crc kubenswrapper[4810]: I1201 15:08:47.185225 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-67hfc"] Dec 01 15:08:47 crc kubenswrapper[4810]: E1201 15:08:47.186172 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b54b60a-4f86-4c96-a7b5-2598e5284158" containerName="extract-content" Dec 01 15:08:47 crc kubenswrapper[4810]: I1201 15:08:47.186189 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b54b60a-4f86-4c96-a7b5-2598e5284158" containerName="extract-content" Dec 01 15:08:47 crc kubenswrapper[4810]: E1201 15:08:47.186208 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b54b60a-4f86-4c96-a7b5-2598e5284158" containerName="extract-utilities" Dec 01 15:08:47 crc kubenswrapper[4810]: I1201 15:08:47.186216 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b54b60a-4f86-4c96-a7b5-2598e5284158" containerName="extract-utilities" Dec 01 15:08:47 crc kubenswrapper[4810]: E1201 15:08:47.186230 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b54b60a-4f86-4c96-a7b5-2598e5284158" containerName="registry-server" Dec 01 15:08:47 crc kubenswrapper[4810]: I1201 15:08:47.186237 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b54b60a-4f86-4c96-a7b5-2598e5284158" containerName="registry-server" Dec 01 15:08:47 crc kubenswrapper[4810]: I1201 15:08:47.186434 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b54b60a-4f86-4c96-a7b5-2598e5284158" containerName="registry-server" Dec 01 15:08:47 crc kubenswrapper[4810]: I1201 15:08:47.187715 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67hfc" Dec 01 15:08:47 crc kubenswrapper[4810]: I1201 15:08:47.215530 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-67hfc"] Dec 01 15:08:47 crc kubenswrapper[4810]: I1201 15:08:47.227290 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7hbt\" (UniqueName: \"kubernetes.io/projected/27fa22eb-4a8f-44f6-b7c7-17a011422b59-kube-api-access-c7hbt\") pod \"redhat-operators-67hfc\" (UID: \"27fa22eb-4a8f-44f6-b7c7-17a011422b59\") " pod="openshift-marketplace/redhat-operators-67hfc" Dec 01 15:08:47 crc kubenswrapper[4810]: I1201 15:08:47.227344 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27fa22eb-4a8f-44f6-b7c7-17a011422b59-catalog-content\") pod \"redhat-operators-67hfc\" (UID: \"27fa22eb-4a8f-44f6-b7c7-17a011422b59\") " pod="openshift-marketplace/redhat-operators-67hfc" Dec 01 15:08:47 crc kubenswrapper[4810]: I1201 15:08:47.227402 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27fa22eb-4a8f-44f6-b7c7-17a011422b59-utilities\") pod \"redhat-operators-67hfc\" (UID: \"27fa22eb-4a8f-44f6-b7c7-17a011422b59\") " pod="openshift-marketplace/redhat-operators-67hfc" Dec 01 15:08:47 crc kubenswrapper[4810]: I1201 15:08:47.328503 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7hbt\" (UniqueName: \"kubernetes.io/projected/27fa22eb-4a8f-44f6-b7c7-17a011422b59-kube-api-access-c7hbt\") pod \"redhat-operators-67hfc\" (UID: \"27fa22eb-4a8f-44f6-b7c7-17a011422b59\") " pod="openshift-marketplace/redhat-operators-67hfc" Dec 01 15:08:47 crc kubenswrapper[4810]: I1201 15:08:47.328564 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27fa22eb-4a8f-44f6-b7c7-17a011422b59-catalog-content\") pod \"redhat-operators-67hfc\" (UID: \"27fa22eb-4a8f-44f6-b7c7-17a011422b59\") " pod="openshift-marketplace/redhat-operators-67hfc" Dec 01 15:08:47 crc kubenswrapper[4810]: I1201 15:08:47.328617 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27fa22eb-4a8f-44f6-b7c7-17a011422b59-utilities\") pod \"redhat-operators-67hfc\" (UID: \"27fa22eb-4a8f-44f6-b7c7-17a011422b59\") " pod="openshift-marketplace/redhat-operators-67hfc" Dec 01 15:08:47 crc kubenswrapper[4810]: I1201 15:08:47.329126 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27fa22eb-4a8f-44f6-b7c7-17a011422b59-utilities\") pod \"redhat-operators-67hfc\" (UID: \"27fa22eb-4a8f-44f6-b7c7-17a011422b59\") " pod="openshift-marketplace/redhat-operators-67hfc" Dec 01 15:08:47 crc kubenswrapper[4810]: I1201 15:08:47.329203 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27fa22eb-4a8f-44f6-b7c7-17a011422b59-catalog-content\") pod \"redhat-operators-67hfc\" (UID: \"27fa22eb-4a8f-44f6-b7c7-17a011422b59\") " pod="openshift-marketplace/redhat-operators-67hfc" Dec 01 15:08:47 crc kubenswrapper[4810]: I1201 15:08:47.354396 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7hbt\" (UniqueName: \"kubernetes.io/projected/27fa22eb-4a8f-44f6-b7c7-17a011422b59-kube-api-access-c7hbt\") pod \"redhat-operators-67hfc\" (UID: \"27fa22eb-4a8f-44f6-b7c7-17a011422b59\") " pod="openshift-marketplace/redhat-operators-67hfc" Dec 01 15:08:47 crc kubenswrapper[4810]: I1201 15:08:47.505009 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67hfc" Dec 01 15:08:47 crc kubenswrapper[4810]: I1201 15:08:47.755438 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-67hfc"] Dec 01 15:08:48 crc kubenswrapper[4810]: I1201 15:08:48.169698 4810 generic.go:334] "Generic (PLEG): container finished" podID="27fa22eb-4a8f-44f6-b7c7-17a011422b59" containerID="3a1c5739d9de79aaf4270c870d1ca174a4e5aa752ebfb90eea0a62079f76b3a2" exitCode=0 Dec 01 15:08:48 crc kubenswrapper[4810]: I1201 15:08:48.169754 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67hfc" event={"ID":"27fa22eb-4a8f-44f6-b7c7-17a011422b59","Type":"ContainerDied","Data":"3a1c5739d9de79aaf4270c870d1ca174a4e5aa752ebfb90eea0a62079f76b3a2"} Dec 01 15:08:48 crc kubenswrapper[4810]: I1201 15:08:48.169964 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67hfc" event={"ID":"27fa22eb-4a8f-44f6-b7c7-17a011422b59","Type":"ContainerStarted","Data":"f051e4c99c7513b0098dcf81a2bd391149c4638e54f2ecd67a8b11b5bbd5f39b"} Dec 01 15:08:48 crc kubenswrapper[4810]: I1201 15:08:48.171444 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 15:08:50 crc kubenswrapper[4810]: I1201 15:08:50.189971 4810 generic.go:334] "Generic (PLEG): container finished" podID="27fa22eb-4a8f-44f6-b7c7-17a011422b59" containerID="0a254ebe05a9a68b94eaf033f1b6cb5f056f78e810f0744f2b33e69057cebfc9" exitCode=0 Dec 01 15:08:50 crc kubenswrapper[4810]: I1201 15:08:50.190328 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67hfc" event={"ID":"27fa22eb-4a8f-44f6-b7c7-17a011422b59","Type":"ContainerDied","Data":"0a254ebe05a9a68b94eaf033f1b6cb5f056f78e810f0744f2b33e69057cebfc9"} Dec 01 15:08:51 crc kubenswrapper[4810]: I1201 15:08:51.208409 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67hfc" event={"ID":"27fa22eb-4a8f-44f6-b7c7-17a011422b59","Type":"ContainerStarted","Data":"c7e58db874f94054312bdd4a246e4ad4b2ed80d77318a7dd3956a71f77644559"} Dec 01 15:08:51 crc kubenswrapper[4810]: I1201 15:08:51.231948 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-67hfc" podStartSLOduration=1.7118863370000001 podStartE2EDuration="4.231928483s" podCreationTimestamp="2025-12-01 15:08:47 +0000 UTC" firstStartedPulling="2025-12-01 15:08:48.171147243 +0000 UTC m=+2093.934656846" lastFinishedPulling="2025-12-01 15:08:50.691189389 +0000 UTC m=+2096.454698992" observedRunningTime="2025-12-01 15:08:51.229176768 +0000 UTC m=+2096.992686371" watchObservedRunningTime="2025-12-01 15:08:51.231928483 +0000 UTC m=+2096.995438086" Dec 01 15:08:57 crc kubenswrapper[4810]: I1201 15:08:57.505360 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-67hfc" Dec 01 15:08:57 crc kubenswrapper[4810]: I1201 15:08:57.505977 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-67hfc" Dec 01 15:08:57 crc kubenswrapper[4810]: I1201 15:08:57.564698 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-67hfc" Dec 01 15:08:58 crc kubenswrapper[4810]: I1201 15:08:58.299592 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-67hfc" Dec 01 15:08:58 crc kubenswrapper[4810]: I1201 15:08:58.348412 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-67hfc"] Dec 01 15:09:00 crc kubenswrapper[4810]: I1201 15:09:00.271307 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-67hfc" podUID="27fa22eb-4a8f-44f6-b7c7-17a011422b59" containerName="registry-server" containerID="cri-o://c7e58db874f94054312bdd4a246e4ad4b2ed80d77318a7dd3956a71f77644559" gracePeriod=2 Dec 01 15:09:02 crc kubenswrapper[4810]: I1201 15:09:02.288216 4810 generic.go:334] "Generic (PLEG): container finished" podID="27fa22eb-4a8f-44f6-b7c7-17a011422b59" containerID="c7e58db874f94054312bdd4a246e4ad4b2ed80d77318a7dd3956a71f77644559" exitCode=0 Dec 01 15:09:02 crc kubenswrapper[4810]: I1201 15:09:02.288273 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67hfc" event={"ID":"27fa22eb-4a8f-44f6-b7c7-17a011422b59","Type":"ContainerDied","Data":"c7e58db874f94054312bdd4a246e4ad4b2ed80d77318a7dd3956a71f77644559"} Dec 01 15:09:02 crc kubenswrapper[4810]: I1201 15:09:02.484519 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67hfc" Dec 01 15:09:02 crc kubenswrapper[4810]: I1201 15:09:02.535443 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27fa22eb-4a8f-44f6-b7c7-17a011422b59-utilities\") pod \"27fa22eb-4a8f-44f6-b7c7-17a011422b59\" (UID: \"27fa22eb-4a8f-44f6-b7c7-17a011422b59\") " Dec 01 15:09:02 crc kubenswrapper[4810]: I1201 15:09:02.535534 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7hbt\" (UniqueName: \"kubernetes.io/projected/27fa22eb-4a8f-44f6-b7c7-17a011422b59-kube-api-access-c7hbt\") pod \"27fa22eb-4a8f-44f6-b7c7-17a011422b59\" (UID: \"27fa22eb-4a8f-44f6-b7c7-17a011422b59\") " Dec 01 15:09:02 crc kubenswrapper[4810]: I1201 15:09:02.535561 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27fa22eb-4a8f-44f6-b7c7-17a011422b59-catalog-content\") pod \"27fa22eb-4a8f-44f6-b7c7-17a011422b59\" (UID: \"27fa22eb-4a8f-44f6-b7c7-17a011422b59\") " Dec 01 15:09:02 crc kubenswrapper[4810]: I1201 15:09:02.536335 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27fa22eb-4a8f-44f6-b7c7-17a011422b59-utilities" (OuterVolumeSpecName: "utilities") pod "27fa22eb-4a8f-44f6-b7c7-17a011422b59" (UID: "27fa22eb-4a8f-44f6-b7c7-17a011422b59"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:09:02 crc kubenswrapper[4810]: I1201 15:09:02.540543 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27fa22eb-4a8f-44f6-b7c7-17a011422b59-kube-api-access-c7hbt" (OuterVolumeSpecName: "kube-api-access-c7hbt") pod "27fa22eb-4a8f-44f6-b7c7-17a011422b59" (UID: "27fa22eb-4a8f-44f6-b7c7-17a011422b59"). InnerVolumeSpecName "kube-api-access-c7hbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:09:02 crc kubenswrapper[4810]: I1201 15:09:02.637737 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7hbt\" (UniqueName: \"kubernetes.io/projected/27fa22eb-4a8f-44f6-b7c7-17a011422b59-kube-api-access-c7hbt\") on node \"crc\" DevicePath \"\"" Dec 01 15:09:02 crc kubenswrapper[4810]: I1201 15:09:02.637787 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27fa22eb-4a8f-44f6-b7c7-17a011422b59-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:09:02 crc kubenswrapper[4810]: I1201 15:09:02.648797 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27fa22eb-4a8f-44f6-b7c7-17a011422b59-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "27fa22eb-4a8f-44f6-b7c7-17a011422b59" (UID: "27fa22eb-4a8f-44f6-b7c7-17a011422b59"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:09:02 crc kubenswrapper[4810]: I1201 15:09:02.739464 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27fa22eb-4a8f-44f6-b7c7-17a011422b59-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:09:03 crc kubenswrapper[4810]: I1201 15:09:03.297655 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67hfc" event={"ID":"27fa22eb-4a8f-44f6-b7c7-17a011422b59","Type":"ContainerDied","Data":"f051e4c99c7513b0098dcf81a2bd391149c4638e54f2ecd67a8b11b5bbd5f39b"} Dec 01 15:09:03 crc kubenswrapper[4810]: I1201 15:09:03.297705 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67hfc" Dec 01 15:09:03 crc kubenswrapper[4810]: I1201 15:09:03.298027 4810 scope.go:117] "RemoveContainer" containerID="c7e58db874f94054312bdd4a246e4ad4b2ed80d77318a7dd3956a71f77644559" Dec 01 15:09:03 crc kubenswrapper[4810]: I1201 15:09:03.328041 4810 scope.go:117] "RemoveContainer" containerID="0a254ebe05a9a68b94eaf033f1b6cb5f056f78e810f0744f2b33e69057cebfc9" Dec 01 15:09:03 crc kubenswrapper[4810]: I1201 15:09:03.333349 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-67hfc"] Dec 01 15:09:03 crc kubenswrapper[4810]: I1201 15:09:03.340045 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-67hfc"] Dec 01 15:09:03 crc kubenswrapper[4810]: I1201 15:09:03.347172 4810 scope.go:117] "RemoveContainer" containerID="3a1c5739d9de79aaf4270c870d1ca174a4e5aa752ebfb90eea0a62079f76b3a2" Dec 01 15:09:04 crc kubenswrapper[4810]: I1201 15:09:04.503898 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27fa22eb-4a8f-44f6-b7c7-17a011422b59" path="/var/lib/kubelet/pods/27fa22eb-4a8f-44f6-b7c7-17a011422b59/volumes" Dec 01 15:09:27 crc kubenswrapper[4810]: I1201 15:09:27.853919 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4n6sh"] Dec 01 15:09:27 crc kubenswrapper[4810]: E1201 15:09:27.854974 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27fa22eb-4a8f-44f6-b7c7-17a011422b59" containerName="registry-server" Dec 01 15:09:27 crc kubenswrapper[4810]: I1201 15:09:27.854995 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="27fa22eb-4a8f-44f6-b7c7-17a011422b59" containerName="registry-server" Dec 01 15:09:27 crc kubenswrapper[4810]: E1201 15:09:27.855017 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27fa22eb-4a8f-44f6-b7c7-17a011422b59" containerName="extract-utilities" Dec 01 15:09:27 crc kubenswrapper[4810]: I1201 15:09:27.855027 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="27fa22eb-4a8f-44f6-b7c7-17a011422b59" containerName="extract-utilities" Dec 01 15:09:27 crc kubenswrapper[4810]: E1201 15:09:27.855052 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27fa22eb-4a8f-44f6-b7c7-17a011422b59" containerName="extract-content" Dec 01 15:09:27 crc kubenswrapper[4810]: I1201 15:09:27.855061 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="27fa22eb-4a8f-44f6-b7c7-17a011422b59" containerName="extract-content" Dec 01 15:09:27 crc kubenswrapper[4810]: I1201 15:09:27.855307 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="27fa22eb-4a8f-44f6-b7c7-17a011422b59" containerName="registry-server" Dec 01 15:09:27 crc kubenswrapper[4810]: I1201 15:09:27.856590 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4n6sh" Dec 01 15:09:27 crc kubenswrapper[4810]: I1201 15:09:27.869035 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4n6sh"] Dec 01 15:09:27 crc kubenswrapper[4810]: I1201 15:09:27.901607 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740-catalog-content\") pod \"redhat-marketplace-4n6sh\" (UID: \"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740\") " pod="openshift-marketplace/redhat-marketplace-4n6sh" Dec 01 15:09:27 crc kubenswrapper[4810]: I1201 15:09:27.901869 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2dwm\" (UniqueName: \"kubernetes.io/projected/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740-kube-api-access-x2dwm\") pod \"redhat-marketplace-4n6sh\" (UID: \"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740\") " pod="openshift-marketplace/redhat-marketplace-4n6sh" Dec 01 15:09:27 crc kubenswrapper[4810]: I1201 15:09:27.901959 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740-utilities\") pod \"redhat-marketplace-4n6sh\" (UID: \"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740\") " pod="openshift-marketplace/redhat-marketplace-4n6sh" Dec 01 15:09:28 crc kubenswrapper[4810]: I1201 15:09:28.002841 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740-catalog-content\") pod \"redhat-marketplace-4n6sh\" (UID: \"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740\") " pod="openshift-marketplace/redhat-marketplace-4n6sh" Dec 01 15:09:28 crc kubenswrapper[4810]: I1201 15:09:28.002894 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2dwm\" (UniqueName: \"kubernetes.io/projected/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740-kube-api-access-x2dwm\") pod \"redhat-marketplace-4n6sh\" (UID: \"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740\") " pod="openshift-marketplace/redhat-marketplace-4n6sh" Dec 01 15:09:28 crc kubenswrapper[4810]: I1201 15:09:28.002915 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740-utilities\") pod \"redhat-marketplace-4n6sh\" (UID: \"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740\") " pod="openshift-marketplace/redhat-marketplace-4n6sh" Dec 01 15:09:28 crc kubenswrapper[4810]: I1201 15:09:28.003323 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740-utilities\") pod \"redhat-marketplace-4n6sh\" (UID: \"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740\") " pod="openshift-marketplace/redhat-marketplace-4n6sh" Dec 01 15:09:28 crc kubenswrapper[4810]: I1201 15:09:28.003659 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740-catalog-content\") pod \"redhat-marketplace-4n6sh\" (UID: \"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740\") " pod="openshift-marketplace/redhat-marketplace-4n6sh" Dec 01 15:09:28 crc kubenswrapper[4810]: I1201 15:09:28.027214 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2dwm\" (UniqueName: \"kubernetes.io/projected/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740-kube-api-access-x2dwm\") pod \"redhat-marketplace-4n6sh\" (UID: \"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740\") " pod="openshift-marketplace/redhat-marketplace-4n6sh" Dec 01 15:09:28 crc kubenswrapper[4810]: I1201 15:09:28.174294 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4n6sh" Dec 01 15:09:28 crc kubenswrapper[4810]: I1201 15:09:28.627659 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4n6sh"] Dec 01 15:09:29 crc kubenswrapper[4810]: I1201 15:09:29.503692 4810 generic.go:334] "Generic (PLEG): container finished" podID="bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740" containerID="330aef484db2bd27bd43483b156f184ec7727d8660dcc96e7f9d8cb42ea554f8" exitCode=0 Dec 01 15:09:29 crc kubenswrapper[4810]: I1201 15:09:29.503780 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4n6sh" event={"ID":"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740","Type":"ContainerDied","Data":"330aef484db2bd27bd43483b156f184ec7727d8660dcc96e7f9d8cb42ea554f8"} Dec 01 15:09:29 crc kubenswrapper[4810]: I1201 15:09:29.504073 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4n6sh" event={"ID":"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740","Type":"ContainerStarted","Data":"804c61264613152b3dd8525eb3130794f2a889de4c39ee928347b65a8f8ab557"} Dec 01 15:09:31 crc kubenswrapper[4810]: I1201 15:09:31.521035 4810 generic.go:334] "Generic (PLEG): container finished" podID="bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740" containerID="cae9f0181ed6f3c819a274af16099b84fc5429b729a8e05cb1bde79609fa85fc" exitCode=0 Dec 01 15:09:31 crc kubenswrapper[4810]: I1201 15:09:31.521109 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4n6sh" event={"ID":"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740","Type":"ContainerDied","Data":"cae9f0181ed6f3c819a274af16099b84fc5429b729a8e05cb1bde79609fa85fc"} Dec 01 15:09:32 crc kubenswrapper[4810]: I1201 15:09:32.530439 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4n6sh" event={"ID":"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740","Type":"ContainerStarted","Data":"e1c10741fa896a59aae24d6054c61bae7d43586222b6c4858b23eea4c27b1f4a"} Dec 01 15:09:32 crc kubenswrapper[4810]: I1201 15:09:32.547934 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4n6sh" podStartSLOduration=3.00332305 podStartE2EDuration="5.54791245s" podCreationTimestamp="2025-12-01 15:09:27 +0000 UTC" firstStartedPulling="2025-12-01 15:09:29.50666402 +0000 UTC m=+2135.270173663" lastFinishedPulling="2025-12-01 15:09:32.05125346 +0000 UTC m=+2137.814763063" observedRunningTime="2025-12-01 15:09:32.545779563 +0000 UTC m=+2138.309289166" watchObservedRunningTime="2025-12-01 15:09:32.54791245 +0000 UTC m=+2138.311422053" Dec 01 15:09:32 crc kubenswrapper[4810]: I1201 15:09:32.972059 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:09:32 crc kubenswrapper[4810]: I1201 15:09:32.972330 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:09:38 crc kubenswrapper[4810]: I1201 15:09:38.174863 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4n6sh" Dec 01 15:09:38 crc kubenswrapper[4810]: I1201 15:09:38.175234 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4n6sh" Dec 01 15:09:38 crc kubenswrapper[4810]: I1201 15:09:38.236454 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4n6sh" Dec 01 15:09:38 crc kubenswrapper[4810]: I1201 15:09:38.649898 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4n6sh" Dec 01 15:09:38 crc kubenswrapper[4810]: I1201 15:09:38.716362 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4n6sh"] Dec 01 15:09:40 crc kubenswrapper[4810]: I1201 15:09:40.603863 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4n6sh" podUID="bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740" containerName="registry-server" containerID="cri-o://e1c10741fa896a59aae24d6054c61bae7d43586222b6c4858b23eea4c27b1f4a" gracePeriod=2 Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.546441 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4n6sh" Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.611939 4810 generic.go:334] "Generic (PLEG): container finished" podID="bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740" containerID="e1c10741fa896a59aae24d6054c61bae7d43586222b6c4858b23eea4c27b1f4a" exitCode=0 Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.612008 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4n6sh" event={"ID":"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740","Type":"ContainerDied","Data":"e1c10741fa896a59aae24d6054c61bae7d43586222b6c4858b23eea4c27b1f4a"} Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.612049 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4n6sh" event={"ID":"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740","Type":"ContainerDied","Data":"804c61264613152b3dd8525eb3130794f2a889de4c39ee928347b65a8f8ab557"} Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.612103 4810 scope.go:117] "RemoveContainer" containerID="e1c10741fa896a59aae24d6054c61bae7d43586222b6c4858b23eea4c27b1f4a" Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.612288 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4n6sh" Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.629685 4810 scope.go:117] "RemoveContainer" containerID="cae9f0181ed6f3c819a274af16099b84fc5429b729a8e05cb1bde79609fa85fc" Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.643309 4810 scope.go:117] "RemoveContainer" containerID="330aef484db2bd27bd43483b156f184ec7727d8660dcc96e7f9d8cb42ea554f8" Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.666928 4810 scope.go:117] "RemoveContainer" containerID="e1c10741fa896a59aae24d6054c61bae7d43586222b6c4858b23eea4c27b1f4a" Dec 01 15:09:41 crc kubenswrapper[4810]: E1201 15:09:41.667410 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1c10741fa896a59aae24d6054c61bae7d43586222b6c4858b23eea4c27b1f4a\": container with ID starting with e1c10741fa896a59aae24d6054c61bae7d43586222b6c4858b23eea4c27b1f4a not found: ID does not exist" containerID="e1c10741fa896a59aae24d6054c61bae7d43586222b6c4858b23eea4c27b1f4a" Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.667450 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1c10741fa896a59aae24d6054c61bae7d43586222b6c4858b23eea4c27b1f4a"} err="failed to get container status \"e1c10741fa896a59aae24d6054c61bae7d43586222b6c4858b23eea4c27b1f4a\": rpc error: code = NotFound desc = could not find container \"e1c10741fa896a59aae24d6054c61bae7d43586222b6c4858b23eea4c27b1f4a\": container with ID starting with e1c10741fa896a59aae24d6054c61bae7d43586222b6c4858b23eea4c27b1f4a not found: ID does not exist" Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.667491 4810 scope.go:117] "RemoveContainer" containerID="cae9f0181ed6f3c819a274af16099b84fc5429b729a8e05cb1bde79609fa85fc" Dec 01 15:09:41 crc kubenswrapper[4810]: E1201 15:09:41.667889 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cae9f0181ed6f3c819a274af16099b84fc5429b729a8e05cb1bde79609fa85fc\": container with ID starting with cae9f0181ed6f3c819a274af16099b84fc5429b729a8e05cb1bde79609fa85fc not found: ID does not exist" containerID="cae9f0181ed6f3c819a274af16099b84fc5429b729a8e05cb1bde79609fa85fc" Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.667940 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cae9f0181ed6f3c819a274af16099b84fc5429b729a8e05cb1bde79609fa85fc"} err="failed to get container status \"cae9f0181ed6f3c819a274af16099b84fc5429b729a8e05cb1bde79609fa85fc\": rpc error: code = NotFound desc = could not find container \"cae9f0181ed6f3c819a274af16099b84fc5429b729a8e05cb1bde79609fa85fc\": container with ID starting with cae9f0181ed6f3c819a274af16099b84fc5429b729a8e05cb1bde79609fa85fc not found: ID does not exist" Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.667970 4810 scope.go:117] "RemoveContainer" containerID="330aef484db2bd27bd43483b156f184ec7727d8660dcc96e7f9d8cb42ea554f8" Dec 01 15:09:41 crc kubenswrapper[4810]: E1201 15:09:41.668269 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"330aef484db2bd27bd43483b156f184ec7727d8660dcc96e7f9d8cb42ea554f8\": container with ID starting with 330aef484db2bd27bd43483b156f184ec7727d8660dcc96e7f9d8cb42ea554f8 not found: ID does not exist" containerID="330aef484db2bd27bd43483b156f184ec7727d8660dcc96e7f9d8cb42ea554f8" Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.668300 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"330aef484db2bd27bd43483b156f184ec7727d8660dcc96e7f9d8cb42ea554f8"} err="failed to get container status \"330aef484db2bd27bd43483b156f184ec7727d8660dcc96e7f9d8cb42ea554f8\": rpc error: code = NotFound desc = could not find container \"330aef484db2bd27bd43483b156f184ec7727d8660dcc96e7f9d8cb42ea554f8\": container with ID starting with 330aef484db2bd27bd43483b156f184ec7727d8660dcc96e7f9d8cb42ea554f8 not found: ID does not exist" Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.691935 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740-catalog-content\") pod \"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740\" (UID: \"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740\") " Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.691981 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740-utilities\") pod \"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740\" (UID: \"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740\") " Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.692063 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2dwm\" (UniqueName: \"kubernetes.io/projected/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740-kube-api-access-x2dwm\") pod \"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740\" (UID: \"bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740\") " Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.693027 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740-utilities" (OuterVolumeSpecName: "utilities") pod "bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740" (UID: "bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.698539 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740-kube-api-access-x2dwm" (OuterVolumeSpecName: "kube-api-access-x2dwm") pod "bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740" (UID: "bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740"). InnerVolumeSpecName "kube-api-access-x2dwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.713373 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740" (UID: "bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.793955 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2dwm\" (UniqueName: \"kubernetes.io/projected/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740-kube-api-access-x2dwm\") on node \"crc\" DevicePath \"\"" Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.793996 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.794007 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.945992 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4n6sh"] Dec 01 15:09:41 crc kubenswrapper[4810]: I1201 15:09:41.952980 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4n6sh"] Dec 01 15:09:42 crc kubenswrapper[4810]: I1201 15:09:42.501979 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740" path="/var/lib/kubelet/pods/bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740/volumes" Dec 01 15:10:02 crc kubenswrapper[4810]: I1201 15:10:02.972671 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:10:02 crc kubenswrapper[4810]: I1201 15:10:02.974100 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:10:32 crc kubenswrapper[4810]: I1201 15:10:32.971839 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:10:32 crc kubenswrapper[4810]: I1201 15:10:32.972266 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:10:32 crc kubenswrapper[4810]: I1201 15:10:32.972311 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 15:10:32 crc kubenswrapper[4810]: I1201 15:10:32.972870 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cdf338bcebff0a98fe1226fd221bbb927f85e462320f38cb1d0863a7b290222f"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 15:10:32 crc kubenswrapper[4810]: I1201 15:10:32.972920 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://cdf338bcebff0a98fe1226fd221bbb927f85e462320f38cb1d0863a7b290222f" gracePeriod=600 Dec 01 15:10:33 crc kubenswrapper[4810]: I1201 15:10:33.180705 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="cdf338bcebff0a98fe1226fd221bbb927f85e462320f38cb1d0863a7b290222f" exitCode=0 Dec 01 15:10:33 crc kubenswrapper[4810]: I1201 15:10:33.180792 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"cdf338bcebff0a98fe1226fd221bbb927f85e462320f38cb1d0863a7b290222f"} Dec 01 15:10:33 crc kubenswrapper[4810]: I1201 15:10:33.180864 4810 scope.go:117] "RemoveContainer" containerID="5847c925dc3cad77262b2d20d3d4a9074d4135bc2d6ccffb2ba346945c1c0bef" Dec 01 15:10:34 crc kubenswrapper[4810]: I1201 15:10:34.189694 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5"} Dec 01 15:12:22 crc kubenswrapper[4810]: I1201 15:12:22.543269 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zcdww"] Dec 01 15:12:22 crc kubenswrapper[4810]: E1201 15:12:22.544504 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740" containerName="registry-server" Dec 01 15:12:22 crc kubenswrapper[4810]: I1201 15:12:22.544529 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740" containerName="registry-server" Dec 01 15:12:22 crc kubenswrapper[4810]: E1201 15:12:22.544561 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740" containerName="extract-content" Dec 01 15:12:22 crc kubenswrapper[4810]: I1201 15:12:22.544574 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740" containerName="extract-content" Dec 01 15:12:22 crc kubenswrapper[4810]: E1201 15:12:22.544601 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740" containerName="extract-utilities" Dec 01 15:12:22 crc kubenswrapper[4810]: I1201 15:12:22.544613 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740" containerName="extract-utilities" Dec 01 15:12:22 crc kubenswrapper[4810]: I1201 15:12:22.544871 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc4d3f90-a49f-4faf-bc8c-3ce83bbb6740" containerName="registry-server" Dec 01 15:12:22 crc kubenswrapper[4810]: I1201 15:12:22.546972 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zcdww" Dec 01 15:12:22 crc kubenswrapper[4810]: I1201 15:12:22.560627 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0e41ca0-4806-49a1-bb76-55f140d4ad99-utilities\") pod \"certified-operators-zcdww\" (UID: \"d0e41ca0-4806-49a1-bb76-55f140d4ad99\") " pod="openshift-marketplace/certified-operators-zcdww" Dec 01 15:12:22 crc kubenswrapper[4810]: I1201 15:12:22.560696 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8wf2\" (UniqueName: \"kubernetes.io/projected/d0e41ca0-4806-49a1-bb76-55f140d4ad99-kube-api-access-b8wf2\") pod \"certified-operators-zcdww\" (UID: \"d0e41ca0-4806-49a1-bb76-55f140d4ad99\") " pod="openshift-marketplace/certified-operators-zcdww" Dec 01 15:12:22 crc kubenswrapper[4810]: I1201 15:12:22.560739 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0e41ca0-4806-49a1-bb76-55f140d4ad99-catalog-content\") pod \"certified-operators-zcdww\" (UID: \"d0e41ca0-4806-49a1-bb76-55f140d4ad99\") " pod="openshift-marketplace/certified-operators-zcdww" Dec 01 15:12:22 crc kubenswrapper[4810]: I1201 15:12:22.572173 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zcdww"] Dec 01 15:12:22 crc kubenswrapper[4810]: I1201 15:12:22.662709 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0e41ca0-4806-49a1-bb76-55f140d4ad99-utilities\") pod \"certified-operators-zcdww\" (UID: \"d0e41ca0-4806-49a1-bb76-55f140d4ad99\") " pod="openshift-marketplace/certified-operators-zcdww" Dec 01 15:12:22 crc kubenswrapper[4810]: I1201 15:12:22.662755 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8wf2\" (UniqueName: \"kubernetes.io/projected/d0e41ca0-4806-49a1-bb76-55f140d4ad99-kube-api-access-b8wf2\") pod \"certified-operators-zcdww\" (UID: \"d0e41ca0-4806-49a1-bb76-55f140d4ad99\") " pod="openshift-marketplace/certified-operators-zcdww" Dec 01 15:12:22 crc kubenswrapper[4810]: I1201 15:12:22.662783 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0e41ca0-4806-49a1-bb76-55f140d4ad99-catalog-content\") pod \"certified-operators-zcdww\" (UID: \"d0e41ca0-4806-49a1-bb76-55f140d4ad99\") " pod="openshift-marketplace/certified-operators-zcdww" Dec 01 15:12:22 crc kubenswrapper[4810]: I1201 15:12:22.663287 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0e41ca0-4806-49a1-bb76-55f140d4ad99-utilities\") pod \"certified-operators-zcdww\" (UID: \"d0e41ca0-4806-49a1-bb76-55f140d4ad99\") " pod="openshift-marketplace/certified-operators-zcdww" Dec 01 15:12:22 crc kubenswrapper[4810]: I1201 15:12:22.663294 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0e41ca0-4806-49a1-bb76-55f140d4ad99-catalog-content\") pod \"certified-operators-zcdww\" (UID: \"d0e41ca0-4806-49a1-bb76-55f140d4ad99\") " pod="openshift-marketplace/certified-operators-zcdww" Dec 01 15:12:22 crc kubenswrapper[4810]: I1201 15:12:22.687762 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8wf2\" (UniqueName: \"kubernetes.io/projected/d0e41ca0-4806-49a1-bb76-55f140d4ad99-kube-api-access-b8wf2\") pod \"certified-operators-zcdww\" (UID: \"d0e41ca0-4806-49a1-bb76-55f140d4ad99\") " pod="openshift-marketplace/certified-operators-zcdww" Dec 01 15:12:22 crc kubenswrapper[4810]: I1201 15:12:22.887343 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zcdww" Dec 01 15:12:23 crc kubenswrapper[4810]: I1201 15:12:23.361280 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zcdww"] Dec 01 15:12:24 crc kubenswrapper[4810]: I1201 15:12:24.062617 4810 generic.go:334] "Generic (PLEG): container finished" podID="d0e41ca0-4806-49a1-bb76-55f140d4ad99" containerID="1713876d8eea0f3772be1a21fe71192f68eb5984ed846d1fc97a861c12fc7483" exitCode=0 Dec 01 15:12:24 crc kubenswrapper[4810]: I1201 15:12:24.062695 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zcdww" event={"ID":"d0e41ca0-4806-49a1-bb76-55f140d4ad99","Type":"ContainerDied","Data":"1713876d8eea0f3772be1a21fe71192f68eb5984ed846d1fc97a861c12fc7483"} Dec 01 15:12:24 crc kubenswrapper[4810]: I1201 15:12:24.062978 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zcdww" event={"ID":"d0e41ca0-4806-49a1-bb76-55f140d4ad99","Type":"ContainerStarted","Data":"2ddcd44f3d83e5e0caefb41af9690a578544a689fd5f8b5124f3ac049d6cf510"} Dec 01 15:12:30 crc kubenswrapper[4810]: I1201 15:12:30.107592 4810 generic.go:334] "Generic (PLEG): container finished" podID="d0e41ca0-4806-49a1-bb76-55f140d4ad99" containerID="f071254a2c54da8de4c155fdd989f0555480d90cc553521063e705508da94da5" exitCode=0 Dec 01 15:12:30 crc kubenswrapper[4810]: I1201 15:12:30.107679 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zcdww" event={"ID":"d0e41ca0-4806-49a1-bb76-55f140d4ad99","Type":"ContainerDied","Data":"f071254a2c54da8de4c155fdd989f0555480d90cc553521063e705508da94da5"} Dec 01 15:12:31 crc kubenswrapper[4810]: I1201 15:12:31.119642 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zcdww" event={"ID":"d0e41ca0-4806-49a1-bb76-55f140d4ad99","Type":"ContainerStarted","Data":"351aa4d079c8d533d4855417d5a0ecdb3fa2637be834c9fdd03b0e6ab9646ec6"} Dec 01 15:12:31 crc kubenswrapper[4810]: I1201 15:12:31.143664 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zcdww" podStartSLOduration=2.69469435 podStartE2EDuration="9.143645112s" podCreationTimestamp="2025-12-01 15:12:22 +0000 UTC" firstStartedPulling="2025-12-01 15:12:24.065671794 +0000 UTC m=+2309.829181407" lastFinishedPulling="2025-12-01 15:12:30.514622566 +0000 UTC m=+2316.278132169" observedRunningTime="2025-12-01 15:12:31.140431316 +0000 UTC m=+2316.903940929" watchObservedRunningTime="2025-12-01 15:12:31.143645112 +0000 UTC m=+2316.907154715" Dec 01 15:12:32 crc kubenswrapper[4810]: I1201 15:12:32.888569 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zcdww" Dec 01 15:12:32 crc kubenswrapper[4810]: I1201 15:12:32.888644 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zcdww" Dec 01 15:12:32 crc kubenswrapper[4810]: I1201 15:12:32.939111 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zcdww" Dec 01 15:12:42 crc kubenswrapper[4810]: I1201 15:12:42.954198 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zcdww" Dec 01 15:12:43 crc kubenswrapper[4810]: I1201 15:12:43.022221 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zcdww"] Dec 01 15:12:43 crc kubenswrapper[4810]: I1201 15:12:43.079600 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w9g5p"] Dec 01 15:12:43 crc kubenswrapper[4810]: I1201 15:12:43.079931 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-w9g5p" podUID="6f441f5b-94ae-4932-b534-b8afa7716e44" containerName="registry-server" containerID="cri-o://69a4f4484793971aed60c5adede195f46aefdd27c0e3c367b2f8f03d48705816" gracePeriod=2 Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.001878 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w9g5p" Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.164689 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5psc\" (UniqueName: \"kubernetes.io/projected/6f441f5b-94ae-4932-b534-b8afa7716e44-kube-api-access-l5psc\") pod \"6f441f5b-94ae-4932-b534-b8afa7716e44\" (UID: \"6f441f5b-94ae-4932-b534-b8afa7716e44\") " Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.164836 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f441f5b-94ae-4932-b534-b8afa7716e44-catalog-content\") pod \"6f441f5b-94ae-4932-b534-b8afa7716e44\" (UID: \"6f441f5b-94ae-4932-b534-b8afa7716e44\") " Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.164859 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f441f5b-94ae-4932-b534-b8afa7716e44-utilities\") pod \"6f441f5b-94ae-4932-b534-b8afa7716e44\" (UID: \"6f441f5b-94ae-4932-b534-b8afa7716e44\") " Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.166023 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f441f5b-94ae-4932-b534-b8afa7716e44-utilities" (OuterVolumeSpecName: "utilities") pod "6f441f5b-94ae-4932-b534-b8afa7716e44" (UID: "6f441f5b-94ae-4932-b534-b8afa7716e44"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.172871 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f441f5b-94ae-4932-b534-b8afa7716e44-kube-api-access-l5psc" (OuterVolumeSpecName: "kube-api-access-l5psc") pod "6f441f5b-94ae-4932-b534-b8afa7716e44" (UID: "6f441f5b-94ae-4932-b534-b8afa7716e44"). InnerVolumeSpecName "kube-api-access-l5psc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.216013 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f441f5b-94ae-4932-b534-b8afa7716e44-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6f441f5b-94ae-4932-b534-b8afa7716e44" (UID: "6f441f5b-94ae-4932-b534-b8afa7716e44"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.227522 4810 generic.go:334] "Generic (PLEG): container finished" podID="6f441f5b-94ae-4932-b534-b8afa7716e44" containerID="69a4f4484793971aed60c5adede195f46aefdd27c0e3c367b2f8f03d48705816" exitCode=0 Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.227560 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w9g5p" event={"ID":"6f441f5b-94ae-4932-b534-b8afa7716e44","Type":"ContainerDied","Data":"69a4f4484793971aed60c5adede195f46aefdd27c0e3c367b2f8f03d48705816"} Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.227583 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w9g5p" event={"ID":"6f441f5b-94ae-4932-b534-b8afa7716e44","Type":"ContainerDied","Data":"99759ba92d8172a42251ac6e878885664a782c0f7453f030d173d7587710bb6c"} Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.227626 4810 scope.go:117] "RemoveContainer" containerID="69a4f4484793971aed60c5adede195f46aefdd27c0e3c367b2f8f03d48705816" Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.227720 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w9g5p" Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.253080 4810 scope.go:117] "RemoveContainer" containerID="1aeb0490544d249d321a133e660bfcc3b3b4db53baa64a6dcf093b09110af6d2" Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.256342 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w9g5p"] Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.266781 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5psc\" (UniqueName: \"kubernetes.io/projected/6f441f5b-94ae-4932-b534-b8afa7716e44-kube-api-access-l5psc\") on node \"crc\" DevicePath \"\"" Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.266823 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f441f5b-94ae-4932-b534-b8afa7716e44-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.266836 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f441f5b-94ae-4932-b534-b8afa7716e44-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.271309 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-w9g5p"] Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.278107 4810 scope.go:117] "RemoveContainer" containerID="2a7cd1ac7379b2a904a3bd510f246057ee387912db8529cbfc8f4b1e76a34e74" Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.303109 4810 scope.go:117] "RemoveContainer" containerID="69a4f4484793971aed60c5adede195f46aefdd27c0e3c367b2f8f03d48705816" Dec 01 15:12:44 crc kubenswrapper[4810]: E1201 15:12:44.303585 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69a4f4484793971aed60c5adede195f46aefdd27c0e3c367b2f8f03d48705816\": container with ID starting with 69a4f4484793971aed60c5adede195f46aefdd27c0e3c367b2f8f03d48705816 not found: ID does not exist" containerID="69a4f4484793971aed60c5adede195f46aefdd27c0e3c367b2f8f03d48705816" Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.303636 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69a4f4484793971aed60c5adede195f46aefdd27c0e3c367b2f8f03d48705816"} err="failed to get container status \"69a4f4484793971aed60c5adede195f46aefdd27c0e3c367b2f8f03d48705816\": rpc error: code = NotFound desc = could not find container \"69a4f4484793971aed60c5adede195f46aefdd27c0e3c367b2f8f03d48705816\": container with ID starting with 69a4f4484793971aed60c5adede195f46aefdd27c0e3c367b2f8f03d48705816 not found: ID does not exist" Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.303657 4810 scope.go:117] "RemoveContainer" containerID="1aeb0490544d249d321a133e660bfcc3b3b4db53baa64a6dcf093b09110af6d2" Dec 01 15:12:44 crc kubenswrapper[4810]: E1201 15:12:44.304034 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1aeb0490544d249d321a133e660bfcc3b3b4db53baa64a6dcf093b09110af6d2\": container with ID starting with 1aeb0490544d249d321a133e660bfcc3b3b4db53baa64a6dcf093b09110af6d2 not found: ID does not exist" containerID="1aeb0490544d249d321a133e660bfcc3b3b4db53baa64a6dcf093b09110af6d2" Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.304102 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1aeb0490544d249d321a133e660bfcc3b3b4db53baa64a6dcf093b09110af6d2"} err="failed to get container status \"1aeb0490544d249d321a133e660bfcc3b3b4db53baa64a6dcf093b09110af6d2\": rpc error: code = NotFound desc = could not find container \"1aeb0490544d249d321a133e660bfcc3b3b4db53baa64a6dcf093b09110af6d2\": container with ID starting with 1aeb0490544d249d321a133e660bfcc3b3b4db53baa64a6dcf093b09110af6d2 not found: ID does not exist" Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.304121 4810 scope.go:117] "RemoveContainer" containerID="2a7cd1ac7379b2a904a3bd510f246057ee387912db8529cbfc8f4b1e76a34e74" Dec 01 15:12:44 crc kubenswrapper[4810]: E1201 15:12:44.304419 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a7cd1ac7379b2a904a3bd510f246057ee387912db8529cbfc8f4b1e76a34e74\": container with ID starting with 2a7cd1ac7379b2a904a3bd510f246057ee387912db8529cbfc8f4b1e76a34e74 not found: ID does not exist" containerID="2a7cd1ac7379b2a904a3bd510f246057ee387912db8529cbfc8f4b1e76a34e74" Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.304527 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a7cd1ac7379b2a904a3bd510f246057ee387912db8529cbfc8f4b1e76a34e74"} err="failed to get container status \"2a7cd1ac7379b2a904a3bd510f246057ee387912db8529cbfc8f4b1e76a34e74\": rpc error: code = NotFound desc = could not find container \"2a7cd1ac7379b2a904a3bd510f246057ee387912db8529cbfc8f4b1e76a34e74\": container with ID starting with 2a7cd1ac7379b2a904a3bd510f246057ee387912db8529cbfc8f4b1e76a34e74 not found: ID does not exist" Dec 01 15:12:44 crc kubenswrapper[4810]: I1201 15:12:44.500277 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f441f5b-94ae-4932-b534-b8afa7716e44" path="/var/lib/kubelet/pods/6f441f5b-94ae-4932-b534-b8afa7716e44/volumes" Dec 01 15:13:02 crc kubenswrapper[4810]: I1201 15:13:02.972721 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:13:02 crc kubenswrapper[4810]: I1201 15:13:02.973284 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:13:32 crc kubenswrapper[4810]: I1201 15:13:32.972537 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:13:32 crc kubenswrapper[4810]: I1201 15:13:32.973332 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:14:02 crc kubenswrapper[4810]: I1201 15:14:02.972576 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:14:02 crc kubenswrapper[4810]: I1201 15:14:02.973350 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:14:02 crc kubenswrapper[4810]: I1201 15:14:02.973426 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 15:14:02 crc kubenswrapper[4810]: I1201 15:14:02.974315 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 15:14:02 crc kubenswrapper[4810]: I1201 15:14:02.974415 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" gracePeriod=600 Dec 01 15:14:03 crc kubenswrapper[4810]: E1201 15:14:03.114431 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:14:03 crc kubenswrapper[4810]: I1201 15:14:03.828155 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" exitCode=0 Dec 01 15:14:03 crc kubenswrapper[4810]: I1201 15:14:03.828209 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5"} Dec 01 15:14:03 crc kubenswrapper[4810]: I1201 15:14:03.828271 4810 scope.go:117] "RemoveContainer" containerID="cdf338bcebff0a98fe1226fd221bbb927f85e462320f38cb1d0863a7b290222f" Dec 01 15:14:03 crc kubenswrapper[4810]: I1201 15:14:03.828879 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:14:03 crc kubenswrapper[4810]: E1201 15:14:03.829141 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:14:05 crc kubenswrapper[4810]: I1201 15:14:05.884220 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6vngw"] Dec 01 15:14:05 crc kubenswrapper[4810]: E1201 15:14:05.884972 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f441f5b-94ae-4932-b534-b8afa7716e44" containerName="registry-server" Dec 01 15:14:05 crc kubenswrapper[4810]: I1201 15:14:05.885007 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f441f5b-94ae-4932-b534-b8afa7716e44" containerName="registry-server" Dec 01 15:14:05 crc kubenswrapper[4810]: E1201 15:14:05.885021 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f441f5b-94ae-4932-b534-b8afa7716e44" containerName="extract-utilities" Dec 01 15:14:05 crc kubenswrapper[4810]: I1201 15:14:05.885027 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f441f5b-94ae-4932-b534-b8afa7716e44" containerName="extract-utilities" Dec 01 15:14:05 crc kubenswrapper[4810]: E1201 15:14:05.885039 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f441f5b-94ae-4932-b534-b8afa7716e44" containerName="extract-content" Dec 01 15:14:05 crc kubenswrapper[4810]: I1201 15:14:05.885045 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f441f5b-94ae-4932-b534-b8afa7716e44" containerName="extract-content" Dec 01 15:14:05 crc kubenswrapper[4810]: I1201 15:14:05.885244 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f441f5b-94ae-4932-b534-b8afa7716e44" containerName="registry-server" Dec 01 15:14:05 crc kubenswrapper[4810]: I1201 15:14:05.886507 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vngw" Dec 01 15:14:05 crc kubenswrapper[4810]: I1201 15:14:05.895529 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6vngw"] Dec 01 15:14:06 crc kubenswrapper[4810]: I1201 15:14:06.039170 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e-catalog-content\") pod \"community-operators-6vngw\" (UID: \"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e\") " pod="openshift-marketplace/community-operators-6vngw" Dec 01 15:14:06 crc kubenswrapper[4810]: I1201 15:14:06.039503 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx8mc\" (UniqueName: \"kubernetes.io/projected/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e-kube-api-access-qx8mc\") pod \"community-operators-6vngw\" (UID: \"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e\") " pod="openshift-marketplace/community-operators-6vngw" Dec 01 15:14:06 crc kubenswrapper[4810]: I1201 15:14:06.039601 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e-utilities\") pod \"community-operators-6vngw\" (UID: \"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e\") " pod="openshift-marketplace/community-operators-6vngw" Dec 01 15:14:06 crc kubenswrapper[4810]: I1201 15:14:06.140982 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e-utilities\") pod \"community-operators-6vngw\" (UID: \"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e\") " pod="openshift-marketplace/community-operators-6vngw" Dec 01 15:14:06 crc kubenswrapper[4810]: I1201 15:14:06.141060 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e-catalog-content\") pod \"community-operators-6vngw\" (UID: \"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e\") " pod="openshift-marketplace/community-operators-6vngw" Dec 01 15:14:06 crc kubenswrapper[4810]: I1201 15:14:06.141154 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx8mc\" (UniqueName: \"kubernetes.io/projected/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e-kube-api-access-qx8mc\") pod \"community-operators-6vngw\" (UID: \"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e\") " pod="openshift-marketplace/community-operators-6vngw" Dec 01 15:14:06 crc kubenswrapper[4810]: I1201 15:14:06.141502 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e-utilities\") pod \"community-operators-6vngw\" (UID: \"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e\") " pod="openshift-marketplace/community-operators-6vngw" Dec 01 15:14:06 crc kubenswrapper[4810]: I1201 15:14:06.141684 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e-catalog-content\") pod \"community-operators-6vngw\" (UID: \"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e\") " pod="openshift-marketplace/community-operators-6vngw" Dec 01 15:14:06 crc kubenswrapper[4810]: I1201 15:14:06.162272 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx8mc\" (UniqueName: \"kubernetes.io/projected/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e-kube-api-access-qx8mc\") pod \"community-operators-6vngw\" (UID: \"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e\") " pod="openshift-marketplace/community-operators-6vngw" Dec 01 15:14:06 crc kubenswrapper[4810]: I1201 15:14:06.246579 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vngw" Dec 01 15:14:06 crc kubenswrapper[4810]: I1201 15:14:06.666908 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6vngw"] Dec 01 15:14:06 crc kubenswrapper[4810]: I1201 15:14:06.856509 4810 generic.go:334] "Generic (PLEG): container finished" podID="8ab7c438-ec65-4e1e-ad29-379ba4ddd02e" containerID="e2af957a9c9d52dbfe9e1c5a479e146e1367c535beb6ea1e1e341afac591cf55" exitCode=0 Dec 01 15:14:06 crc kubenswrapper[4810]: I1201 15:14:06.856619 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vngw" event={"ID":"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e","Type":"ContainerDied","Data":"e2af957a9c9d52dbfe9e1c5a479e146e1367c535beb6ea1e1e341afac591cf55"} Dec 01 15:14:06 crc kubenswrapper[4810]: I1201 15:14:06.856832 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vngw" event={"ID":"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e","Type":"ContainerStarted","Data":"6040774dc49082eddf82fbddbf6a72f460f4aab17b015b549797719e97d7c61c"} Dec 01 15:14:06 crc kubenswrapper[4810]: I1201 15:14:06.858677 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 15:14:10 crc kubenswrapper[4810]: I1201 15:14:10.887232 4810 generic.go:334] "Generic (PLEG): container finished" podID="8ab7c438-ec65-4e1e-ad29-379ba4ddd02e" containerID="f5a19743a393a588fd41badaefd677c7892a774a00e76e4929d2cc1d7e21aee6" exitCode=0 Dec 01 15:14:10 crc kubenswrapper[4810]: I1201 15:14:10.887301 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vngw" event={"ID":"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e","Type":"ContainerDied","Data":"f5a19743a393a588fd41badaefd677c7892a774a00e76e4929d2cc1d7e21aee6"} Dec 01 15:14:11 crc kubenswrapper[4810]: I1201 15:14:11.896922 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vngw" event={"ID":"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e","Type":"ContainerStarted","Data":"10e20b6875ec34e50fbf418ef27485b5a3d6468329c93d4f9c33d584f2187a91"} Dec 01 15:14:11 crc kubenswrapper[4810]: I1201 15:14:11.915917 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6vngw" podStartSLOduration=2.287392729 podStartE2EDuration="6.91589747s" podCreationTimestamp="2025-12-01 15:14:05 +0000 UTC" firstStartedPulling="2025-12-01 15:14:06.858412746 +0000 UTC m=+2412.621922349" lastFinishedPulling="2025-12-01 15:14:11.486917467 +0000 UTC m=+2417.250427090" observedRunningTime="2025-12-01 15:14:11.913578607 +0000 UTC m=+2417.677088230" watchObservedRunningTime="2025-12-01 15:14:11.91589747 +0000 UTC m=+2417.679407083" Dec 01 15:14:16 crc kubenswrapper[4810]: I1201 15:14:16.247894 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6vngw" Dec 01 15:14:16 crc kubenswrapper[4810]: I1201 15:14:16.248597 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6vngw" Dec 01 15:14:16 crc kubenswrapper[4810]: I1201 15:14:16.295766 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6vngw" Dec 01 15:14:16 crc kubenswrapper[4810]: I1201 15:14:16.978694 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6vngw" Dec 01 15:14:19 crc kubenswrapper[4810]: I1201 15:14:19.491386 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:14:19 crc kubenswrapper[4810]: E1201 15:14:19.492024 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:14:23 crc kubenswrapper[4810]: I1201 15:14:23.238863 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6vngw"] Dec 01 15:14:23 crc kubenswrapper[4810]: I1201 15:14:23.820865 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gpx9d"] Dec 01 15:14:23 crc kubenswrapper[4810]: I1201 15:14:23.821331 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gpx9d" podUID="92abf8d2-574f-4375-bc2a-3436ab810e23" containerName="registry-server" containerID="cri-o://50b016a6c36e3f48c87f1fed5a95478a1c90242abf91ff3ef84deaf10fc17c95" gracePeriod=2 Dec 01 15:14:24 crc kubenswrapper[4810]: I1201 15:14:24.998813 4810 generic.go:334] "Generic (PLEG): container finished" podID="92abf8d2-574f-4375-bc2a-3436ab810e23" containerID="50b016a6c36e3f48c87f1fed5a95478a1c90242abf91ff3ef84deaf10fc17c95" exitCode=0 Dec 01 15:14:24 crc kubenswrapper[4810]: I1201 15:14:24.998880 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpx9d" event={"ID":"92abf8d2-574f-4375-bc2a-3436ab810e23","Type":"ContainerDied","Data":"50b016a6c36e3f48c87f1fed5a95478a1c90242abf91ff3ef84deaf10fc17c95"} Dec 01 15:14:25 crc kubenswrapper[4810]: I1201 15:14:25.643174 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gpx9d" Dec 01 15:14:25 crc kubenswrapper[4810]: I1201 15:14:25.814435 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n44vh\" (UniqueName: \"kubernetes.io/projected/92abf8d2-574f-4375-bc2a-3436ab810e23-kube-api-access-n44vh\") pod \"92abf8d2-574f-4375-bc2a-3436ab810e23\" (UID: \"92abf8d2-574f-4375-bc2a-3436ab810e23\") " Dec 01 15:14:25 crc kubenswrapper[4810]: I1201 15:14:25.814515 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92abf8d2-574f-4375-bc2a-3436ab810e23-catalog-content\") pod \"92abf8d2-574f-4375-bc2a-3436ab810e23\" (UID: \"92abf8d2-574f-4375-bc2a-3436ab810e23\") " Dec 01 15:14:25 crc kubenswrapper[4810]: I1201 15:14:25.814621 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92abf8d2-574f-4375-bc2a-3436ab810e23-utilities\") pod \"92abf8d2-574f-4375-bc2a-3436ab810e23\" (UID: \"92abf8d2-574f-4375-bc2a-3436ab810e23\") " Dec 01 15:14:25 crc kubenswrapper[4810]: I1201 15:14:25.815688 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92abf8d2-574f-4375-bc2a-3436ab810e23-utilities" (OuterVolumeSpecName: "utilities") pod "92abf8d2-574f-4375-bc2a-3436ab810e23" (UID: "92abf8d2-574f-4375-bc2a-3436ab810e23"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:14:25 crc kubenswrapper[4810]: I1201 15:14:25.821727 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92abf8d2-574f-4375-bc2a-3436ab810e23-kube-api-access-n44vh" (OuterVolumeSpecName: "kube-api-access-n44vh") pod "92abf8d2-574f-4375-bc2a-3436ab810e23" (UID: "92abf8d2-574f-4375-bc2a-3436ab810e23"). InnerVolumeSpecName "kube-api-access-n44vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:14:25 crc kubenswrapper[4810]: I1201 15:14:25.862519 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92abf8d2-574f-4375-bc2a-3436ab810e23-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92abf8d2-574f-4375-bc2a-3436ab810e23" (UID: "92abf8d2-574f-4375-bc2a-3436ab810e23"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:14:25 crc kubenswrapper[4810]: I1201 15:14:25.915940 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n44vh\" (UniqueName: \"kubernetes.io/projected/92abf8d2-574f-4375-bc2a-3436ab810e23-kube-api-access-n44vh\") on node \"crc\" DevicePath \"\"" Dec 01 15:14:25 crc kubenswrapper[4810]: I1201 15:14:25.915974 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92abf8d2-574f-4375-bc2a-3436ab810e23-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:14:25 crc kubenswrapper[4810]: I1201 15:14:25.915986 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92abf8d2-574f-4375-bc2a-3436ab810e23-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:14:26 crc kubenswrapper[4810]: I1201 15:14:26.013907 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpx9d" event={"ID":"92abf8d2-574f-4375-bc2a-3436ab810e23","Type":"ContainerDied","Data":"c8a898448bdc0f5b4c3ca97c943b68af6120dde46ce5ee2ee4b940c26bb9e134"} Dec 01 15:14:26 crc kubenswrapper[4810]: I1201 15:14:26.013989 4810 scope.go:117] "RemoveContainer" containerID="50b016a6c36e3f48c87f1fed5a95478a1c90242abf91ff3ef84deaf10fc17c95" Dec 01 15:14:26 crc kubenswrapper[4810]: I1201 15:14:26.014599 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gpx9d" Dec 01 15:14:26 crc kubenswrapper[4810]: I1201 15:14:26.039453 4810 scope.go:117] "RemoveContainer" containerID="8aa21bb82a328fd9c77663983da79f3bc90b15b398bf87e79db90bccd484dd21" Dec 01 15:14:26 crc kubenswrapper[4810]: I1201 15:14:26.048524 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gpx9d"] Dec 01 15:14:26 crc kubenswrapper[4810]: I1201 15:14:26.053990 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gpx9d"] Dec 01 15:14:26 crc kubenswrapper[4810]: I1201 15:14:26.058839 4810 scope.go:117] "RemoveContainer" containerID="76696122146d81d98b2a5332a03623b7536e9640b8b3e90b54d619d4c3e04cda" Dec 01 15:14:26 crc kubenswrapper[4810]: I1201 15:14:26.502920 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92abf8d2-574f-4375-bc2a-3436ab810e23" path="/var/lib/kubelet/pods/92abf8d2-574f-4375-bc2a-3436ab810e23/volumes" Dec 01 15:14:33 crc kubenswrapper[4810]: I1201 15:14:33.491210 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:14:33 crc kubenswrapper[4810]: E1201 15:14:33.492024 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:14:48 crc kubenswrapper[4810]: I1201 15:14:48.492455 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:14:48 crc kubenswrapper[4810]: E1201 15:14:48.493899 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.143372 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9"] Dec 01 15:15:00 crc kubenswrapper[4810]: E1201 15:15:00.144498 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92abf8d2-574f-4375-bc2a-3436ab810e23" containerName="extract-utilities" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.144516 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="92abf8d2-574f-4375-bc2a-3436ab810e23" containerName="extract-utilities" Dec 01 15:15:00 crc kubenswrapper[4810]: E1201 15:15:00.144539 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92abf8d2-574f-4375-bc2a-3436ab810e23" containerName="registry-server" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.144548 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="92abf8d2-574f-4375-bc2a-3436ab810e23" containerName="registry-server" Dec 01 15:15:00 crc kubenswrapper[4810]: E1201 15:15:00.144560 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92abf8d2-574f-4375-bc2a-3436ab810e23" containerName="extract-content" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.144569 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="92abf8d2-574f-4375-bc2a-3436ab810e23" containerName="extract-content" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.144769 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="92abf8d2-574f-4375-bc2a-3436ab810e23" containerName="registry-server" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.145368 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.148750 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.148780 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.156631 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9"] Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.280769 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2c3a243-4fa8-4c09-90ba-fb901628b7da-secret-volume\") pod \"collect-profiles-29410035-sbvh9\" (UID: \"c2c3a243-4fa8-4c09-90ba-fb901628b7da\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.280814 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tc7d\" (UniqueName: \"kubernetes.io/projected/c2c3a243-4fa8-4c09-90ba-fb901628b7da-kube-api-access-2tc7d\") pod \"collect-profiles-29410035-sbvh9\" (UID: \"c2c3a243-4fa8-4c09-90ba-fb901628b7da\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.280867 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2c3a243-4fa8-4c09-90ba-fb901628b7da-config-volume\") pod \"collect-profiles-29410035-sbvh9\" (UID: \"c2c3a243-4fa8-4c09-90ba-fb901628b7da\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.381806 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2c3a243-4fa8-4c09-90ba-fb901628b7da-secret-volume\") pod \"collect-profiles-29410035-sbvh9\" (UID: \"c2c3a243-4fa8-4c09-90ba-fb901628b7da\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.381852 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tc7d\" (UniqueName: \"kubernetes.io/projected/c2c3a243-4fa8-4c09-90ba-fb901628b7da-kube-api-access-2tc7d\") pod \"collect-profiles-29410035-sbvh9\" (UID: \"c2c3a243-4fa8-4c09-90ba-fb901628b7da\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.381923 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2c3a243-4fa8-4c09-90ba-fb901628b7da-config-volume\") pod \"collect-profiles-29410035-sbvh9\" (UID: \"c2c3a243-4fa8-4c09-90ba-fb901628b7da\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.382902 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2c3a243-4fa8-4c09-90ba-fb901628b7da-config-volume\") pod \"collect-profiles-29410035-sbvh9\" (UID: \"c2c3a243-4fa8-4c09-90ba-fb901628b7da\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.388208 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2c3a243-4fa8-4c09-90ba-fb901628b7da-secret-volume\") pod \"collect-profiles-29410035-sbvh9\" (UID: \"c2c3a243-4fa8-4c09-90ba-fb901628b7da\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.404319 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tc7d\" (UniqueName: \"kubernetes.io/projected/c2c3a243-4fa8-4c09-90ba-fb901628b7da-kube-api-access-2tc7d\") pod \"collect-profiles-29410035-sbvh9\" (UID: \"c2c3a243-4fa8-4c09-90ba-fb901628b7da\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.467227 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9" Dec 01 15:15:00 crc kubenswrapper[4810]: I1201 15:15:00.909902 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9"] Dec 01 15:15:01 crc kubenswrapper[4810]: I1201 15:15:01.270368 4810 generic.go:334] "Generic (PLEG): container finished" podID="c2c3a243-4fa8-4c09-90ba-fb901628b7da" containerID="59c98a47ec8e218fb94ced748c247d8b6232e1e05323abf0ee48987c383faaf3" exitCode=0 Dec 01 15:15:01 crc kubenswrapper[4810]: I1201 15:15:01.270413 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9" event={"ID":"c2c3a243-4fa8-4c09-90ba-fb901628b7da","Type":"ContainerDied","Data":"59c98a47ec8e218fb94ced748c247d8b6232e1e05323abf0ee48987c383faaf3"} Dec 01 15:15:01 crc kubenswrapper[4810]: I1201 15:15:01.270731 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9" event={"ID":"c2c3a243-4fa8-4c09-90ba-fb901628b7da","Type":"ContainerStarted","Data":"683ba61155fc113b935542e97e4075637a87e4a8a4ede73e850ec1ae2c2a5c72"} Dec 01 15:15:02 crc kubenswrapper[4810]: I1201 15:15:02.522921 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9" Dec 01 15:15:02 crc kubenswrapper[4810]: I1201 15:15:02.610362 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tc7d\" (UniqueName: \"kubernetes.io/projected/c2c3a243-4fa8-4c09-90ba-fb901628b7da-kube-api-access-2tc7d\") pod \"c2c3a243-4fa8-4c09-90ba-fb901628b7da\" (UID: \"c2c3a243-4fa8-4c09-90ba-fb901628b7da\") " Dec 01 15:15:02 crc kubenswrapper[4810]: I1201 15:15:02.610483 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2c3a243-4fa8-4c09-90ba-fb901628b7da-config-volume\") pod \"c2c3a243-4fa8-4c09-90ba-fb901628b7da\" (UID: \"c2c3a243-4fa8-4c09-90ba-fb901628b7da\") " Dec 01 15:15:02 crc kubenswrapper[4810]: I1201 15:15:02.610537 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2c3a243-4fa8-4c09-90ba-fb901628b7da-secret-volume\") pod \"c2c3a243-4fa8-4c09-90ba-fb901628b7da\" (UID: \"c2c3a243-4fa8-4c09-90ba-fb901628b7da\") " Dec 01 15:15:02 crc kubenswrapper[4810]: I1201 15:15:02.611242 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2c3a243-4fa8-4c09-90ba-fb901628b7da-config-volume" (OuterVolumeSpecName: "config-volume") pod "c2c3a243-4fa8-4c09-90ba-fb901628b7da" (UID: "c2c3a243-4fa8-4c09-90ba-fb901628b7da"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:15:02 crc kubenswrapper[4810]: I1201 15:15:02.615993 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2c3a243-4fa8-4c09-90ba-fb901628b7da-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c2c3a243-4fa8-4c09-90ba-fb901628b7da" (UID: "c2c3a243-4fa8-4c09-90ba-fb901628b7da"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:15:02 crc kubenswrapper[4810]: I1201 15:15:02.616334 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2c3a243-4fa8-4c09-90ba-fb901628b7da-kube-api-access-2tc7d" (OuterVolumeSpecName: "kube-api-access-2tc7d") pod "c2c3a243-4fa8-4c09-90ba-fb901628b7da" (UID: "c2c3a243-4fa8-4c09-90ba-fb901628b7da"). InnerVolumeSpecName "kube-api-access-2tc7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:15:02 crc kubenswrapper[4810]: I1201 15:15:02.712151 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tc7d\" (UniqueName: \"kubernetes.io/projected/c2c3a243-4fa8-4c09-90ba-fb901628b7da-kube-api-access-2tc7d\") on node \"crc\" DevicePath \"\"" Dec 01 15:15:02 crc kubenswrapper[4810]: I1201 15:15:02.712183 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2c3a243-4fa8-4c09-90ba-fb901628b7da-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 15:15:02 crc kubenswrapper[4810]: I1201 15:15:02.712193 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2c3a243-4fa8-4c09-90ba-fb901628b7da-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 15:15:03 crc kubenswrapper[4810]: I1201 15:15:03.290951 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9" event={"ID":"c2c3a243-4fa8-4c09-90ba-fb901628b7da","Type":"ContainerDied","Data":"683ba61155fc113b935542e97e4075637a87e4a8a4ede73e850ec1ae2c2a5c72"} Dec 01 15:15:03 crc kubenswrapper[4810]: I1201 15:15:03.290986 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="683ba61155fc113b935542e97e4075637a87e4a8a4ede73e850ec1ae2c2a5c72" Dec 01 15:15:03 crc kubenswrapper[4810]: I1201 15:15:03.291050 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9" Dec 01 15:15:03 crc kubenswrapper[4810]: I1201 15:15:03.490978 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:15:03 crc kubenswrapper[4810]: E1201 15:15:03.491196 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:15:03 crc kubenswrapper[4810]: I1201 15:15:03.581851 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld"] Dec 01 15:15:03 crc kubenswrapper[4810]: I1201 15:15:03.587999 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29409990-xjzld"] Dec 01 15:15:04 crc kubenswrapper[4810]: I1201 15:15:04.499957 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="484c6c08-536f-4374-971e-1fb4ef397b64" path="/var/lib/kubelet/pods/484c6c08-536f-4374-971e-1fb4ef397b64/volumes" Dec 01 15:15:07 crc kubenswrapper[4810]: I1201 15:15:07.754362 4810 scope.go:117] "RemoveContainer" containerID="b3feacdd440d69f9f0233048aea0de0190a3da88ac5e9deaeb3af2daf12d95dc" Dec 01 15:15:18 crc kubenswrapper[4810]: I1201 15:15:18.491679 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:15:18 crc kubenswrapper[4810]: E1201 15:15:18.492978 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:15:31 crc kubenswrapper[4810]: I1201 15:15:31.492234 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:15:31 crc kubenswrapper[4810]: E1201 15:15:31.493950 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:15:44 crc kubenswrapper[4810]: I1201 15:15:44.528582 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:15:44 crc kubenswrapper[4810]: E1201 15:15:44.529777 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:15:58 crc kubenswrapper[4810]: I1201 15:15:58.491683 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:15:58 crc kubenswrapper[4810]: E1201 15:15:58.492369 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:16:12 crc kubenswrapper[4810]: I1201 15:16:12.490929 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:16:12 crc kubenswrapper[4810]: E1201 15:16:12.491928 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:16:23 crc kubenswrapper[4810]: I1201 15:16:23.490973 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:16:23 crc kubenswrapper[4810]: E1201 15:16:23.491576 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:16:36 crc kubenswrapper[4810]: I1201 15:16:36.491286 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:16:36 crc kubenswrapper[4810]: E1201 15:16:36.492157 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:16:47 crc kubenswrapper[4810]: I1201 15:16:47.491173 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:16:47 crc kubenswrapper[4810]: E1201 15:16:47.492815 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:16:59 crc kubenswrapper[4810]: I1201 15:16:59.099794 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:16:59 crc kubenswrapper[4810]: E1201 15:16:59.100373 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:17:13 crc kubenswrapper[4810]: I1201 15:17:13.491533 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:17:13 crc kubenswrapper[4810]: E1201 15:17:13.492375 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:17:26 crc kubenswrapper[4810]: I1201 15:17:26.490746 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:17:26 crc kubenswrapper[4810]: E1201 15:17:26.491586 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:17:37 crc kubenswrapper[4810]: I1201 15:17:37.491018 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:17:37 crc kubenswrapper[4810]: E1201 15:17:37.491786 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:17:51 crc kubenswrapper[4810]: I1201 15:17:51.490596 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:17:51 crc kubenswrapper[4810]: E1201 15:17:51.491360 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:18:02 crc kubenswrapper[4810]: I1201 15:18:02.491899 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:18:02 crc kubenswrapper[4810]: E1201 15:18:02.492500 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:18:16 crc kubenswrapper[4810]: I1201 15:18:16.490828 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:18:16 crc kubenswrapper[4810]: E1201 15:18:16.491721 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:18:29 crc kubenswrapper[4810]: I1201 15:18:29.490889 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:18:29 crc kubenswrapper[4810]: E1201 15:18:29.491762 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:18:40 crc kubenswrapper[4810]: I1201 15:18:40.491170 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:18:40 crc kubenswrapper[4810]: E1201 15:18:40.491995 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:18:54 crc kubenswrapper[4810]: I1201 15:18:54.500942 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:18:54 crc kubenswrapper[4810]: E1201 15:18:54.501955 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:19:06 crc kubenswrapper[4810]: I1201 15:19:06.490666 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:19:07 crc kubenswrapper[4810]: I1201 15:19:07.532732 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"1b6adf628cd09401f04ff4153fb1f5fd44fef07fe578fc9a26d88fa6d79cedb2"} Dec 01 15:21:32 crc kubenswrapper[4810]: I1201 15:21:32.975685 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:21:32 crc kubenswrapper[4810]: I1201 15:21:32.976281 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:22:02 crc kubenswrapper[4810]: I1201 15:22:02.975503 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:22:02 crc kubenswrapper[4810]: I1201 15:22:02.976084 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:22:32 crc kubenswrapper[4810]: I1201 15:22:32.973165 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:22:32 crc kubenswrapper[4810]: I1201 15:22:32.973924 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:22:32 crc kubenswrapper[4810]: I1201 15:22:32.973990 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 15:22:32 crc kubenswrapper[4810]: I1201 15:22:32.974858 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1b6adf628cd09401f04ff4153fb1f5fd44fef07fe578fc9a26d88fa6d79cedb2"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 15:22:32 crc kubenswrapper[4810]: I1201 15:22:32.975024 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://1b6adf628cd09401f04ff4153fb1f5fd44fef07fe578fc9a26d88fa6d79cedb2" gracePeriod=600 Dec 01 15:22:34 crc kubenswrapper[4810]: I1201 15:22:34.001372 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="1b6adf628cd09401f04ff4153fb1f5fd44fef07fe578fc9a26d88fa6d79cedb2" exitCode=0 Dec 01 15:22:34 crc kubenswrapper[4810]: I1201 15:22:34.001435 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"1b6adf628cd09401f04ff4153fb1f5fd44fef07fe578fc9a26d88fa6d79cedb2"} Dec 01 15:22:34 crc kubenswrapper[4810]: I1201 15:22:34.001885 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9"} Dec 01 15:22:34 crc kubenswrapper[4810]: I1201 15:22:34.001925 4810 scope.go:117] "RemoveContainer" containerID="c2b775d3efec0b1415122fad53d167d5c9d6d2c0925ac1ecb3b29d82e3b498b5" Dec 01 15:24:13 crc kubenswrapper[4810]: I1201 15:24:13.170056 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qpjw7"] Dec 01 15:24:13 crc kubenswrapper[4810]: E1201 15:24:13.171038 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2c3a243-4fa8-4c09-90ba-fb901628b7da" containerName="collect-profiles" Dec 01 15:24:13 crc kubenswrapper[4810]: I1201 15:24:13.171056 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2c3a243-4fa8-4c09-90ba-fb901628b7da" containerName="collect-profiles" Dec 01 15:24:13 crc kubenswrapper[4810]: I1201 15:24:13.171233 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2c3a243-4fa8-4c09-90ba-fb901628b7da" containerName="collect-profiles" Dec 01 15:24:13 crc kubenswrapper[4810]: I1201 15:24:13.173754 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qpjw7" Dec 01 15:24:13 crc kubenswrapper[4810]: I1201 15:24:13.182989 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qpjw7"] Dec 01 15:24:13 crc kubenswrapper[4810]: I1201 15:24:13.345422 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e-catalog-content\") pod \"redhat-operators-qpjw7\" (UID: \"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e\") " pod="openshift-marketplace/redhat-operators-qpjw7" Dec 01 15:24:13 crc kubenswrapper[4810]: I1201 15:24:13.345556 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e-utilities\") pod \"redhat-operators-qpjw7\" (UID: \"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e\") " pod="openshift-marketplace/redhat-operators-qpjw7" Dec 01 15:24:13 crc kubenswrapper[4810]: I1201 15:24:13.345633 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7tqk\" (UniqueName: \"kubernetes.io/projected/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e-kube-api-access-n7tqk\") pod \"redhat-operators-qpjw7\" (UID: \"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e\") " pod="openshift-marketplace/redhat-operators-qpjw7" Dec 01 15:24:13 crc kubenswrapper[4810]: I1201 15:24:13.446819 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7tqk\" (UniqueName: \"kubernetes.io/projected/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e-kube-api-access-n7tqk\") pod \"redhat-operators-qpjw7\" (UID: \"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e\") " pod="openshift-marketplace/redhat-operators-qpjw7" Dec 01 15:24:13 crc kubenswrapper[4810]: I1201 15:24:13.446893 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e-catalog-content\") pod \"redhat-operators-qpjw7\" (UID: \"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e\") " pod="openshift-marketplace/redhat-operators-qpjw7" Dec 01 15:24:13 crc kubenswrapper[4810]: I1201 15:24:13.446962 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e-utilities\") pod \"redhat-operators-qpjw7\" (UID: \"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e\") " pod="openshift-marketplace/redhat-operators-qpjw7" Dec 01 15:24:13 crc kubenswrapper[4810]: I1201 15:24:13.447563 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e-catalog-content\") pod \"redhat-operators-qpjw7\" (UID: \"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e\") " pod="openshift-marketplace/redhat-operators-qpjw7" Dec 01 15:24:13 crc kubenswrapper[4810]: I1201 15:24:13.447610 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e-utilities\") pod \"redhat-operators-qpjw7\" (UID: \"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e\") " pod="openshift-marketplace/redhat-operators-qpjw7" Dec 01 15:24:13 crc kubenswrapper[4810]: I1201 15:24:13.471904 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7tqk\" (UniqueName: \"kubernetes.io/projected/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e-kube-api-access-n7tqk\") pod \"redhat-operators-qpjw7\" (UID: \"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e\") " pod="openshift-marketplace/redhat-operators-qpjw7" Dec 01 15:24:13 crc kubenswrapper[4810]: I1201 15:24:13.503808 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qpjw7" Dec 01 15:24:13 crc kubenswrapper[4810]: I1201 15:24:13.740304 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qpjw7"] Dec 01 15:24:13 crc kubenswrapper[4810]: I1201 15:24:13.767855 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpjw7" event={"ID":"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e","Type":"ContainerStarted","Data":"6f72dc1e1a0b0de280294aadc657f9795ada8fb11c496fa8ce3d406618adc2bb"} Dec 01 15:24:14 crc kubenswrapper[4810]: I1201 15:24:14.777211 4810 generic.go:334] "Generic (PLEG): container finished" podID="db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e" containerID="4e35b824467864ddf631f32f9fc63f67c0fd45e0470d9351eb0192d4d421813f" exitCode=0 Dec 01 15:24:14 crc kubenswrapper[4810]: I1201 15:24:14.777264 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpjw7" event={"ID":"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e","Type":"ContainerDied","Data":"4e35b824467864ddf631f32f9fc63f67c0fd45e0470d9351eb0192d4d421813f"} Dec 01 15:24:14 crc kubenswrapper[4810]: I1201 15:24:14.780387 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 15:24:16 crc kubenswrapper[4810]: I1201 15:24:16.794452 4810 generic.go:334] "Generic (PLEG): container finished" podID="db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e" containerID="5caa2e71d7d0b5e0622686d2d8524ed2be52fe02f649a854949536f3410166cf" exitCode=0 Dec 01 15:24:16 crc kubenswrapper[4810]: I1201 15:24:16.794516 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpjw7" event={"ID":"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e","Type":"ContainerDied","Data":"5caa2e71d7d0b5e0622686d2d8524ed2be52fe02f649a854949536f3410166cf"} Dec 01 15:24:17 crc kubenswrapper[4810]: I1201 15:24:17.809719 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpjw7" event={"ID":"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e","Type":"ContainerStarted","Data":"1e5b472fb143bfd390bc5658b41f9f9eaa75302ad6383e452b6ab48136daa26a"} Dec 01 15:24:17 crc kubenswrapper[4810]: I1201 15:24:17.835249 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qpjw7" podStartSLOduration=2.351454234 podStartE2EDuration="4.835225823s" podCreationTimestamp="2025-12-01 15:24:13 +0000 UTC" firstStartedPulling="2025-12-01 15:24:14.780114049 +0000 UTC m=+3020.543623652" lastFinishedPulling="2025-12-01 15:24:17.263885638 +0000 UTC m=+3023.027395241" observedRunningTime="2025-12-01 15:24:17.829924599 +0000 UTC m=+3023.593434212" watchObservedRunningTime="2025-12-01 15:24:17.835225823 +0000 UTC m=+3023.598735426" Dec 01 15:24:23 crc kubenswrapper[4810]: I1201 15:24:23.505087 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qpjw7" Dec 01 15:24:23 crc kubenswrapper[4810]: I1201 15:24:23.505862 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qpjw7" Dec 01 15:24:23 crc kubenswrapper[4810]: I1201 15:24:23.566759 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qpjw7" Dec 01 15:24:23 crc kubenswrapper[4810]: I1201 15:24:23.899553 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qpjw7" Dec 01 15:24:23 crc kubenswrapper[4810]: I1201 15:24:23.959781 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qpjw7"] Dec 01 15:24:25 crc kubenswrapper[4810]: I1201 15:24:25.865425 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qpjw7" podUID="db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e" containerName="registry-server" containerID="cri-o://1e5b472fb143bfd390bc5658b41f9f9eaa75302ad6383e452b6ab48136daa26a" gracePeriod=2 Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.854922 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qpjw7" Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.880809 4810 generic.go:334] "Generic (PLEG): container finished" podID="db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e" containerID="1e5b472fb143bfd390bc5658b41f9f9eaa75302ad6383e452b6ab48136daa26a" exitCode=0 Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.880857 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpjw7" event={"ID":"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e","Type":"ContainerDied","Data":"1e5b472fb143bfd390bc5658b41f9f9eaa75302ad6383e452b6ab48136daa26a"} Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.880884 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qpjw7" Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.880911 4810 scope.go:117] "RemoveContainer" containerID="1e5b472fb143bfd390bc5658b41f9f9eaa75302ad6383e452b6ab48136daa26a" Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.880895 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qpjw7" event={"ID":"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e","Type":"ContainerDied","Data":"6f72dc1e1a0b0de280294aadc657f9795ada8fb11c496fa8ce3d406618adc2bb"} Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.900407 4810 scope.go:117] "RemoveContainer" containerID="5caa2e71d7d0b5e0622686d2d8524ed2be52fe02f649a854949536f3410166cf" Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.918703 4810 scope.go:117] "RemoveContainer" containerID="4e35b824467864ddf631f32f9fc63f67c0fd45e0470d9351eb0192d4d421813f" Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.945543 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e-catalog-content\") pod \"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e\" (UID: \"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e\") " Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.945610 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e-utilities\") pod \"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e\" (UID: \"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e\") " Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.945699 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7tqk\" (UniqueName: \"kubernetes.io/projected/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e-kube-api-access-n7tqk\") pod \"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e\" (UID: \"db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e\") " Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.947196 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e-utilities" (OuterVolumeSpecName: "utilities") pod "db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e" (UID: "db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.952441 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e-kube-api-access-n7tqk" (OuterVolumeSpecName: "kube-api-access-n7tqk") pod "db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e" (UID: "db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e"). InnerVolumeSpecName "kube-api-access-n7tqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.957737 4810 scope.go:117] "RemoveContainer" containerID="1e5b472fb143bfd390bc5658b41f9f9eaa75302ad6383e452b6ab48136daa26a" Dec 01 15:24:26 crc kubenswrapper[4810]: E1201 15:24:26.958502 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e5b472fb143bfd390bc5658b41f9f9eaa75302ad6383e452b6ab48136daa26a\": container with ID starting with 1e5b472fb143bfd390bc5658b41f9f9eaa75302ad6383e452b6ab48136daa26a not found: ID does not exist" containerID="1e5b472fb143bfd390bc5658b41f9f9eaa75302ad6383e452b6ab48136daa26a" Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.958543 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e5b472fb143bfd390bc5658b41f9f9eaa75302ad6383e452b6ab48136daa26a"} err="failed to get container status \"1e5b472fb143bfd390bc5658b41f9f9eaa75302ad6383e452b6ab48136daa26a\": rpc error: code = NotFound desc = could not find container \"1e5b472fb143bfd390bc5658b41f9f9eaa75302ad6383e452b6ab48136daa26a\": container with ID starting with 1e5b472fb143bfd390bc5658b41f9f9eaa75302ad6383e452b6ab48136daa26a not found: ID does not exist" Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.958637 4810 scope.go:117] "RemoveContainer" containerID="5caa2e71d7d0b5e0622686d2d8524ed2be52fe02f649a854949536f3410166cf" Dec 01 15:24:26 crc kubenswrapper[4810]: E1201 15:24:26.959316 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5caa2e71d7d0b5e0622686d2d8524ed2be52fe02f649a854949536f3410166cf\": container with ID starting with 5caa2e71d7d0b5e0622686d2d8524ed2be52fe02f649a854949536f3410166cf not found: ID does not exist" containerID="5caa2e71d7d0b5e0622686d2d8524ed2be52fe02f649a854949536f3410166cf" Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.959390 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5caa2e71d7d0b5e0622686d2d8524ed2be52fe02f649a854949536f3410166cf"} err="failed to get container status \"5caa2e71d7d0b5e0622686d2d8524ed2be52fe02f649a854949536f3410166cf\": rpc error: code = NotFound desc = could not find container \"5caa2e71d7d0b5e0622686d2d8524ed2be52fe02f649a854949536f3410166cf\": container with ID starting with 5caa2e71d7d0b5e0622686d2d8524ed2be52fe02f649a854949536f3410166cf not found: ID does not exist" Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.959442 4810 scope.go:117] "RemoveContainer" containerID="4e35b824467864ddf631f32f9fc63f67c0fd45e0470d9351eb0192d4d421813f" Dec 01 15:24:26 crc kubenswrapper[4810]: E1201 15:24:26.960129 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e35b824467864ddf631f32f9fc63f67c0fd45e0470d9351eb0192d4d421813f\": container with ID starting with 4e35b824467864ddf631f32f9fc63f67c0fd45e0470d9351eb0192d4d421813f not found: ID does not exist" containerID="4e35b824467864ddf631f32f9fc63f67c0fd45e0470d9351eb0192d4d421813f" Dec 01 15:24:26 crc kubenswrapper[4810]: I1201 15:24:26.960179 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e35b824467864ddf631f32f9fc63f67c0fd45e0470d9351eb0192d4d421813f"} err="failed to get container status \"4e35b824467864ddf631f32f9fc63f67c0fd45e0470d9351eb0192d4d421813f\": rpc error: code = NotFound desc = could not find container \"4e35b824467864ddf631f32f9fc63f67c0fd45e0470d9351eb0192d4d421813f\": container with ID starting with 4e35b824467864ddf631f32f9fc63f67c0fd45e0470d9351eb0192d4d421813f not found: ID does not exist" Dec 01 15:24:27 crc kubenswrapper[4810]: I1201 15:24:27.046778 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:24:27 crc kubenswrapper[4810]: I1201 15:24:27.046819 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7tqk\" (UniqueName: \"kubernetes.io/projected/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e-kube-api-access-n7tqk\") on node \"crc\" DevicePath \"\"" Dec 01 15:24:27 crc kubenswrapper[4810]: I1201 15:24:27.649837 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e" (UID: "db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:24:27 crc kubenswrapper[4810]: I1201 15:24:27.660763 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:24:27 crc kubenswrapper[4810]: I1201 15:24:27.825505 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qpjw7"] Dec 01 15:24:27 crc kubenswrapper[4810]: I1201 15:24:27.830901 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qpjw7"] Dec 01 15:24:28 crc kubenswrapper[4810]: I1201 15:24:28.500596 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e" path="/var/lib/kubelet/pods/db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e/volumes" Dec 01 15:24:36 crc kubenswrapper[4810]: I1201 15:24:36.321848 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fsds7"] Dec 01 15:24:36 crc kubenswrapper[4810]: E1201 15:24:36.323349 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e" containerName="registry-server" Dec 01 15:24:36 crc kubenswrapper[4810]: I1201 15:24:36.323366 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e" containerName="registry-server" Dec 01 15:24:36 crc kubenswrapper[4810]: E1201 15:24:36.323383 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e" containerName="extract-utilities" Dec 01 15:24:36 crc kubenswrapper[4810]: I1201 15:24:36.323392 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e" containerName="extract-utilities" Dec 01 15:24:36 crc kubenswrapper[4810]: E1201 15:24:36.323408 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e" containerName="extract-content" Dec 01 15:24:36 crc kubenswrapper[4810]: I1201 15:24:36.323415 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e" containerName="extract-content" Dec 01 15:24:36 crc kubenswrapper[4810]: I1201 15:24:36.323632 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="db3f9f1e-227f-42a5-8a4c-aa2783a0cd7e" containerName="registry-server" Dec 01 15:24:36 crc kubenswrapper[4810]: I1201 15:24:36.326071 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fsds7" Dec 01 15:24:36 crc kubenswrapper[4810]: I1201 15:24:36.329864 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fsds7"] Dec 01 15:24:36 crc kubenswrapper[4810]: I1201 15:24:36.497141 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvhlv\" (UniqueName: \"kubernetes.io/projected/2969563a-1f7d-44b4-8a27-b674b18a7090-kube-api-access-mvhlv\") pod \"community-operators-fsds7\" (UID: \"2969563a-1f7d-44b4-8a27-b674b18a7090\") " pod="openshift-marketplace/community-operators-fsds7" Dec 01 15:24:36 crc kubenswrapper[4810]: I1201 15:24:36.497279 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2969563a-1f7d-44b4-8a27-b674b18a7090-catalog-content\") pod \"community-operators-fsds7\" (UID: \"2969563a-1f7d-44b4-8a27-b674b18a7090\") " pod="openshift-marketplace/community-operators-fsds7" Dec 01 15:24:36 crc kubenswrapper[4810]: I1201 15:24:36.497413 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2969563a-1f7d-44b4-8a27-b674b18a7090-utilities\") pod \"community-operators-fsds7\" (UID: \"2969563a-1f7d-44b4-8a27-b674b18a7090\") " pod="openshift-marketplace/community-operators-fsds7" Dec 01 15:24:36 crc kubenswrapper[4810]: I1201 15:24:36.598619 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvhlv\" (UniqueName: \"kubernetes.io/projected/2969563a-1f7d-44b4-8a27-b674b18a7090-kube-api-access-mvhlv\") pod \"community-operators-fsds7\" (UID: \"2969563a-1f7d-44b4-8a27-b674b18a7090\") " pod="openshift-marketplace/community-operators-fsds7" Dec 01 15:24:36 crc kubenswrapper[4810]: I1201 15:24:36.599068 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2969563a-1f7d-44b4-8a27-b674b18a7090-catalog-content\") pod \"community-operators-fsds7\" (UID: \"2969563a-1f7d-44b4-8a27-b674b18a7090\") " pod="openshift-marketplace/community-operators-fsds7" Dec 01 15:24:36 crc kubenswrapper[4810]: I1201 15:24:36.599512 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2969563a-1f7d-44b4-8a27-b674b18a7090-catalog-content\") pod \"community-operators-fsds7\" (UID: \"2969563a-1f7d-44b4-8a27-b674b18a7090\") " pod="openshift-marketplace/community-operators-fsds7" Dec 01 15:24:36 crc kubenswrapper[4810]: I1201 15:24:36.599663 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2969563a-1f7d-44b4-8a27-b674b18a7090-utilities\") pod \"community-operators-fsds7\" (UID: \"2969563a-1f7d-44b4-8a27-b674b18a7090\") " pod="openshift-marketplace/community-operators-fsds7" Dec 01 15:24:36 crc kubenswrapper[4810]: I1201 15:24:36.599937 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2969563a-1f7d-44b4-8a27-b674b18a7090-utilities\") pod \"community-operators-fsds7\" (UID: \"2969563a-1f7d-44b4-8a27-b674b18a7090\") " pod="openshift-marketplace/community-operators-fsds7" Dec 01 15:24:36 crc kubenswrapper[4810]: I1201 15:24:36.623113 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvhlv\" (UniqueName: \"kubernetes.io/projected/2969563a-1f7d-44b4-8a27-b674b18a7090-kube-api-access-mvhlv\") pod \"community-operators-fsds7\" (UID: \"2969563a-1f7d-44b4-8a27-b674b18a7090\") " pod="openshift-marketplace/community-operators-fsds7" Dec 01 15:24:36 crc kubenswrapper[4810]: I1201 15:24:36.647462 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fsds7" Dec 01 15:24:37 crc kubenswrapper[4810]: I1201 15:24:37.091088 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fsds7"] Dec 01 15:24:37 crc kubenswrapper[4810]: I1201 15:24:37.966058 4810 generic.go:334] "Generic (PLEG): container finished" podID="2969563a-1f7d-44b4-8a27-b674b18a7090" containerID="9181f32f0e8cd1779ccd66f7361479b7db809d644c73f8c354cb81ae35833fdb" exitCode=0 Dec 01 15:24:37 crc kubenswrapper[4810]: I1201 15:24:37.966164 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fsds7" event={"ID":"2969563a-1f7d-44b4-8a27-b674b18a7090","Type":"ContainerDied","Data":"9181f32f0e8cd1779ccd66f7361479b7db809d644c73f8c354cb81ae35833fdb"} Dec 01 15:24:37 crc kubenswrapper[4810]: I1201 15:24:37.966384 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fsds7" event={"ID":"2969563a-1f7d-44b4-8a27-b674b18a7090","Type":"ContainerStarted","Data":"e810d3d3cc4caa5d294cfbdd76a72eaa75c166e9cab85be5baec663278dc6e70"} Dec 01 15:24:38 crc kubenswrapper[4810]: I1201 15:24:38.979077 4810 generic.go:334] "Generic (PLEG): container finished" podID="2969563a-1f7d-44b4-8a27-b674b18a7090" containerID="3095a9fdb0e88560e6794b8233c468f67d2cc2053d821eeffd46622420346f4b" exitCode=0 Dec 01 15:24:38 crc kubenswrapper[4810]: I1201 15:24:38.979309 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fsds7" event={"ID":"2969563a-1f7d-44b4-8a27-b674b18a7090","Type":"ContainerDied","Data":"3095a9fdb0e88560e6794b8233c468f67d2cc2053d821eeffd46622420346f4b"} Dec 01 15:24:39 crc kubenswrapper[4810]: I1201 15:24:39.987576 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fsds7" event={"ID":"2969563a-1f7d-44b4-8a27-b674b18a7090","Type":"ContainerStarted","Data":"59f3876b16132bf82bb6ead05bb924fbadb25ed07384fad8b43600645e75ab4a"} Dec 01 15:24:40 crc kubenswrapper[4810]: I1201 15:24:40.009427 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fsds7" podStartSLOduration=2.225272901 podStartE2EDuration="4.009403527s" podCreationTimestamp="2025-12-01 15:24:36 +0000 UTC" firstStartedPulling="2025-12-01 15:24:37.968195249 +0000 UTC m=+3043.731704852" lastFinishedPulling="2025-12-01 15:24:39.752325875 +0000 UTC m=+3045.515835478" observedRunningTime="2025-12-01 15:24:40.003156008 +0000 UTC m=+3045.766665621" watchObservedRunningTime="2025-12-01 15:24:40.009403527 +0000 UTC m=+3045.772913130" Dec 01 15:24:46 crc kubenswrapper[4810]: I1201 15:24:46.648391 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fsds7" Dec 01 15:24:46 crc kubenswrapper[4810]: I1201 15:24:46.648989 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fsds7" Dec 01 15:24:46 crc kubenswrapper[4810]: I1201 15:24:46.691750 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fsds7" Dec 01 15:24:47 crc kubenswrapper[4810]: I1201 15:24:47.069129 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fsds7" Dec 01 15:24:47 crc kubenswrapper[4810]: I1201 15:24:47.110992 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fsds7"] Dec 01 15:24:49 crc kubenswrapper[4810]: I1201 15:24:49.046093 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fsds7" podUID="2969563a-1f7d-44b4-8a27-b674b18a7090" containerName="registry-server" containerID="cri-o://59f3876b16132bf82bb6ead05bb924fbadb25ed07384fad8b43600645e75ab4a" gracePeriod=2 Dec 01 15:24:49 crc kubenswrapper[4810]: I1201 15:24:49.912728 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fsds7" Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.010274 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2969563a-1f7d-44b4-8a27-b674b18a7090-catalog-content\") pod \"2969563a-1f7d-44b4-8a27-b674b18a7090\" (UID: \"2969563a-1f7d-44b4-8a27-b674b18a7090\") " Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.010433 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2969563a-1f7d-44b4-8a27-b674b18a7090-utilities\") pod \"2969563a-1f7d-44b4-8a27-b674b18a7090\" (UID: \"2969563a-1f7d-44b4-8a27-b674b18a7090\") " Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.010631 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvhlv\" (UniqueName: \"kubernetes.io/projected/2969563a-1f7d-44b4-8a27-b674b18a7090-kube-api-access-mvhlv\") pod \"2969563a-1f7d-44b4-8a27-b674b18a7090\" (UID: \"2969563a-1f7d-44b4-8a27-b674b18a7090\") " Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.011776 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2969563a-1f7d-44b4-8a27-b674b18a7090-utilities" (OuterVolumeSpecName: "utilities") pod "2969563a-1f7d-44b4-8a27-b674b18a7090" (UID: "2969563a-1f7d-44b4-8a27-b674b18a7090"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.015168 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2969563a-1f7d-44b4-8a27-b674b18a7090-kube-api-access-mvhlv" (OuterVolumeSpecName: "kube-api-access-mvhlv") pod "2969563a-1f7d-44b4-8a27-b674b18a7090" (UID: "2969563a-1f7d-44b4-8a27-b674b18a7090"). InnerVolumeSpecName "kube-api-access-mvhlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.060593 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2969563a-1f7d-44b4-8a27-b674b18a7090-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2969563a-1f7d-44b4-8a27-b674b18a7090" (UID: "2969563a-1f7d-44b4-8a27-b674b18a7090"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.062374 4810 generic.go:334] "Generic (PLEG): container finished" podID="2969563a-1f7d-44b4-8a27-b674b18a7090" containerID="59f3876b16132bf82bb6ead05bb924fbadb25ed07384fad8b43600645e75ab4a" exitCode=0 Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.062418 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fsds7" event={"ID":"2969563a-1f7d-44b4-8a27-b674b18a7090","Type":"ContainerDied","Data":"59f3876b16132bf82bb6ead05bb924fbadb25ed07384fad8b43600645e75ab4a"} Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.062439 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fsds7" Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.062451 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fsds7" event={"ID":"2969563a-1f7d-44b4-8a27-b674b18a7090","Type":"ContainerDied","Data":"e810d3d3cc4caa5d294cfbdd76a72eaa75c166e9cab85be5baec663278dc6e70"} Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.062510 4810 scope.go:117] "RemoveContainer" containerID="59f3876b16132bf82bb6ead05bb924fbadb25ed07384fad8b43600645e75ab4a" Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.086912 4810 scope.go:117] "RemoveContainer" containerID="3095a9fdb0e88560e6794b8233c468f67d2cc2053d821eeffd46622420346f4b" Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.093742 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fsds7"] Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.100681 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fsds7"] Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.112985 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvhlv\" (UniqueName: \"kubernetes.io/projected/2969563a-1f7d-44b4-8a27-b674b18a7090-kube-api-access-mvhlv\") on node \"crc\" DevicePath \"\"" Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.113175 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2969563a-1f7d-44b4-8a27-b674b18a7090-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.113283 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2969563a-1f7d-44b4-8a27-b674b18a7090-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.122271 4810 scope.go:117] "RemoveContainer" containerID="9181f32f0e8cd1779ccd66f7361479b7db809d644c73f8c354cb81ae35833fdb" Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.136251 4810 scope.go:117] "RemoveContainer" containerID="59f3876b16132bf82bb6ead05bb924fbadb25ed07384fad8b43600645e75ab4a" Dec 01 15:24:50 crc kubenswrapper[4810]: E1201 15:24:50.137055 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59f3876b16132bf82bb6ead05bb924fbadb25ed07384fad8b43600645e75ab4a\": container with ID starting with 59f3876b16132bf82bb6ead05bb924fbadb25ed07384fad8b43600645e75ab4a not found: ID does not exist" containerID="59f3876b16132bf82bb6ead05bb924fbadb25ed07384fad8b43600645e75ab4a" Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.137176 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59f3876b16132bf82bb6ead05bb924fbadb25ed07384fad8b43600645e75ab4a"} err="failed to get container status \"59f3876b16132bf82bb6ead05bb924fbadb25ed07384fad8b43600645e75ab4a\": rpc error: code = NotFound desc = could not find container \"59f3876b16132bf82bb6ead05bb924fbadb25ed07384fad8b43600645e75ab4a\": container with ID starting with 59f3876b16132bf82bb6ead05bb924fbadb25ed07384fad8b43600645e75ab4a not found: ID does not exist" Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.137311 4810 scope.go:117] "RemoveContainer" containerID="3095a9fdb0e88560e6794b8233c468f67d2cc2053d821eeffd46622420346f4b" Dec 01 15:24:50 crc kubenswrapper[4810]: E1201 15:24:50.139030 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3095a9fdb0e88560e6794b8233c468f67d2cc2053d821eeffd46622420346f4b\": container with ID starting with 3095a9fdb0e88560e6794b8233c468f67d2cc2053d821eeffd46622420346f4b not found: ID does not exist" containerID="3095a9fdb0e88560e6794b8233c468f67d2cc2053d821eeffd46622420346f4b" Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.140005 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3095a9fdb0e88560e6794b8233c468f67d2cc2053d821eeffd46622420346f4b"} err="failed to get container status \"3095a9fdb0e88560e6794b8233c468f67d2cc2053d821eeffd46622420346f4b\": rpc error: code = NotFound desc = could not find container \"3095a9fdb0e88560e6794b8233c468f67d2cc2053d821eeffd46622420346f4b\": container with ID starting with 3095a9fdb0e88560e6794b8233c468f67d2cc2053d821eeffd46622420346f4b not found: ID does not exist" Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.140118 4810 scope.go:117] "RemoveContainer" containerID="9181f32f0e8cd1779ccd66f7361479b7db809d644c73f8c354cb81ae35833fdb" Dec 01 15:24:50 crc kubenswrapper[4810]: E1201 15:24:50.141211 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9181f32f0e8cd1779ccd66f7361479b7db809d644c73f8c354cb81ae35833fdb\": container with ID starting with 9181f32f0e8cd1779ccd66f7361479b7db809d644c73f8c354cb81ae35833fdb not found: ID does not exist" containerID="9181f32f0e8cd1779ccd66f7361479b7db809d644c73f8c354cb81ae35833fdb" Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.141291 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9181f32f0e8cd1779ccd66f7361479b7db809d644c73f8c354cb81ae35833fdb"} err="failed to get container status \"9181f32f0e8cd1779ccd66f7361479b7db809d644c73f8c354cb81ae35833fdb\": rpc error: code = NotFound desc = could not find container \"9181f32f0e8cd1779ccd66f7361479b7db809d644c73f8c354cb81ae35833fdb\": container with ID starting with 9181f32f0e8cd1779ccd66f7361479b7db809d644c73f8c354cb81ae35833fdb not found: ID does not exist" Dec 01 15:24:50 crc kubenswrapper[4810]: I1201 15:24:50.504724 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2969563a-1f7d-44b4-8a27-b674b18a7090" path="/var/lib/kubelet/pods/2969563a-1f7d-44b4-8a27-b674b18a7090/volumes" Dec 01 15:25:02 crc kubenswrapper[4810]: I1201 15:25:02.972686 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:25:03 crc kubenswrapper[4810]: I1201 15:25:02.973389 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:25:32 crc kubenswrapper[4810]: I1201 15:25:32.971875 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:25:32 crc kubenswrapper[4810]: I1201 15:25:32.972709 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:26:02 crc kubenswrapper[4810]: I1201 15:26:02.972304 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:26:02 crc kubenswrapper[4810]: I1201 15:26:02.972973 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:26:02 crc kubenswrapper[4810]: I1201 15:26:02.973025 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 15:26:02 crc kubenswrapper[4810]: I1201 15:26:02.973671 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 15:26:02 crc kubenswrapper[4810]: I1201 15:26:02.973728 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" gracePeriod=600 Dec 01 15:26:03 crc kubenswrapper[4810]: E1201 15:26:03.103553 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:26:03 crc kubenswrapper[4810]: I1201 15:26:03.568976 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" exitCode=0 Dec 01 15:26:03 crc kubenswrapper[4810]: I1201 15:26:03.569017 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9"} Dec 01 15:26:03 crc kubenswrapper[4810]: I1201 15:26:03.569102 4810 scope.go:117] "RemoveContainer" containerID="1b6adf628cd09401f04ff4153fb1f5fd44fef07fe578fc9a26d88fa6d79cedb2" Dec 01 15:26:03 crc kubenswrapper[4810]: I1201 15:26:03.569455 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:26:03 crc kubenswrapper[4810]: E1201 15:26:03.569731 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:26:16 crc kubenswrapper[4810]: I1201 15:26:16.495506 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:26:16 crc kubenswrapper[4810]: E1201 15:26:16.496214 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:26:31 crc kubenswrapper[4810]: I1201 15:26:31.491248 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:26:31 crc kubenswrapper[4810]: E1201 15:26:31.493103 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:26:44 crc kubenswrapper[4810]: I1201 15:26:44.496281 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:26:44 crc kubenswrapper[4810]: E1201 15:26:44.497072 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:26:58 crc kubenswrapper[4810]: I1201 15:26:58.490772 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:26:58 crc kubenswrapper[4810]: E1201 15:26:58.491419 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:27:10 crc kubenswrapper[4810]: I1201 15:27:10.491142 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:27:10 crc kubenswrapper[4810]: E1201 15:27:10.491925 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:27:23 crc kubenswrapper[4810]: I1201 15:27:23.490849 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:27:23 crc kubenswrapper[4810]: E1201 15:27:23.491634 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:27:37 crc kubenswrapper[4810]: I1201 15:27:37.491798 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:27:37 crc kubenswrapper[4810]: E1201 15:27:37.492704 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:27:52 crc kubenswrapper[4810]: I1201 15:27:52.491786 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:27:52 crc kubenswrapper[4810]: E1201 15:27:52.492682 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:28:07 crc kubenswrapper[4810]: I1201 15:28:07.491201 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:28:07 crc kubenswrapper[4810]: E1201 15:28:07.492039 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:28:19 crc kubenswrapper[4810]: I1201 15:28:19.492182 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:28:19 crc kubenswrapper[4810]: E1201 15:28:19.493150 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:28:20 crc kubenswrapper[4810]: I1201 15:28:20.552333 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qnw6v"] Dec 01 15:28:20 crc kubenswrapper[4810]: E1201 15:28:20.553204 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2969563a-1f7d-44b4-8a27-b674b18a7090" containerName="extract-utilities" Dec 01 15:28:20 crc kubenswrapper[4810]: I1201 15:28:20.553226 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2969563a-1f7d-44b4-8a27-b674b18a7090" containerName="extract-utilities" Dec 01 15:28:20 crc kubenswrapper[4810]: E1201 15:28:20.553249 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2969563a-1f7d-44b4-8a27-b674b18a7090" containerName="extract-content" Dec 01 15:28:20 crc kubenswrapper[4810]: I1201 15:28:20.553261 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2969563a-1f7d-44b4-8a27-b674b18a7090" containerName="extract-content" Dec 01 15:28:20 crc kubenswrapper[4810]: E1201 15:28:20.553304 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2969563a-1f7d-44b4-8a27-b674b18a7090" containerName="registry-server" Dec 01 15:28:20 crc kubenswrapper[4810]: I1201 15:28:20.553315 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2969563a-1f7d-44b4-8a27-b674b18a7090" containerName="registry-server" Dec 01 15:28:20 crc kubenswrapper[4810]: I1201 15:28:20.553612 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="2969563a-1f7d-44b4-8a27-b674b18a7090" containerName="registry-server" Dec 01 15:28:20 crc kubenswrapper[4810]: I1201 15:28:20.555273 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qnw6v" Dec 01 15:28:20 crc kubenswrapper[4810]: I1201 15:28:20.566415 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qnw6v"] Dec 01 15:28:20 crc kubenswrapper[4810]: I1201 15:28:20.668116 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/203927d5-efbb-4467-8856-9e5ed186f423-utilities\") pod \"redhat-marketplace-qnw6v\" (UID: \"203927d5-efbb-4467-8856-9e5ed186f423\") " pod="openshift-marketplace/redhat-marketplace-qnw6v" Dec 01 15:28:20 crc kubenswrapper[4810]: I1201 15:28:20.668215 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl66t\" (UniqueName: \"kubernetes.io/projected/203927d5-efbb-4467-8856-9e5ed186f423-kube-api-access-sl66t\") pod \"redhat-marketplace-qnw6v\" (UID: \"203927d5-efbb-4467-8856-9e5ed186f423\") " pod="openshift-marketplace/redhat-marketplace-qnw6v" Dec 01 15:28:20 crc kubenswrapper[4810]: I1201 15:28:20.668285 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/203927d5-efbb-4467-8856-9e5ed186f423-catalog-content\") pod \"redhat-marketplace-qnw6v\" (UID: \"203927d5-efbb-4467-8856-9e5ed186f423\") " pod="openshift-marketplace/redhat-marketplace-qnw6v" Dec 01 15:28:20 crc kubenswrapper[4810]: I1201 15:28:20.769872 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/203927d5-efbb-4467-8856-9e5ed186f423-utilities\") pod \"redhat-marketplace-qnw6v\" (UID: \"203927d5-efbb-4467-8856-9e5ed186f423\") " pod="openshift-marketplace/redhat-marketplace-qnw6v" Dec 01 15:28:20 crc kubenswrapper[4810]: I1201 15:28:20.770500 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/203927d5-efbb-4467-8856-9e5ed186f423-utilities\") pod \"redhat-marketplace-qnw6v\" (UID: \"203927d5-efbb-4467-8856-9e5ed186f423\") " pod="openshift-marketplace/redhat-marketplace-qnw6v" Dec 01 15:28:20 crc kubenswrapper[4810]: I1201 15:28:20.770637 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl66t\" (UniqueName: \"kubernetes.io/projected/203927d5-efbb-4467-8856-9e5ed186f423-kube-api-access-sl66t\") pod \"redhat-marketplace-qnw6v\" (UID: \"203927d5-efbb-4467-8856-9e5ed186f423\") " pod="openshift-marketplace/redhat-marketplace-qnw6v" Dec 01 15:28:20 crc kubenswrapper[4810]: I1201 15:28:20.771020 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/203927d5-efbb-4467-8856-9e5ed186f423-catalog-content\") pod \"redhat-marketplace-qnw6v\" (UID: \"203927d5-efbb-4467-8856-9e5ed186f423\") " pod="openshift-marketplace/redhat-marketplace-qnw6v" Dec 01 15:28:20 crc kubenswrapper[4810]: I1201 15:28:20.771323 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/203927d5-efbb-4467-8856-9e5ed186f423-catalog-content\") pod \"redhat-marketplace-qnw6v\" (UID: \"203927d5-efbb-4467-8856-9e5ed186f423\") " pod="openshift-marketplace/redhat-marketplace-qnw6v" Dec 01 15:28:20 crc kubenswrapper[4810]: I1201 15:28:20.795658 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl66t\" (UniqueName: \"kubernetes.io/projected/203927d5-efbb-4467-8856-9e5ed186f423-kube-api-access-sl66t\") pod \"redhat-marketplace-qnw6v\" (UID: \"203927d5-efbb-4467-8856-9e5ed186f423\") " pod="openshift-marketplace/redhat-marketplace-qnw6v" Dec 01 15:28:20 crc kubenswrapper[4810]: I1201 15:28:20.878839 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qnw6v" Dec 01 15:28:21 crc kubenswrapper[4810]: I1201 15:28:21.306647 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qnw6v"] Dec 01 15:28:21 crc kubenswrapper[4810]: I1201 15:28:21.591978 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qnw6v" event={"ID":"203927d5-efbb-4467-8856-9e5ed186f423","Type":"ContainerStarted","Data":"f04469b2d9db687bd001656a4c3d37eebc096403f56b4c1bfe23cb4123f208f0"} Dec 01 15:28:22 crc kubenswrapper[4810]: I1201 15:28:22.598738 4810 generic.go:334] "Generic (PLEG): container finished" podID="203927d5-efbb-4467-8856-9e5ed186f423" containerID="4675a2337f3b2aadbbbdfe10f538a8246a4cf1c601176542fa4bdef4d70157f6" exitCode=0 Dec 01 15:28:22 crc kubenswrapper[4810]: I1201 15:28:22.598826 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qnw6v" event={"ID":"203927d5-efbb-4467-8856-9e5ed186f423","Type":"ContainerDied","Data":"4675a2337f3b2aadbbbdfe10f538a8246a4cf1c601176542fa4bdef4d70157f6"} Dec 01 15:28:23 crc kubenswrapper[4810]: I1201 15:28:23.608404 4810 generic.go:334] "Generic (PLEG): container finished" podID="203927d5-efbb-4467-8856-9e5ed186f423" containerID="74d003210268ba39f865162aaf157555de087541e2dfde343125fb46c8b3f955" exitCode=0 Dec 01 15:28:23 crc kubenswrapper[4810]: I1201 15:28:23.608497 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qnw6v" event={"ID":"203927d5-efbb-4467-8856-9e5ed186f423","Type":"ContainerDied","Data":"74d003210268ba39f865162aaf157555de087541e2dfde343125fb46c8b3f955"} Dec 01 15:28:24 crc kubenswrapper[4810]: I1201 15:28:24.615066 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qnw6v" event={"ID":"203927d5-efbb-4467-8856-9e5ed186f423","Type":"ContainerStarted","Data":"ccdf3c4a837e968d8357948b2473c0f4ef95ddedbbc92c8f3fc96dddc051b997"} Dec 01 15:28:30 crc kubenswrapper[4810]: I1201 15:28:30.880002 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qnw6v" Dec 01 15:28:30 crc kubenswrapper[4810]: I1201 15:28:30.880569 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qnw6v" Dec 01 15:28:30 crc kubenswrapper[4810]: I1201 15:28:30.935464 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qnw6v" Dec 01 15:28:30 crc kubenswrapper[4810]: I1201 15:28:30.957723 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qnw6v" podStartSLOduration=9.369438117 podStartE2EDuration="10.957698341s" podCreationTimestamp="2025-12-01 15:28:20 +0000 UTC" firstStartedPulling="2025-12-01 15:28:22.600196985 +0000 UTC m=+3268.363706588" lastFinishedPulling="2025-12-01 15:28:24.188457169 +0000 UTC m=+3269.951966812" observedRunningTime="2025-12-01 15:28:24.632500081 +0000 UTC m=+3270.396009704" watchObservedRunningTime="2025-12-01 15:28:30.957698341 +0000 UTC m=+3276.721207984" Dec 01 15:28:31 crc kubenswrapper[4810]: I1201 15:28:31.697008 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qnw6v" Dec 01 15:28:31 crc kubenswrapper[4810]: I1201 15:28:31.752061 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qnw6v"] Dec 01 15:28:33 crc kubenswrapper[4810]: I1201 15:28:33.491623 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:28:33 crc kubenswrapper[4810]: E1201 15:28:33.492102 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:28:33 crc kubenswrapper[4810]: I1201 15:28:33.701547 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qnw6v" podUID="203927d5-efbb-4467-8856-9e5ed186f423" containerName="registry-server" containerID="cri-o://ccdf3c4a837e968d8357948b2473c0f4ef95ddedbbc92c8f3fc96dddc051b997" gracePeriod=2 Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.590152 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qnw6v" Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.709566 4810 generic.go:334] "Generic (PLEG): container finished" podID="203927d5-efbb-4467-8856-9e5ed186f423" containerID="ccdf3c4a837e968d8357948b2473c0f4ef95ddedbbc92c8f3fc96dddc051b997" exitCode=0 Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.709619 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qnw6v" event={"ID":"203927d5-efbb-4467-8856-9e5ed186f423","Type":"ContainerDied","Data":"ccdf3c4a837e968d8357948b2473c0f4ef95ddedbbc92c8f3fc96dddc051b997"} Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.709665 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qnw6v" event={"ID":"203927d5-efbb-4467-8856-9e5ed186f423","Type":"ContainerDied","Data":"f04469b2d9db687bd001656a4c3d37eebc096403f56b4c1bfe23cb4123f208f0"} Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.709690 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qnw6v" Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.709713 4810 scope.go:117] "RemoveContainer" containerID="ccdf3c4a837e968d8357948b2473c0f4ef95ddedbbc92c8f3fc96dddc051b997" Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.727046 4810 scope.go:117] "RemoveContainer" containerID="74d003210268ba39f865162aaf157555de087541e2dfde343125fb46c8b3f955" Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.744893 4810 scope.go:117] "RemoveContainer" containerID="4675a2337f3b2aadbbbdfe10f538a8246a4cf1c601176542fa4bdef4d70157f6" Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.758805 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/203927d5-efbb-4467-8856-9e5ed186f423-utilities\") pod \"203927d5-efbb-4467-8856-9e5ed186f423\" (UID: \"203927d5-efbb-4467-8856-9e5ed186f423\") " Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.758955 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sl66t\" (UniqueName: \"kubernetes.io/projected/203927d5-efbb-4467-8856-9e5ed186f423-kube-api-access-sl66t\") pod \"203927d5-efbb-4467-8856-9e5ed186f423\" (UID: \"203927d5-efbb-4467-8856-9e5ed186f423\") " Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.758985 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/203927d5-efbb-4467-8856-9e5ed186f423-catalog-content\") pod \"203927d5-efbb-4467-8856-9e5ed186f423\" (UID: \"203927d5-efbb-4467-8856-9e5ed186f423\") " Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.759803 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/203927d5-efbb-4467-8856-9e5ed186f423-utilities" (OuterVolumeSpecName: "utilities") pod "203927d5-efbb-4467-8856-9e5ed186f423" (UID: "203927d5-efbb-4467-8856-9e5ed186f423"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.765115 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/203927d5-efbb-4467-8856-9e5ed186f423-kube-api-access-sl66t" (OuterVolumeSpecName: "kube-api-access-sl66t") pod "203927d5-efbb-4467-8856-9e5ed186f423" (UID: "203927d5-efbb-4467-8856-9e5ed186f423"). InnerVolumeSpecName "kube-api-access-sl66t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.768638 4810 scope.go:117] "RemoveContainer" containerID="ccdf3c4a837e968d8357948b2473c0f4ef95ddedbbc92c8f3fc96dddc051b997" Dec 01 15:28:34 crc kubenswrapper[4810]: E1201 15:28:34.769059 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccdf3c4a837e968d8357948b2473c0f4ef95ddedbbc92c8f3fc96dddc051b997\": container with ID starting with ccdf3c4a837e968d8357948b2473c0f4ef95ddedbbc92c8f3fc96dddc051b997 not found: ID does not exist" containerID="ccdf3c4a837e968d8357948b2473c0f4ef95ddedbbc92c8f3fc96dddc051b997" Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.769164 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccdf3c4a837e968d8357948b2473c0f4ef95ddedbbc92c8f3fc96dddc051b997"} err="failed to get container status \"ccdf3c4a837e968d8357948b2473c0f4ef95ddedbbc92c8f3fc96dddc051b997\": rpc error: code = NotFound desc = could not find container \"ccdf3c4a837e968d8357948b2473c0f4ef95ddedbbc92c8f3fc96dddc051b997\": container with ID starting with ccdf3c4a837e968d8357948b2473c0f4ef95ddedbbc92c8f3fc96dddc051b997 not found: ID does not exist" Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.769255 4810 scope.go:117] "RemoveContainer" containerID="74d003210268ba39f865162aaf157555de087541e2dfde343125fb46c8b3f955" Dec 01 15:28:34 crc kubenswrapper[4810]: E1201 15:28:34.769550 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74d003210268ba39f865162aaf157555de087541e2dfde343125fb46c8b3f955\": container with ID starting with 74d003210268ba39f865162aaf157555de087541e2dfde343125fb46c8b3f955 not found: ID does not exist" containerID="74d003210268ba39f865162aaf157555de087541e2dfde343125fb46c8b3f955" Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.769642 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74d003210268ba39f865162aaf157555de087541e2dfde343125fb46c8b3f955"} err="failed to get container status \"74d003210268ba39f865162aaf157555de087541e2dfde343125fb46c8b3f955\": rpc error: code = NotFound desc = could not find container \"74d003210268ba39f865162aaf157555de087541e2dfde343125fb46c8b3f955\": container with ID starting with 74d003210268ba39f865162aaf157555de087541e2dfde343125fb46c8b3f955 not found: ID does not exist" Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.769732 4810 scope.go:117] "RemoveContainer" containerID="4675a2337f3b2aadbbbdfe10f538a8246a4cf1c601176542fa4bdef4d70157f6" Dec 01 15:28:34 crc kubenswrapper[4810]: E1201 15:28:34.770098 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4675a2337f3b2aadbbbdfe10f538a8246a4cf1c601176542fa4bdef4d70157f6\": container with ID starting with 4675a2337f3b2aadbbbdfe10f538a8246a4cf1c601176542fa4bdef4d70157f6 not found: ID does not exist" containerID="4675a2337f3b2aadbbbdfe10f538a8246a4cf1c601176542fa4bdef4d70157f6" Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.770194 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4675a2337f3b2aadbbbdfe10f538a8246a4cf1c601176542fa4bdef4d70157f6"} err="failed to get container status \"4675a2337f3b2aadbbbdfe10f538a8246a4cf1c601176542fa4bdef4d70157f6\": rpc error: code = NotFound desc = could not find container \"4675a2337f3b2aadbbbdfe10f538a8246a4cf1c601176542fa4bdef4d70157f6\": container with ID starting with 4675a2337f3b2aadbbbdfe10f538a8246a4cf1c601176542fa4bdef4d70157f6 not found: ID does not exist" Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.777047 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/203927d5-efbb-4467-8856-9e5ed186f423-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "203927d5-efbb-4467-8856-9e5ed186f423" (UID: "203927d5-efbb-4467-8856-9e5ed186f423"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.860366 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/203927d5-efbb-4467-8856-9e5ed186f423-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.860400 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/203927d5-efbb-4467-8856-9e5ed186f423-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:28:34 crc kubenswrapper[4810]: I1201 15:28:34.860418 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sl66t\" (UniqueName: \"kubernetes.io/projected/203927d5-efbb-4467-8856-9e5ed186f423-kube-api-access-sl66t\") on node \"crc\" DevicePath \"\"" Dec 01 15:28:35 crc kubenswrapper[4810]: I1201 15:28:35.044658 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qnw6v"] Dec 01 15:28:35 crc kubenswrapper[4810]: I1201 15:28:35.051384 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qnw6v"] Dec 01 15:28:36 crc kubenswrapper[4810]: I1201 15:28:36.501669 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="203927d5-efbb-4467-8856-9e5ed186f423" path="/var/lib/kubelet/pods/203927d5-efbb-4467-8856-9e5ed186f423/volumes" Dec 01 15:28:44 crc kubenswrapper[4810]: I1201 15:28:44.494989 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:28:44 crc kubenswrapper[4810]: E1201 15:28:44.495832 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:28:57 crc kubenswrapper[4810]: I1201 15:28:57.490507 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:28:57 crc kubenswrapper[4810]: E1201 15:28:57.491774 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:29:08 crc kubenswrapper[4810]: I1201 15:29:08.492215 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:29:08 crc kubenswrapper[4810]: E1201 15:29:08.493132 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:29:19 crc kubenswrapper[4810]: I1201 15:29:19.491080 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:29:19 crc kubenswrapper[4810]: E1201 15:29:19.491862 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:29:33 crc kubenswrapper[4810]: I1201 15:29:33.491487 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:29:33 crc kubenswrapper[4810]: E1201 15:29:33.492337 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:29:48 crc kubenswrapper[4810]: I1201 15:29:48.491308 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:29:48 crc kubenswrapper[4810]: E1201 15:29:48.492025 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.144086 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn"] Dec 01 15:30:00 crc kubenswrapper[4810]: E1201 15:30:00.144944 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="203927d5-efbb-4467-8856-9e5ed186f423" containerName="registry-server" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.144962 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="203927d5-efbb-4467-8856-9e5ed186f423" containerName="registry-server" Dec 01 15:30:00 crc kubenswrapper[4810]: E1201 15:30:00.144996 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="203927d5-efbb-4467-8856-9e5ed186f423" containerName="extract-utilities" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.145007 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="203927d5-efbb-4467-8856-9e5ed186f423" containerName="extract-utilities" Dec 01 15:30:00 crc kubenswrapper[4810]: E1201 15:30:00.145026 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="203927d5-efbb-4467-8856-9e5ed186f423" containerName="extract-content" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.145034 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="203927d5-efbb-4467-8856-9e5ed186f423" containerName="extract-content" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.145216 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="203927d5-efbb-4467-8856-9e5ed186f423" containerName="registry-server" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.145837 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.148435 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.148598 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.153444 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn"] Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.295545 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwvzq\" (UniqueName: \"kubernetes.io/projected/70be3d5e-ec64-4b22-8cc7-c071e13ed43c-kube-api-access-xwvzq\") pod \"collect-profiles-29410050-g29pn\" (UID: \"70be3d5e-ec64-4b22-8cc7-c071e13ed43c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.295615 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/70be3d5e-ec64-4b22-8cc7-c071e13ed43c-secret-volume\") pod \"collect-profiles-29410050-g29pn\" (UID: \"70be3d5e-ec64-4b22-8cc7-c071e13ed43c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.295975 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/70be3d5e-ec64-4b22-8cc7-c071e13ed43c-config-volume\") pod \"collect-profiles-29410050-g29pn\" (UID: \"70be3d5e-ec64-4b22-8cc7-c071e13ed43c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.405464 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwvzq\" (UniqueName: \"kubernetes.io/projected/70be3d5e-ec64-4b22-8cc7-c071e13ed43c-kube-api-access-xwvzq\") pod \"collect-profiles-29410050-g29pn\" (UID: \"70be3d5e-ec64-4b22-8cc7-c071e13ed43c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.405587 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/70be3d5e-ec64-4b22-8cc7-c071e13ed43c-secret-volume\") pod \"collect-profiles-29410050-g29pn\" (UID: \"70be3d5e-ec64-4b22-8cc7-c071e13ed43c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.405644 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/70be3d5e-ec64-4b22-8cc7-c071e13ed43c-config-volume\") pod \"collect-profiles-29410050-g29pn\" (UID: \"70be3d5e-ec64-4b22-8cc7-c071e13ed43c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.407024 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/70be3d5e-ec64-4b22-8cc7-c071e13ed43c-config-volume\") pod \"collect-profiles-29410050-g29pn\" (UID: \"70be3d5e-ec64-4b22-8cc7-c071e13ed43c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.413209 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/70be3d5e-ec64-4b22-8cc7-c071e13ed43c-secret-volume\") pod \"collect-profiles-29410050-g29pn\" (UID: \"70be3d5e-ec64-4b22-8cc7-c071e13ed43c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.424411 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwvzq\" (UniqueName: \"kubernetes.io/projected/70be3d5e-ec64-4b22-8cc7-c071e13ed43c-kube-api-access-xwvzq\") pod \"collect-profiles-29410050-g29pn\" (UID: \"70be3d5e-ec64-4b22-8cc7-c071e13ed43c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.463339 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn" Dec 01 15:30:00 crc kubenswrapper[4810]: I1201 15:30:00.902142 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn"] Dec 01 15:30:01 crc kubenswrapper[4810]: I1201 15:30:01.328288 4810 generic.go:334] "Generic (PLEG): container finished" podID="70be3d5e-ec64-4b22-8cc7-c071e13ed43c" containerID="2195a8be29b3f937a0e6244ce829f622b65caef3a438114604d3e7440f5401dc" exitCode=0 Dec 01 15:30:01 crc kubenswrapper[4810]: I1201 15:30:01.328347 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn" event={"ID":"70be3d5e-ec64-4b22-8cc7-c071e13ed43c","Type":"ContainerDied","Data":"2195a8be29b3f937a0e6244ce829f622b65caef3a438114604d3e7440f5401dc"} Dec 01 15:30:01 crc kubenswrapper[4810]: I1201 15:30:01.328616 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn" event={"ID":"70be3d5e-ec64-4b22-8cc7-c071e13ed43c","Type":"ContainerStarted","Data":"eba72621db134da3eda643d324e34d01593536523249458246642f6b88159d81"} Dec 01 15:30:02 crc kubenswrapper[4810]: I1201 15:30:02.609551 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn" Dec 01 15:30:02 crc kubenswrapper[4810]: I1201 15:30:02.645139 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/70be3d5e-ec64-4b22-8cc7-c071e13ed43c-secret-volume\") pod \"70be3d5e-ec64-4b22-8cc7-c071e13ed43c\" (UID: \"70be3d5e-ec64-4b22-8cc7-c071e13ed43c\") " Dec 01 15:30:02 crc kubenswrapper[4810]: I1201 15:30:02.645305 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwvzq\" (UniqueName: \"kubernetes.io/projected/70be3d5e-ec64-4b22-8cc7-c071e13ed43c-kube-api-access-xwvzq\") pod \"70be3d5e-ec64-4b22-8cc7-c071e13ed43c\" (UID: \"70be3d5e-ec64-4b22-8cc7-c071e13ed43c\") " Dec 01 15:30:02 crc kubenswrapper[4810]: I1201 15:30:02.645333 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/70be3d5e-ec64-4b22-8cc7-c071e13ed43c-config-volume\") pod \"70be3d5e-ec64-4b22-8cc7-c071e13ed43c\" (UID: \"70be3d5e-ec64-4b22-8cc7-c071e13ed43c\") " Dec 01 15:30:02 crc kubenswrapper[4810]: I1201 15:30:02.646118 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70be3d5e-ec64-4b22-8cc7-c071e13ed43c-config-volume" (OuterVolumeSpecName: "config-volume") pod "70be3d5e-ec64-4b22-8cc7-c071e13ed43c" (UID: "70be3d5e-ec64-4b22-8cc7-c071e13ed43c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:30:02 crc kubenswrapper[4810]: I1201 15:30:02.650932 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70be3d5e-ec64-4b22-8cc7-c071e13ed43c-kube-api-access-xwvzq" (OuterVolumeSpecName: "kube-api-access-xwvzq") pod "70be3d5e-ec64-4b22-8cc7-c071e13ed43c" (UID: "70be3d5e-ec64-4b22-8cc7-c071e13ed43c"). InnerVolumeSpecName "kube-api-access-xwvzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:30:02 crc kubenswrapper[4810]: I1201 15:30:02.651621 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70be3d5e-ec64-4b22-8cc7-c071e13ed43c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "70be3d5e-ec64-4b22-8cc7-c071e13ed43c" (UID: "70be3d5e-ec64-4b22-8cc7-c071e13ed43c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:30:02 crc kubenswrapper[4810]: I1201 15:30:02.746583 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/70be3d5e-ec64-4b22-8cc7-c071e13ed43c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 15:30:02 crc kubenswrapper[4810]: I1201 15:30:02.746913 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwvzq\" (UniqueName: \"kubernetes.io/projected/70be3d5e-ec64-4b22-8cc7-c071e13ed43c-kube-api-access-xwvzq\") on node \"crc\" DevicePath \"\"" Dec 01 15:30:02 crc kubenswrapper[4810]: I1201 15:30:02.747016 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/70be3d5e-ec64-4b22-8cc7-c071e13ed43c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 15:30:03 crc kubenswrapper[4810]: I1201 15:30:03.342870 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn" event={"ID":"70be3d5e-ec64-4b22-8cc7-c071e13ed43c","Type":"ContainerDied","Data":"eba72621db134da3eda643d324e34d01593536523249458246642f6b88159d81"} Dec 01 15:30:03 crc kubenswrapper[4810]: I1201 15:30:03.342906 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eba72621db134da3eda643d324e34d01593536523249458246642f6b88159d81" Dec 01 15:30:03 crc kubenswrapper[4810]: I1201 15:30:03.342950 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn" Dec 01 15:30:03 crc kubenswrapper[4810]: I1201 15:30:03.491399 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:30:03 crc kubenswrapper[4810]: E1201 15:30:03.491689 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:30:03 crc kubenswrapper[4810]: I1201 15:30:03.693336 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt"] Dec 01 15:30:03 crc kubenswrapper[4810]: I1201 15:30:03.700714 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410005-fcbwt"] Dec 01 15:30:04 crc kubenswrapper[4810]: I1201 15:30:04.500683 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2067075-0beb-49b1-be3f-2a1577801d4a" path="/var/lib/kubelet/pods/f2067075-0beb-49b1-be3f-2a1577801d4a/volumes" Dec 01 15:30:08 crc kubenswrapper[4810]: I1201 15:30:08.033636 4810 scope.go:117] "RemoveContainer" containerID="c55bddc8877be42932cecfaf18ff65a25c0c87b44045e74379df3031e2f55a6b" Dec 01 15:30:17 crc kubenswrapper[4810]: I1201 15:30:17.491809 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:30:17 crc kubenswrapper[4810]: E1201 15:30:17.492639 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:30:31 crc kubenswrapper[4810]: I1201 15:30:31.491375 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:30:31 crc kubenswrapper[4810]: E1201 15:30:31.492257 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:30:42 crc kubenswrapper[4810]: I1201 15:30:42.491406 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:30:42 crc kubenswrapper[4810]: E1201 15:30:42.492223 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:30:55 crc kubenswrapper[4810]: I1201 15:30:55.491407 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:30:55 crc kubenswrapper[4810]: E1201 15:30:55.492174 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:31:07 crc kubenswrapper[4810]: I1201 15:31:07.560231 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dx5mq"] Dec 01 15:31:07 crc kubenswrapper[4810]: E1201 15:31:07.565043 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70be3d5e-ec64-4b22-8cc7-c071e13ed43c" containerName="collect-profiles" Dec 01 15:31:07 crc kubenswrapper[4810]: I1201 15:31:07.565074 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="70be3d5e-ec64-4b22-8cc7-c071e13ed43c" containerName="collect-profiles" Dec 01 15:31:07 crc kubenswrapper[4810]: I1201 15:31:07.565611 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="70be3d5e-ec64-4b22-8cc7-c071e13ed43c" containerName="collect-profiles" Dec 01 15:31:07 crc kubenswrapper[4810]: I1201 15:31:07.567788 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dx5mq" Dec 01 15:31:07 crc kubenswrapper[4810]: I1201 15:31:07.585131 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dx5mq"] Dec 01 15:31:07 crc kubenswrapper[4810]: I1201 15:31:07.648266 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nw8d\" (UniqueName: \"kubernetes.io/projected/0118223f-ecf1-493d-a1a5-5f30c7022721-kube-api-access-4nw8d\") pod \"certified-operators-dx5mq\" (UID: \"0118223f-ecf1-493d-a1a5-5f30c7022721\") " pod="openshift-marketplace/certified-operators-dx5mq" Dec 01 15:31:07 crc kubenswrapper[4810]: I1201 15:31:07.648489 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0118223f-ecf1-493d-a1a5-5f30c7022721-utilities\") pod \"certified-operators-dx5mq\" (UID: \"0118223f-ecf1-493d-a1a5-5f30c7022721\") " pod="openshift-marketplace/certified-operators-dx5mq" Dec 01 15:31:07 crc kubenswrapper[4810]: I1201 15:31:07.648516 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0118223f-ecf1-493d-a1a5-5f30c7022721-catalog-content\") pod \"certified-operators-dx5mq\" (UID: \"0118223f-ecf1-493d-a1a5-5f30c7022721\") " pod="openshift-marketplace/certified-operators-dx5mq" Dec 01 15:31:07 crc kubenswrapper[4810]: I1201 15:31:07.749703 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0118223f-ecf1-493d-a1a5-5f30c7022721-utilities\") pod \"certified-operators-dx5mq\" (UID: \"0118223f-ecf1-493d-a1a5-5f30c7022721\") " pod="openshift-marketplace/certified-operators-dx5mq" Dec 01 15:31:07 crc kubenswrapper[4810]: I1201 15:31:07.749754 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0118223f-ecf1-493d-a1a5-5f30c7022721-catalog-content\") pod \"certified-operators-dx5mq\" (UID: \"0118223f-ecf1-493d-a1a5-5f30c7022721\") " pod="openshift-marketplace/certified-operators-dx5mq" Dec 01 15:31:07 crc kubenswrapper[4810]: I1201 15:31:07.749795 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nw8d\" (UniqueName: \"kubernetes.io/projected/0118223f-ecf1-493d-a1a5-5f30c7022721-kube-api-access-4nw8d\") pod \"certified-operators-dx5mq\" (UID: \"0118223f-ecf1-493d-a1a5-5f30c7022721\") " pod="openshift-marketplace/certified-operators-dx5mq" Dec 01 15:31:07 crc kubenswrapper[4810]: I1201 15:31:07.750259 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0118223f-ecf1-493d-a1a5-5f30c7022721-utilities\") pod \"certified-operators-dx5mq\" (UID: \"0118223f-ecf1-493d-a1a5-5f30c7022721\") " pod="openshift-marketplace/certified-operators-dx5mq" Dec 01 15:31:07 crc kubenswrapper[4810]: I1201 15:31:07.750332 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0118223f-ecf1-493d-a1a5-5f30c7022721-catalog-content\") pod \"certified-operators-dx5mq\" (UID: \"0118223f-ecf1-493d-a1a5-5f30c7022721\") " pod="openshift-marketplace/certified-operators-dx5mq" Dec 01 15:31:07 crc kubenswrapper[4810]: I1201 15:31:07.770356 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nw8d\" (UniqueName: \"kubernetes.io/projected/0118223f-ecf1-493d-a1a5-5f30c7022721-kube-api-access-4nw8d\") pod \"certified-operators-dx5mq\" (UID: \"0118223f-ecf1-493d-a1a5-5f30c7022721\") " pod="openshift-marketplace/certified-operators-dx5mq" Dec 01 15:31:07 crc kubenswrapper[4810]: I1201 15:31:07.937727 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dx5mq" Dec 01 15:31:08 crc kubenswrapper[4810]: I1201 15:31:08.209201 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dx5mq"] Dec 01 15:31:08 crc kubenswrapper[4810]: I1201 15:31:08.837056 4810 generic.go:334] "Generic (PLEG): container finished" podID="0118223f-ecf1-493d-a1a5-5f30c7022721" containerID="bf509144ee0e17e772e74f778113c75579edf0c3b7ded32789a73e1e38712b6c" exitCode=0 Dec 01 15:31:08 crc kubenswrapper[4810]: I1201 15:31:08.837120 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dx5mq" event={"ID":"0118223f-ecf1-493d-a1a5-5f30c7022721","Type":"ContainerDied","Data":"bf509144ee0e17e772e74f778113c75579edf0c3b7ded32789a73e1e38712b6c"} Dec 01 15:31:08 crc kubenswrapper[4810]: I1201 15:31:08.837370 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dx5mq" event={"ID":"0118223f-ecf1-493d-a1a5-5f30c7022721","Type":"ContainerStarted","Data":"38f12b19a2f3ff23450b2c4fba07d5edfb77b63894468ca466184d8050d74354"} Dec 01 15:31:08 crc kubenswrapper[4810]: I1201 15:31:08.838816 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 15:31:09 crc kubenswrapper[4810]: I1201 15:31:09.491797 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:31:09 crc kubenswrapper[4810]: I1201 15:31:09.847255 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"4173ab52e4f8b87a8612cb1c2625b99d0a9ec9652f422817094bb8e9a3f7c869"} Dec 01 15:31:09 crc kubenswrapper[4810]: I1201 15:31:09.851463 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dx5mq" event={"ID":"0118223f-ecf1-493d-a1a5-5f30c7022721","Type":"ContainerStarted","Data":"a6cdcd83a4c6b11a503da28ac18391033317f7520ec8c9f5d0e77bb15f2268c9"} Dec 01 15:31:10 crc kubenswrapper[4810]: I1201 15:31:10.859814 4810 generic.go:334] "Generic (PLEG): container finished" podID="0118223f-ecf1-493d-a1a5-5f30c7022721" containerID="a6cdcd83a4c6b11a503da28ac18391033317f7520ec8c9f5d0e77bb15f2268c9" exitCode=0 Dec 01 15:31:10 crc kubenswrapper[4810]: I1201 15:31:10.859933 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dx5mq" event={"ID":"0118223f-ecf1-493d-a1a5-5f30c7022721","Type":"ContainerDied","Data":"a6cdcd83a4c6b11a503da28ac18391033317f7520ec8c9f5d0e77bb15f2268c9"} Dec 01 15:31:11 crc kubenswrapper[4810]: I1201 15:31:11.868418 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dx5mq" event={"ID":"0118223f-ecf1-493d-a1a5-5f30c7022721","Type":"ContainerStarted","Data":"252d18b21db78a0371f9672b39ab8416cd0272a6f640096890ecf0bd8ab9744d"} Dec 01 15:31:11 crc kubenswrapper[4810]: I1201 15:31:11.905791 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dx5mq" podStartSLOduration=2.467078483 podStartE2EDuration="4.905770978s" podCreationTimestamp="2025-12-01 15:31:07 +0000 UTC" firstStartedPulling="2025-12-01 15:31:08.838591183 +0000 UTC m=+3434.602100786" lastFinishedPulling="2025-12-01 15:31:11.277283638 +0000 UTC m=+3437.040793281" observedRunningTime="2025-12-01 15:31:11.885238512 +0000 UTC m=+3437.648748145" watchObservedRunningTime="2025-12-01 15:31:11.905770978 +0000 UTC m=+3437.669280581" Dec 01 15:31:17 crc kubenswrapper[4810]: I1201 15:31:17.938058 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dx5mq" Dec 01 15:31:17 crc kubenswrapper[4810]: I1201 15:31:17.938693 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dx5mq" Dec 01 15:31:17 crc kubenswrapper[4810]: I1201 15:31:17.977102 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dx5mq" Dec 01 15:31:18 crc kubenswrapper[4810]: I1201 15:31:18.965251 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dx5mq" Dec 01 15:31:19 crc kubenswrapper[4810]: I1201 15:31:19.013726 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dx5mq"] Dec 01 15:31:20 crc kubenswrapper[4810]: I1201 15:31:20.934532 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dx5mq" podUID="0118223f-ecf1-493d-a1a5-5f30c7022721" containerName="registry-server" containerID="cri-o://252d18b21db78a0371f9672b39ab8416cd0272a6f640096890ecf0bd8ab9744d" gracePeriod=2 Dec 01 15:31:21 crc kubenswrapper[4810]: I1201 15:31:21.834462 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dx5mq" Dec 01 15:31:21 crc kubenswrapper[4810]: I1201 15:31:21.941352 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0118223f-ecf1-493d-a1a5-5f30c7022721-utilities\") pod \"0118223f-ecf1-493d-a1a5-5f30c7022721\" (UID: \"0118223f-ecf1-493d-a1a5-5f30c7022721\") " Dec 01 15:31:21 crc kubenswrapper[4810]: I1201 15:31:21.941540 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0118223f-ecf1-493d-a1a5-5f30c7022721-catalog-content\") pod \"0118223f-ecf1-493d-a1a5-5f30c7022721\" (UID: \"0118223f-ecf1-493d-a1a5-5f30c7022721\") " Dec 01 15:31:21 crc kubenswrapper[4810]: I1201 15:31:21.941594 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nw8d\" (UniqueName: \"kubernetes.io/projected/0118223f-ecf1-493d-a1a5-5f30c7022721-kube-api-access-4nw8d\") pod \"0118223f-ecf1-493d-a1a5-5f30c7022721\" (UID: \"0118223f-ecf1-493d-a1a5-5f30c7022721\") " Dec 01 15:31:21 crc kubenswrapper[4810]: I1201 15:31:21.942347 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0118223f-ecf1-493d-a1a5-5f30c7022721-utilities" (OuterVolumeSpecName: "utilities") pod "0118223f-ecf1-493d-a1a5-5f30c7022721" (UID: "0118223f-ecf1-493d-a1a5-5f30c7022721"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:31:21 crc kubenswrapper[4810]: I1201 15:31:21.948410 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0118223f-ecf1-493d-a1a5-5f30c7022721-kube-api-access-4nw8d" (OuterVolumeSpecName: "kube-api-access-4nw8d") pod "0118223f-ecf1-493d-a1a5-5f30c7022721" (UID: "0118223f-ecf1-493d-a1a5-5f30c7022721"). InnerVolumeSpecName "kube-api-access-4nw8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:31:21 crc kubenswrapper[4810]: I1201 15:31:21.949217 4810 generic.go:334] "Generic (PLEG): container finished" podID="0118223f-ecf1-493d-a1a5-5f30c7022721" containerID="252d18b21db78a0371f9672b39ab8416cd0272a6f640096890ecf0bd8ab9744d" exitCode=0 Dec 01 15:31:21 crc kubenswrapper[4810]: I1201 15:31:21.949273 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dx5mq" event={"ID":"0118223f-ecf1-493d-a1a5-5f30c7022721","Type":"ContainerDied","Data":"252d18b21db78a0371f9672b39ab8416cd0272a6f640096890ecf0bd8ab9744d"} Dec 01 15:31:21 crc kubenswrapper[4810]: I1201 15:31:21.949306 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dx5mq" event={"ID":"0118223f-ecf1-493d-a1a5-5f30c7022721","Type":"ContainerDied","Data":"38f12b19a2f3ff23450b2c4fba07d5edfb77b63894468ca466184d8050d74354"} Dec 01 15:31:21 crc kubenswrapper[4810]: I1201 15:31:21.949327 4810 scope.go:117] "RemoveContainer" containerID="252d18b21db78a0371f9672b39ab8416cd0272a6f640096890ecf0bd8ab9744d" Dec 01 15:31:21 crc kubenswrapper[4810]: I1201 15:31:21.949506 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dx5mq" Dec 01 15:31:21 crc kubenswrapper[4810]: I1201 15:31:21.984264 4810 scope.go:117] "RemoveContainer" containerID="a6cdcd83a4c6b11a503da28ac18391033317f7520ec8c9f5d0e77bb15f2268c9" Dec 01 15:31:22 crc kubenswrapper[4810]: I1201 15:31:22.012229 4810 scope.go:117] "RemoveContainer" containerID="bf509144ee0e17e772e74f778113c75579edf0c3b7ded32789a73e1e38712b6c" Dec 01 15:31:22 crc kubenswrapper[4810]: I1201 15:31:22.014151 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0118223f-ecf1-493d-a1a5-5f30c7022721-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0118223f-ecf1-493d-a1a5-5f30c7022721" (UID: "0118223f-ecf1-493d-a1a5-5f30c7022721"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:31:22 crc kubenswrapper[4810]: I1201 15:31:22.034134 4810 scope.go:117] "RemoveContainer" containerID="252d18b21db78a0371f9672b39ab8416cd0272a6f640096890ecf0bd8ab9744d" Dec 01 15:31:22 crc kubenswrapper[4810]: E1201 15:31:22.034569 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"252d18b21db78a0371f9672b39ab8416cd0272a6f640096890ecf0bd8ab9744d\": container with ID starting with 252d18b21db78a0371f9672b39ab8416cd0272a6f640096890ecf0bd8ab9744d not found: ID does not exist" containerID="252d18b21db78a0371f9672b39ab8416cd0272a6f640096890ecf0bd8ab9744d" Dec 01 15:31:22 crc kubenswrapper[4810]: I1201 15:31:22.034599 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"252d18b21db78a0371f9672b39ab8416cd0272a6f640096890ecf0bd8ab9744d"} err="failed to get container status \"252d18b21db78a0371f9672b39ab8416cd0272a6f640096890ecf0bd8ab9744d\": rpc error: code = NotFound desc = could not find container \"252d18b21db78a0371f9672b39ab8416cd0272a6f640096890ecf0bd8ab9744d\": container with ID starting with 252d18b21db78a0371f9672b39ab8416cd0272a6f640096890ecf0bd8ab9744d not found: ID does not exist" Dec 01 15:31:22 crc kubenswrapper[4810]: I1201 15:31:22.034639 4810 scope.go:117] "RemoveContainer" containerID="a6cdcd83a4c6b11a503da28ac18391033317f7520ec8c9f5d0e77bb15f2268c9" Dec 01 15:31:22 crc kubenswrapper[4810]: E1201 15:31:22.035027 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6cdcd83a4c6b11a503da28ac18391033317f7520ec8c9f5d0e77bb15f2268c9\": container with ID starting with a6cdcd83a4c6b11a503da28ac18391033317f7520ec8c9f5d0e77bb15f2268c9 not found: ID does not exist" containerID="a6cdcd83a4c6b11a503da28ac18391033317f7520ec8c9f5d0e77bb15f2268c9" Dec 01 15:31:22 crc kubenswrapper[4810]: I1201 15:31:22.035070 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6cdcd83a4c6b11a503da28ac18391033317f7520ec8c9f5d0e77bb15f2268c9"} err="failed to get container status \"a6cdcd83a4c6b11a503da28ac18391033317f7520ec8c9f5d0e77bb15f2268c9\": rpc error: code = NotFound desc = could not find container \"a6cdcd83a4c6b11a503da28ac18391033317f7520ec8c9f5d0e77bb15f2268c9\": container with ID starting with a6cdcd83a4c6b11a503da28ac18391033317f7520ec8c9f5d0e77bb15f2268c9 not found: ID does not exist" Dec 01 15:31:22 crc kubenswrapper[4810]: I1201 15:31:22.035088 4810 scope.go:117] "RemoveContainer" containerID="bf509144ee0e17e772e74f778113c75579edf0c3b7ded32789a73e1e38712b6c" Dec 01 15:31:22 crc kubenswrapper[4810]: E1201 15:31:22.035304 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf509144ee0e17e772e74f778113c75579edf0c3b7ded32789a73e1e38712b6c\": container with ID starting with bf509144ee0e17e772e74f778113c75579edf0c3b7ded32789a73e1e38712b6c not found: ID does not exist" containerID="bf509144ee0e17e772e74f778113c75579edf0c3b7ded32789a73e1e38712b6c" Dec 01 15:31:22 crc kubenswrapper[4810]: I1201 15:31:22.035329 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf509144ee0e17e772e74f778113c75579edf0c3b7ded32789a73e1e38712b6c"} err="failed to get container status \"bf509144ee0e17e772e74f778113c75579edf0c3b7ded32789a73e1e38712b6c\": rpc error: code = NotFound desc = could not find container \"bf509144ee0e17e772e74f778113c75579edf0c3b7ded32789a73e1e38712b6c\": container with ID starting with bf509144ee0e17e772e74f778113c75579edf0c3b7ded32789a73e1e38712b6c not found: ID does not exist" Dec 01 15:31:22 crc kubenswrapper[4810]: I1201 15:31:22.042997 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0118223f-ecf1-493d-a1a5-5f30c7022721-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:31:22 crc kubenswrapper[4810]: I1201 15:31:22.043023 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nw8d\" (UniqueName: \"kubernetes.io/projected/0118223f-ecf1-493d-a1a5-5f30c7022721-kube-api-access-4nw8d\") on node \"crc\" DevicePath \"\"" Dec 01 15:31:22 crc kubenswrapper[4810]: I1201 15:31:22.043035 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0118223f-ecf1-493d-a1a5-5f30c7022721-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:31:22 crc kubenswrapper[4810]: I1201 15:31:22.293171 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dx5mq"] Dec 01 15:31:22 crc kubenswrapper[4810]: I1201 15:31:22.300343 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dx5mq"] Dec 01 15:31:22 crc kubenswrapper[4810]: I1201 15:31:22.502646 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0118223f-ecf1-493d-a1a5-5f30c7022721" path="/var/lib/kubelet/pods/0118223f-ecf1-493d-a1a5-5f30c7022721/volumes" Dec 01 15:33:32 crc kubenswrapper[4810]: I1201 15:33:32.972613 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:33:32 crc kubenswrapper[4810]: I1201 15:33:32.973233 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:34:02 crc kubenswrapper[4810]: I1201 15:34:02.972713 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:34:02 crc kubenswrapper[4810]: I1201 15:34:02.973276 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:34:32 crc kubenswrapper[4810]: I1201 15:34:32.972568 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:34:32 crc kubenswrapper[4810]: I1201 15:34:32.973494 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:34:32 crc kubenswrapper[4810]: I1201 15:34:32.973577 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 15:34:32 crc kubenswrapper[4810]: I1201 15:34:32.974281 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4173ab52e4f8b87a8612cb1c2625b99d0a9ec9652f422817094bb8e9a3f7c869"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 15:34:32 crc kubenswrapper[4810]: I1201 15:34:32.974339 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://4173ab52e4f8b87a8612cb1c2625b99d0a9ec9652f422817094bb8e9a3f7c869" gracePeriod=600 Dec 01 15:34:33 crc kubenswrapper[4810]: I1201 15:34:33.365283 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="4173ab52e4f8b87a8612cb1c2625b99d0a9ec9652f422817094bb8e9a3f7c869" exitCode=0 Dec 01 15:34:33 crc kubenswrapper[4810]: I1201 15:34:33.365643 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"4173ab52e4f8b87a8612cb1c2625b99d0a9ec9652f422817094bb8e9a3f7c869"} Dec 01 15:34:33 crc kubenswrapper[4810]: I1201 15:34:33.365669 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096"} Dec 01 15:34:33 crc kubenswrapper[4810]: I1201 15:34:33.365707 4810 scope.go:117] "RemoveContainer" containerID="e45bc785540c7110828a2a301601279fe64a12090db9036f7abf853c3ed590d9" Dec 01 15:35:34 crc kubenswrapper[4810]: I1201 15:35:34.251788 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mkkfc"] Dec 01 15:35:34 crc kubenswrapper[4810]: E1201 15:35:34.252623 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0118223f-ecf1-493d-a1a5-5f30c7022721" containerName="extract-content" Dec 01 15:35:34 crc kubenswrapper[4810]: I1201 15:35:34.252637 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0118223f-ecf1-493d-a1a5-5f30c7022721" containerName="extract-content" Dec 01 15:35:34 crc kubenswrapper[4810]: E1201 15:35:34.252650 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0118223f-ecf1-493d-a1a5-5f30c7022721" containerName="registry-server" Dec 01 15:35:34 crc kubenswrapper[4810]: I1201 15:35:34.252656 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0118223f-ecf1-493d-a1a5-5f30c7022721" containerName="registry-server" Dec 01 15:35:34 crc kubenswrapper[4810]: E1201 15:35:34.252669 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0118223f-ecf1-493d-a1a5-5f30c7022721" containerName="extract-utilities" Dec 01 15:35:34 crc kubenswrapper[4810]: I1201 15:35:34.252677 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0118223f-ecf1-493d-a1a5-5f30c7022721" containerName="extract-utilities" Dec 01 15:35:34 crc kubenswrapper[4810]: I1201 15:35:34.252797 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0118223f-ecf1-493d-a1a5-5f30c7022721" containerName="registry-server" Dec 01 15:35:34 crc kubenswrapper[4810]: I1201 15:35:34.253788 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mkkfc" Dec 01 15:35:34 crc kubenswrapper[4810]: I1201 15:35:34.266105 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mkkfc"] Dec 01 15:35:34 crc kubenswrapper[4810]: I1201 15:35:34.342095 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjl2x\" (UniqueName: \"kubernetes.io/projected/ec5aaa18-c499-41ec-8104-ed0eed4a69b7-kube-api-access-bjl2x\") pod \"redhat-operators-mkkfc\" (UID: \"ec5aaa18-c499-41ec-8104-ed0eed4a69b7\") " pod="openshift-marketplace/redhat-operators-mkkfc" Dec 01 15:35:34 crc kubenswrapper[4810]: I1201 15:35:34.342140 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec5aaa18-c499-41ec-8104-ed0eed4a69b7-utilities\") pod \"redhat-operators-mkkfc\" (UID: \"ec5aaa18-c499-41ec-8104-ed0eed4a69b7\") " pod="openshift-marketplace/redhat-operators-mkkfc" Dec 01 15:35:34 crc kubenswrapper[4810]: I1201 15:35:34.342214 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec5aaa18-c499-41ec-8104-ed0eed4a69b7-catalog-content\") pod \"redhat-operators-mkkfc\" (UID: \"ec5aaa18-c499-41ec-8104-ed0eed4a69b7\") " pod="openshift-marketplace/redhat-operators-mkkfc" Dec 01 15:35:34 crc kubenswrapper[4810]: I1201 15:35:34.443542 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec5aaa18-c499-41ec-8104-ed0eed4a69b7-catalog-content\") pod \"redhat-operators-mkkfc\" (UID: \"ec5aaa18-c499-41ec-8104-ed0eed4a69b7\") " pod="openshift-marketplace/redhat-operators-mkkfc" Dec 01 15:35:34 crc kubenswrapper[4810]: I1201 15:35:34.443667 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjl2x\" (UniqueName: \"kubernetes.io/projected/ec5aaa18-c499-41ec-8104-ed0eed4a69b7-kube-api-access-bjl2x\") pod \"redhat-operators-mkkfc\" (UID: \"ec5aaa18-c499-41ec-8104-ed0eed4a69b7\") " pod="openshift-marketplace/redhat-operators-mkkfc" Dec 01 15:35:34 crc kubenswrapper[4810]: I1201 15:35:34.443702 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec5aaa18-c499-41ec-8104-ed0eed4a69b7-utilities\") pod \"redhat-operators-mkkfc\" (UID: \"ec5aaa18-c499-41ec-8104-ed0eed4a69b7\") " pod="openshift-marketplace/redhat-operators-mkkfc" Dec 01 15:35:34 crc kubenswrapper[4810]: I1201 15:35:34.444124 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec5aaa18-c499-41ec-8104-ed0eed4a69b7-catalog-content\") pod \"redhat-operators-mkkfc\" (UID: \"ec5aaa18-c499-41ec-8104-ed0eed4a69b7\") " pod="openshift-marketplace/redhat-operators-mkkfc" Dec 01 15:35:34 crc kubenswrapper[4810]: I1201 15:35:34.444202 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec5aaa18-c499-41ec-8104-ed0eed4a69b7-utilities\") pod \"redhat-operators-mkkfc\" (UID: \"ec5aaa18-c499-41ec-8104-ed0eed4a69b7\") " pod="openshift-marketplace/redhat-operators-mkkfc" Dec 01 15:35:34 crc kubenswrapper[4810]: I1201 15:35:34.464578 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjl2x\" (UniqueName: \"kubernetes.io/projected/ec5aaa18-c499-41ec-8104-ed0eed4a69b7-kube-api-access-bjl2x\") pod \"redhat-operators-mkkfc\" (UID: \"ec5aaa18-c499-41ec-8104-ed0eed4a69b7\") " pod="openshift-marketplace/redhat-operators-mkkfc" Dec 01 15:35:34 crc kubenswrapper[4810]: I1201 15:35:34.575445 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mkkfc" Dec 01 15:35:34 crc kubenswrapper[4810]: I1201 15:35:34.995587 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mkkfc"] Dec 01 15:35:35 crc kubenswrapper[4810]: I1201 15:35:35.870320 4810 generic.go:334] "Generic (PLEG): container finished" podID="ec5aaa18-c499-41ec-8104-ed0eed4a69b7" containerID="ab6384597f0185d8863505efb9559fa095fe3b2b1464d13561443bbe4d03974b" exitCode=0 Dec 01 15:35:35 crc kubenswrapper[4810]: I1201 15:35:35.870365 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkkfc" event={"ID":"ec5aaa18-c499-41ec-8104-ed0eed4a69b7","Type":"ContainerDied","Data":"ab6384597f0185d8863505efb9559fa095fe3b2b1464d13561443bbe4d03974b"} Dec 01 15:35:35 crc kubenswrapper[4810]: I1201 15:35:35.870388 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkkfc" event={"ID":"ec5aaa18-c499-41ec-8104-ed0eed4a69b7","Type":"ContainerStarted","Data":"40391c49a7c052ceeb5cfc20b1d8c29226d1c3ba2a9de8ef62f6c782e4b648a7"} Dec 01 15:35:36 crc kubenswrapper[4810]: I1201 15:35:36.879791 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkkfc" event={"ID":"ec5aaa18-c499-41ec-8104-ed0eed4a69b7","Type":"ContainerStarted","Data":"859e9297127354f6ddd5ba9cebd995d291aaaff990e89c80add0002e6235bdd6"} Dec 01 15:35:37 crc kubenswrapper[4810]: I1201 15:35:37.888518 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkkfc" event={"ID":"ec5aaa18-c499-41ec-8104-ed0eed4a69b7","Type":"ContainerDied","Data":"859e9297127354f6ddd5ba9cebd995d291aaaff990e89c80add0002e6235bdd6"} Dec 01 15:35:37 crc kubenswrapper[4810]: I1201 15:35:37.888463 4810 generic.go:334] "Generic (PLEG): container finished" podID="ec5aaa18-c499-41ec-8104-ed0eed4a69b7" containerID="859e9297127354f6ddd5ba9cebd995d291aaaff990e89c80add0002e6235bdd6" exitCode=0 Dec 01 15:35:38 crc kubenswrapper[4810]: I1201 15:35:38.901952 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkkfc" event={"ID":"ec5aaa18-c499-41ec-8104-ed0eed4a69b7","Type":"ContainerStarted","Data":"068ec9f7f47068c663215dfa6f0c5d124415de53af296558ed5a0751899a1591"} Dec 01 15:35:38 crc kubenswrapper[4810]: I1201 15:35:38.925381 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mkkfc" podStartSLOduration=2.447498283 podStartE2EDuration="4.92535578s" podCreationTimestamp="2025-12-01 15:35:34 +0000 UTC" firstStartedPulling="2025-12-01 15:35:35.871449693 +0000 UTC m=+3701.634959296" lastFinishedPulling="2025-12-01 15:35:38.34930717 +0000 UTC m=+3704.112816793" observedRunningTime="2025-12-01 15:35:38.921574107 +0000 UTC m=+3704.685083720" watchObservedRunningTime="2025-12-01 15:35:38.92535578 +0000 UTC m=+3704.688865383" Dec 01 15:35:44 crc kubenswrapper[4810]: I1201 15:35:44.575866 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mkkfc" Dec 01 15:35:44 crc kubenswrapper[4810]: I1201 15:35:44.576414 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mkkfc" Dec 01 15:35:44 crc kubenswrapper[4810]: I1201 15:35:44.623337 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mkkfc" Dec 01 15:35:45 crc kubenswrapper[4810]: I1201 15:35:45.005141 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mkkfc" Dec 01 15:35:45 crc kubenswrapper[4810]: I1201 15:35:45.067084 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mkkfc"] Dec 01 15:35:46 crc kubenswrapper[4810]: I1201 15:35:46.963122 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mkkfc" podUID="ec5aaa18-c499-41ec-8104-ed0eed4a69b7" containerName="registry-server" containerID="cri-o://068ec9f7f47068c663215dfa6f0c5d124415de53af296558ed5a0751899a1591" gracePeriod=2 Dec 01 15:35:47 crc kubenswrapper[4810]: I1201 15:35:47.373457 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mkkfc" Dec 01 15:35:47 crc kubenswrapper[4810]: I1201 15:35:47.427531 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjl2x\" (UniqueName: \"kubernetes.io/projected/ec5aaa18-c499-41ec-8104-ed0eed4a69b7-kube-api-access-bjl2x\") pod \"ec5aaa18-c499-41ec-8104-ed0eed4a69b7\" (UID: \"ec5aaa18-c499-41ec-8104-ed0eed4a69b7\") " Dec 01 15:35:47 crc kubenswrapper[4810]: I1201 15:35:47.427590 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec5aaa18-c499-41ec-8104-ed0eed4a69b7-utilities\") pod \"ec5aaa18-c499-41ec-8104-ed0eed4a69b7\" (UID: \"ec5aaa18-c499-41ec-8104-ed0eed4a69b7\") " Dec 01 15:35:47 crc kubenswrapper[4810]: I1201 15:35:47.427629 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec5aaa18-c499-41ec-8104-ed0eed4a69b7-catalog-content\") pod \"ec5aaa18-c499-41ec-8104-ed0eed4a69b7\" (UID: \"ec5aaa18-c499-41ec-8104-ed0eed4a69b7\") " Dec 01 15:35:47 crc kubenswrapper[4810]: I1201 15:35:47.428407 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec5aaa18-c499-41ec-8104-ed0eed4a69b7-utilities" (OuterVolumeSpecName: "utilities") pod "ec5aaa18-c499-41ec-8104-ed0eed4a69b7" (UID: "ec5aaa18-c499-41ec-8104-ed0eed4a69b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:35:47 crc kubenswrapper[4810]: I1201 15:35:47.433121 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec5aaa18-c499-41ec-8104-ed0eed4a69b7-kube-api-access-bjl2x" (OuterVolumeSpecName: "kube-api-access-bjl2x") pod "ec5aaa18-c499-41ec-8104-ed0eed4a69b7" (UID: "ec5aaa18-c499-41ec-8104-ed0eed4a69b7"). InnerVolumeSpecName "kube-api-access-bjl2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:35:47 crc kubenswrapper[4810]: I1201 15:35:47.528830 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjl2x\" (UniqueName: \"kubernetes.io/projected/ec5aaa18-c499-41ec-8104-ed0eed4a69b7-kube-api-access-bjl2x\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:47 crc kubenswrapper[4810]: I1201 15:35:47.528867 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec5aaa18-c499-41ec-8104-ed0eed4a69b7-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:47 crc kubenswrapper[4810]: I1201 15:35:47.974466 4810 generic.go:334] "Generic (PLEG): container finished" podID="ec5aaa18-c499-41ec-8104-ed0eed4a69b7" containerID="068ec9f7f47068c663215dfa6f0c5d124415de53af296558ed5a0751899a1591" exitCode=0 Dec 01 15:35:47 crc kubenswrapper[4810]: I1201 15:35:47.974551 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkkfc" event={"ID":"ec5aaa18-c499-41ec-8104-ed0eed4a69b7","Type":"ContainerDied","Data":"068ec9f7f47068c663215dfa6f0c5d124415de53af296558ed5a0751899a1591"} Dec 01 15:35:47 crc kubenswrapper[4810]: I1201 15:35:47.974642 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkkfc" event={"ID":"ec5aaa18-c499-41ec-8104-ed0eed4a69b7","Type":"ContainerDied","Data":"40391c49a7c052ceeb5cfc20b1d8c29226d1c3ba2a9de8ef62f6c782e4b648a7"} Dec 01 15:35:47 crc kubenswrapper[4810]: I1201 15:35:47.974687 4810 scope.go:117] "RemoveContainer" containerID="068ec9f7f47068c663215dfa6f0c5d124415de53af296558ed5a0751899a1591" Dec 01 15:35:47 crc kubenswrapper[4810]: I1201 15:35:47.975654 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mkkfc" Dec 01 15:35:47 crc kubenswrapper[4810]: I1201 15:35:47.991176 4810 scope.go:117] "RemoveContainer" containerID="859e9297127354f6ddd5ba9cebd995d291aaaff990e89c80add0002e6235bdd6" Dec 01 15:35:48 crc kubenswrapper[4810]: I1201 15:35:48.006649 4810 scope.go:117] "RemoveContainer" containerID="ab6384597f0185d8863505efb9559fa095fe3b2b1464d13561443bbe4d03974b" Dec 01 15:35:48 crc kubenswrapper[4810]: I1201 15:35:48.039240 4810 scope.go:117] "RemoveContainer" containerID="068ec9f7f47068c663215dfa6f0c5d124415de53af296558ed5a0751899a1591" Dec 01 15:35:48 crc kubenswrapper[4810]: E1201 15:35:48.039806 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"068ec9f7f47068c663215dfa6f0c5d124415de53af296558ed5a0751899a1591\": container with ID starting with 068ec9f7f47068c663215dfa6f0c5d124415de53af296558ed5a0751899a1591 not found: ID does not exist" containerID="068ec9f7f47068c663215dfa6f0c5d124415de53af296558ed5a0751899a1591" Dec 01 15:35:48 crc kubenswrapper[4810]: I1201 15:35:48.039923 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"068ec9f7f47068c663215dfa6f0c5d124415de53af296558ed5a0751899a1591"} err="failed to get container status \"068ec9f7f47068c663215dfa6f0c5d124415de53af296558ed5a0751899a1591\": rpc error: code = NotFound desc = could not find container \"068ec9f7f47068c663215dfa6f0c5d124415de53af296558ed5a0751899a1591\": container with ID starting with 068ec9f7f47068c663215dfa6f0c5d124415de53af296558ed5a0751899a1591 not found: ID does not exist" Dec 01 15:35:48 crc kubenswrapper[4810]: I1201 15:35:48.040012 4810 scope.go:117] "RemoveContainer" containerID="859e9297127354f6ddd5ba9cebd995d291aaaff990e89c80add0002e6235bdd6" Dec 01 15:35:48 crc kubenswrapper[4810]: E1201 15:35:48.040441 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"859e9297127354f6ddd5ba9cebd995d291aaaff990e89c80add0002e6235bdd6\": container with ID starting with 859e9297127354f6ddd5ba9cebd995d291aaaff990e89c80add0002e6235bdd6 not found: ID does not exist" containerID="859e9297127354f6ddd5ba9cebd995d291aaaff990e89c80add0002e6235bdd6" Dec 01 15:35:48 crc kubenswrapper[4810]: I1201 15:35:48.040491 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"859e9297127354f6ddd5ba9cebd995d291aaaff990e89c80add0002e6235bdd6"} err="failed to get container status \"859e9297127354f6ddd5ba9cebd995d291aaaff990e89c80add0002e6235bdd6\": rpc error: code = NotFound desc = could not find container \"859e9297127354f6ddd5ba9cebd995d291aaaff990e89c80add0002e6235bdd6\": container with ID starting with 859e9297127354f6ddd5ba9cebd995d291aaaff990e89c80add0002e6235bdd6 not found: ID does not exist" Dec 01 15:35:48 crc kubenswrapper[4810]: I1201 15:35:48.040524 4810 scope.go:117] "RemoveContainer" containerID="ab6384597f0185d8863505efb9559fa095fe3b2b1464d13561443bbe4d03974b" Dec 01 15:35:48 crc kubenswrapper[4810]: E1201 15:35:48.040888 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab6384597f0185d8863505efb9559fa095fe3b2b1464d13561443bbe4d03974b\": container with ID starting with ab6384597f0185d8863505efb9559fa095fe3b2b1464d13561443bbe4d03974b not found: ID does not exist" containerID="ab6384597f0185d8863505efb9559fa095fe3b2b1464d13561443bbe4d03974b" Dec 01 15:35:48 crc kubenswrapper[4810]: I1201 15:35:48.040931 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab6384597f0185d8863505efb9559fa095fe3b2b1464d13561443bbe4d03974b"} err="failed to get container status \"ab6384597f0185d8863505efb9559fa095fe3b2b1464d13561443bbe4d03974b\": rpc error: code = NotFound desc = could not find container \"ab6384597f0185d8863505efb9559fa095fe3b2b1464d13561443bbe4d03974b\": container with ID starting with ab6384597f0185d8863505efb9559fa095fe3b2b1464d13561443bbe4d03974b not found: ID does not exist" Dec 01 15:35:48 crc kubenswrapper[4810]: I1201 15:35:48.439353 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec5aaa18-c499-41ec-8104-ed0eed4a69b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ec5aaa18-c499-41ec-8104-ed0eed4a69b7" (UID: "ec5aaa18-c499-41ec-8104-ed0eed4a69b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:35:48 crc kubenswrapper[4810]: I1201 15:35:48.443654 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec5aaa18-c499-41ec-8104-ed0eed4a69b7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:35:48 crc kubenswrapper[4810]: I1201 15:35:48.593912 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mkkfc"] Dec 01 15:35:48 crc kubenswrapper[4810]: I1201 15:35:48.600556 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mkkfc"] Dec 01 15:35:50 crc kubenswrapper[4810]: I1201 15:35:50.505803 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec5aaa18-c499-41ec-8104-ed0eed4a69b7" path="/var/lib/kubelet/pods/ec5aaa18-c499-41ec-8104-ed0eed4a69b7/volumes" Dec 01 15:35:58 crc kubenswrapper[4810]: I1201 15:35:58.304645 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fqnhb"] Dec 01 15:35:58 crc kubenswrapper[4810]: E1201 15:35:58.307707 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec5aaa18-c499-41ec-8104-ed0eed4a69b7" containerName="registry-server" Dec 01 15:35:58 crc kubenswrapper[4810]: I1201 15:35:58.307730 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec5aaa18-c499-41ec-8104-ed0eed4a69b7" containerName="registry-server" Dec 01 15:35:58 crc kubenswrapper[4810]: E1201 15:35:58.307744 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec5aaa18-c499-41ec-8104-ed0eed4a69b7" containerName="extract-content" Dec 01 15:35:58 crc kubenswrapper[4810]: I1201 15:35:58.307752 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec5aaa18-c499-41ec-8104-ed0eed4a69b7" containerName="extract-content" Dec 01 15:35:58 crc kubenswrapper[4810]: E1201 15:35:58.307781 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec5aaa18-c499-41ec-8104-ed0eed4a69b7" containerName="extract-utilities" Dec 01 15:35:58 crc kubenswrapper[4810]: I1201 15:35:58.307789 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec5aaa18-c499-41ec-8104-ed0eed4a69b7" containerName="extract-utilities" Dec 01 15:35:58 crc kubenswrapper[4810]: I1201 15:35:58.307984 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec5aaa18-c499-41ec-8104-ed0eed4a69b7" containerName="registry-server" Dec 01 15:35:58 crc kubenswrapper[4810]: I1201 15:35:58.310587 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fqnhb" Dec 01 15:35:58 crc kubenswrapper[4810]: I1201 15:35:58.313886 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fqnhb"] Dec 01 15:35:58 crc kubenswrapper[4810]: I1201 15:35:58.386027 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9df5eda5-4d25-48e2-8648-65d61297dd42-utilities\") pod \"community-operators-fqnhb\" (UID: \"9df5eda5-4d25-48e2-8648-65d61297dd42\") " pod="openshift-marketplace/community-operators-fqnhb" Dec 01 15:35:58 crc kubenswrapper[4810]: I1201 15:35:58.386121 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmwmm\" (UniqueName: \"kubernetes.io/projected/9df5eda5-4d25-48e2-8648-65d61297dd42-kube-api-access-tmwmm\") pod \"community-operators-fqnhb\" (UID: \"9df5eda5-4d25-48e2-8648-65d61297dd42\") " pod="openshift-marketplace/community-operators-fqnhb" Dec 01 15:35:58 crc kubenswrapper[4810]: I1201 15:35:58.386160 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9df5eda5-4d25-48e2-8648-65d61297dd42-catalog-content\") pod \"community-operators-fqnhb\" (UID: \"9df5eda5-4d25-48e2-8648-65d61297dd42\") " pod="openshift-marketplace/community-operators-fqnhb" Dec 01 15:35:58 crc kubenswrapper[4810]: I1201 15:35:58.488026 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9df5eda5-4d25-48e2-8648-65d61297dd42-utilities\") pod \"community-operators-fqnhb\" (UID: \"9df5eda5-4d25-48e2-8648-65d61297dd42\") " pod="openshift-marketplace/community-operators-fqnhb" Dec 01 15:35:58 crc kubenswrapper[4810]: I1201 15:35:58.488109 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmwmm\" (UniqueName: \"kubernetes.io/projected/9df5eda5-4d25-48e2-8648-65d61297dd42-kube-api-access-tmwmm\") pod \"community-operators-fqnhb\" (UID: \"9df5eda5-4d25-48e2-8648-65d61297dd42\") " pod="openshift-marketplace/community-operators-fqnhb" Dec 01 15:35:58 crc kubenswrapper[4810]: I1201 15:35:58.488141 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9df5eda5-4d25-48e2-8648-65d61297dd42-catalog-content\") pod \"community-operators-fqnhb\" (UID: \"9df5eda5-4d25-48e2-8648-65d61297dd42\") " pod="openshift-marketplace/community-operators-fqnhb" Dec 01 15:35:58 crc kubenswrapper[4810]: I1201 15:35:58.488555 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9df5eda5-4d25-48e2-8648-65d61297dd42-utilities\") pod \"community-operators-fqnhb\" (UID: \"9df5eda5-4d25-48e2-8648-65d61297dd42\") " pod="openshift-marketplace/community-operators-fqnhb" Dec 01 15:35:58 crc kubenswrapper[4810]: I1201 15:35:58.488621 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9df5eda5-4d25-48e2-8648-65d61297dd42-catalog-content\") pod \"community-operators-fqnhb\" (UID: \"9df5eda5-4d25-48e2-8648-65d61297dd42\") " pod="openshift-marketplace/community-operators-fqnhb" Dec 01 15:35:58 crc kubenswrapper[4810]: I1201 15:35:58.514583 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmwmm\" (UniqueName: \"kubernetes.io/projected/9df5eda5-4d25-48e2-8648-65d61297dd42-kube-api-access-tmwmm\") pod \"community-operators-fqnhb\" (UID: \"9df5eda5-4d25-48e2-8648-65d61297dd42\") " pod="openshift-marketplace/community-operators-fqnhb" Dec 01 15:35:58 crc kubenswrapper[4810]: I1201 15:35:58.636356 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fqnhb" Dec 01 15:35:59 crc kubenswrapper[4810]: I1201 15:35:59.091810 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fqnhb"] Dec 01 15:36:00 crc kubenswrapper[4810]: I1201 15:36:00.067695 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqnhb" event={"ID":"9df5eda5-4d25-48e2-8648-65d61297dd42","Type":"ContainerStarted","Data":"b10c31a2e7a55f0dc5e2c1e326c241f8362e41695724e5d6bc446b4d32733239"} Dec 01 15:36:01 crc kubenswrapper[4810]: I1201 15:36:01.076741 4810 generic.go:334] "Generic (PLEG): container finished" podID="9df5eda5-4d25-48e2-8648-65d61297dd42" containerID="b4f4f96fcb8fdb2752cde01fbc615e67112b2d945933e484f5bca4779a414991" exitCode=0 Dec 01 15:36:01 crc kubenswrapper[4810]: I1201 15:36:01.077014 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqnhb" event={"ID":"9df5eda5-4d25-48e2-8648-65d61297dd42","Type":"ContainerDied","Data":"b4f4f96fcb8fdb2752cde01fbc615e67112b2d945933e484f5bca4779a414991"} Dec 01 15:36:02 crc kubenswrapper[4810]: I1201 15:36:02.084873 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqnhb" event={"ID":"9df5eda5-4d25-48e2-8648-65d61297dd42","Type":"ContainerStarted","Data":"1eb97462996ffa94c5b99d5a801fe56e7300776b38ba3d7e99fa7985e8ff6dad"} Dec 01 15:36:03 crc kubenswrapper[4810]: I1201 15:36:03.096645 4810 generic.go:334] "Generic (PLEG): container finished" podID="9df5eda5-4d25-48e2-8648-65d61297dd42" containerID="1eb97462996ffa94c5b99d5a801fe56e7300776b38ba3d7e99fa7985e8ff6dad" exitCode=0 Dec 01 15:36:03 crc kubenswrapper[4810]: I1201 15:36:03.096685 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqnhb" event={"ID":"9df5eda5-4d25-48e2-8648-65d61297dd42","Type":"ContainerDied","Data":"1eb97462996ffa94c5b99d5a801fe56e7300776b38ba3d7e99fa7985e8ff6dad"} Dec 01 15:36:04 crc kubenswrapper[4810]: I1201 15:36:04.104287 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqnhb" event={"ID":"9df5eda5-4d25-48e2-8648-65d61297dd42","Type":"ContainerStarted","Data":"a51da69eff153c0d11231ccf5fe84eb79b7419b13c4e22137b9af12a0fdc0760"} Dec 01 15:36:04 crc kubenswrapper[4810]: I1201 15:36:04.127634 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fqnhb" podStartSLOduration=3.402288359 podStartE2EDuration="6.127612868s" podCreationTimestamp="2025-12-01 15:35:58 +0000 UTC" firstStartedPulling="2025-12-01 15:36:01.078580707 +0000 UTC m=+3726.842090310" lastFinishedPulling="2025-12-01 15:36:03.803905216 +0000 UTC m=+3729.567414819" observedRunningTime="2025-12-01 15:36:04.121536544 +0000 UTC m=+3729.885046147" watchObservedRunningTime="2025-12-01 15:36:04.127612868 +0000 UTC m=+3729.891122491" Dec 01 15:36:08 crc kubenswrapper[4810]: I1201 15:36:08.636600 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fqnhb" Dec 01 15:36:08 crc kubenswrapper[4810]: I1201 15:36:08.637173 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fqnhb" Dec 01 15:36:08 crc kubenswrapper[4810]: I1201 15:36:08.680953 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fqnhb" Dec 01 15:36:09 crc kubenswrapper[4810]: I1201 15:36:09.189640 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fqnhb" Dec 01 15:36:09 crc kubenswrapper[4810]: I1201 15:36:09.246204 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fqnhb"] Dec 01 15:36:11 crc kubenswrapper[4810]: I1201 15:36:11.167964 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fqnhb" podUID="9df5eda5-4d25-48e2-8648-65d61297dd42" containerName="registry-server" containerID="cri-o://a51da69eff153c0d11231ccf5fe84eb79b7419b13c4e22137b9af12a0fdc0760" gracePeriod=2 Dec 01 15:36:11 crc kubenswrapper[4810]: I1201 15:36:11.547143 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fqnhb" Dec 01 15:36:11 crc kubenswrapper[4810]: I1201 15:36:11.681382 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9df5eda5-4d25-48e2-8648-65d61297dd42-catalog-content\") pod \"9df5eda5-4d25-48e2-8648-65d61297dd42\" (UID: \"9df5eda5-4d25-48e2-8648-65d61297dd42\") " Dec 01 15:36:11 crc kubenswrapper[4810]: I1201 15:36:11.681510 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9df5eda5-4d25-48e2-8648-65d61297dd42-utilities\") pod \"9df5eda5-4d25-48e2-8648-65d61297dd42\" (UID: \"9df5eda5-4d25-48e2-8648-65d61297dd42\") " Dec 01 15:36:11 crc kubenswrapper[4810]: I1201 15:36:11.681644 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmwmm\" (UniqueName: \"kubernetes.io/projected/9df5eda5-4d25-48e2-8648-65d61297dd42-kube-api-access-tmwmm\") pod \"9df5eda5-4d25-48e2-8648-65d61297dd42\" (UID: \"9df5eda5-4d25-48e2-8648-65d61297dd42\") " Dec 01 15:36:11 crc kubenswrapper[4810]: I1201 15:36:11.682276 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9df5eda5-4d25-48e2-8648-65d61297dd42-utilities" (OuterVolumeSpecName: "utilities") pod "9df5eda5-4d25-48e2-8648-65d61297dd42" (UID: "9df5eda5-4d25-48e2-8648-65d61297dd42"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:36:11 crc kubenswrapper[4810]: I1201 15:36:11.687852 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9df5eda5-4d25-48e2-8648-65d61297dd42-kube-api-access-tmwmm" (OuterVolumeSpecName: "kube-api-access-tmwmm") pod "9df5eda5-4d25-48e2-8648-65d61297dd42" (UID: "9df5eda5-4d25-48e2-8648-65d61297dd42"). InnerVolumeSpecName "kube-api-access-tmwmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:36:11 crc kubenswrapper[4810]: I1201 15:36:11.748680 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9df5eda5-4d25-48e2-8648-65d61297dd42-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9df5eda5-4d25-48e2-8648-65d61297dd42" (UID: "9df5eda5-4d25-48e2-8648-65d61297dd42"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:36:11 crc kubenswrapper[4810]: I1201 15:36:11.783887 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9df5eda5-4d25-48e2-8648-65d61297dd42-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:36:11 crc kubenswrapper[4810]: I1201 15:36:11.783942 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmwmm\" (UniqueName: \"kubernetes.io/projected/9df5eda5-4d25-48e2-8648-65d61297dd42-kube-api-access-tmwmm\") on node \"crc\" DevicePath \"\"" Dec 01 15:36:11 crc kubenswrapper[4810]: I1201 15:36:11.783953 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9df5eda5-4d25-48e2-8648-65d61297dd42-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:36:12 crc kubenswrapper[4810]: I1201 15:36:12.178084 4810 generic.go:334] "Generic (PLEG): container finished" podID="9df5eda5-4d25-48e2-8648-65d61297dd42" containerID="a51da69eff153c0d11231ccf5fe84eb79b7419b13c4e22137b9af12a0fdc0760" exitCode=0 Dec 01 15:36:12 crc kubenswrapper[4810]: I1201 15:36:12.178178 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fqnhb" Dec 01 15:36:12 crc kubenswrapper[4810]: I1201 15:36:12.178208 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqnhb" event={"ID":"9df5eda5-4d25-48e2-8648-65d61297dd42","Type":"ContainerDied","Data":"a51da69eff153c0d11231ccf5fe84eb79b7419b13c4e22137b9af12a0fdc0760"} Dec 01 15:36:12 crc kubenswrapper[4810]: I1201 15:36:12.179255 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fqnhb" event={"ID":"9df5eda5-4d25-48e2-8648-65d61297dd42","Type":"ContainerDied","Data":"b10c31a2e7a55f0dc5e2c1e326c241f8362e41695724e5d6bc446b4d32733239"} Dec 01 15:36:12 crc kubenswrapper[4810]: I1201 15:36:12.179291 4810 scope.go:117] "RemoveContainer" containerID="a51da69eff153c0d11231ccf5fe84eb79b7419b13c4e22137b9af12a0fdc0760" Dec 01 15:36:12 crc kubenswrapper[4810]: I1201 15:36:12.206203 4810 scope.go:117] "RemoveContainer" containerID="1eb97462996ffa94c5b99d5a801fe56e7300776b38ba3d7e99fa7985e8ff6dad" Dec 01 15:36:12 crc kubenswrapper[4810]: I1201 15:36:12.211310 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fqnhb"] Dec 01 15:36:12 crc kubenswrapper[4810]: I1201 15:36:12.220078 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fqnhb"] Dec 01 15:36:12 crc kubenswrapper[4810]: I1201 15:36:12.235786 4810 scope.go:117] "RemoveContainer" containerID="b4f4f96fcb8fdb2752cde01fbc615e67112b2d945933e484f5bca4779a414991" Dec 01 15:36:12 crc kubenswrapper[4810]: I1201 15:36:12.250094 4810 scope.go:117] "RemoveContainer" containerID="a51da69eff153c0d11231ccf5fe84eb79b7419b13c4e22137b9af12a0fdc0760" Dec 01 15:36:12 crc kubenswrapper[4810]: E1201 15:36:12.250528 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a51da69eff153c0d11231ccf5fe84eb79b7419b13c4e22137b9af12a0fdc0760\": container with ID starting with a51da69eff153c0d11231ccf5fe84eb79b7419b13c4e22137b9af12a0fdc0760 not found: ID does not exist" containerID="a51da69eff153c0d11231ccf5fe84eb79b7419b13c4e22137b9af12a0fdc0760" Dec 01 15:36:12 crc kubenswrapper[4810]: I1201 15:36:12.250581 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a51da69eff153c0d11231ccf5fe84eb79b7419b13c4e22137b9af12a0fdc0760"} err="failed to get container status \"a51da69eff153c0d11231ccf5fe84eb79b7419b13c4e22137b9af12a0fdc0760\": rpc error: code = NotFound desc = could not find container \"a51da69eff153c0d11231ccf5fe84eb79b7419b13c4e22137b9af12a0fdc0760\": container with ID starting with a51da69eff153c0d11231ccf5fe84eb79b7419b13c4e22137b9af12a0fdc0760 not found: ID does not exist" Dec 01 15:36:12 crc kubenswrapper[4810]: I1201 15:36:12.250612 4810 scope.go:117] "RemoveContainer" containerID="1eb97462996ffa94c5b99d5a801fe56e7300776b38ba3d7e99fa7985e8ff6dad" Dec 01 15:36:12 crc kubenswrapper[4810]: E1201 15:36:12.250985 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1eb97462996ffa94c5b99d5a801fe56e7300776b38ba3d7e99fa7985e8ff6dad\": container with ID starting with 1eb97462996ffa94c5b99d5a801fe56e7300776b38ba3d7e99fa7985e8ff6dad not found: ID does not exist" containerID="1eb97462996ffa94c5b99d5a801fe56e7300776b38ba3d7e99fa7985e8ff6dad" Dec 01 15:36:12 crc kubenswrapper[4810]: I1201 15:36:12.251016 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1eb97462996ffa94c5b99d5a801fe56e7300776b38ba3d7e99fa7985e8ff6dad"} err="failed to get container status \"1eb97462996ffa94c5b99d5a801fe56e7300776b38ba3d7e99fa7985e8ff6dad\": rpc error: code = NotFound desc = could not find container \"1eb97462996ffa94c5b99d5a801fe56e7300776b38ba3d7e99fa7985e8ff6dad\": container with ID starting with 1eb97462996ffa94c5b99d5a801fe56e7300776b38ba3d7e99fa7985e8ff6dad not found: ID does not exist" Dec 01 15:36:12 crc kubenswrapper[4810]: I1201 15:36:12.251036 4810 scope.go:117] "RemoveContainer" containerID="b4f4f96fcb8fdb2752cde01fbc615e67112b2d945933e484f5bca4779a414991" Dec 01 15:36:12 crc kubenswrapper[4810]: E1201 15:36:12.251396 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4f4f96fcb8fdb2752cde01fbc615e67112b2d945933e484f5bca4779a414991\": container with ID starting with b4f4f96fcb8fdb2752cde01fbc615e67112b2d945933e484f5bca4779a414991 not found: ID does not exist" containerID="b4f4f96fcb8fdb2752cde01fbc615e67112b2d945933e484f5bca4779a414991" Dec 01 15:36:12 crc kubenswrapper[4810]: I1201 15:36:12.251431 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4f4f96fcb8fdb2752cde01fbc615e67112b2d945933e484f5bca4779a414991"} err="failed to get container status \"b4f4f96fcb8fdb2752cde01fbc615e67112b2d945933e484f5bca4779a414991\": rpc error: code = NotFound desc = could not find container \"b4f4f96fcb8fdb2752cde01fbc615e67112b2d945933e484f5bca4779a414991\": container with ID starting with b4f4f96fcb8fdb2752cde01fbc615e67112b2d945933e484f5bca4779a414991 not found: ID does not exist" Dec 01 15:36:12 crc kubenswrapper[4810]: I1201 15:36:12.501658 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9df5eda5-4d25-48e2-8648-65d61297dd42" path="/var/lib/kubelet/pods/9df5eda5-4d25-48e2-8648-65d61297dd42/volumes" Dec 01 15:37:02 crc kubenswrapper[4810]: I1201 15:37:02.972891 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:37:02 crc kubenswrapper[4810]: I1201 15:37:02.973807 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:37:32 crc kubenswrapper[4810]: I1201 15:37:32.972441 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:37:32 crc kubenswrapper[4810]: I1201 15:37:32.972989 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:38:02 crc kubenswrapper[4810]: I1201 15:38:02.972074 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:38:02 crc kubenswrapper[4810]: I1201 15:38:02.973297 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:38:02 crc kubenswrapper[4810]: I1201 15:38:02.973395 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 15:38:02 crc kubenswrapper[4810]: I1201 15:38:02.974706 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 15:38:02 crc kubenswrapper[4810]: I1201 15:38:02.974812 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" gracePeriod=600 Dec 01 15:38:03 crc kubenswrapper[4810]: E1201 15:38:03.102075 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:38:03 crc kubenswrapper[4810]: I1201 15:38:03.232068 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" exitCode=0 Dec 01 15:38:03 crc kubenswrapper[4810]: I1201 15:38:03.232148 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096"} Dec 01 15:38:03 crc kubenswrapper[4810]: I1201 15:38:03.232494 4810 scope.go:117] "RemoveContainer" containerID="4173ab52e4f8b87a8612cb1c2625b99d0a9ec9652f422817094bb8e9a3f7c869" Dec 01 15:38:03 crc kubenswrapper[4810]: I1201 15:38:03.238983 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:38:03 crc kubenswrapper[4810]: E1201 15:38:03.239626 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:38:17 crc kubenswrapper[4810]: I1201 15:38:17.491637 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:38:17 crc kubenswrapper[4810]: E1201 15:38:17.492347 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:38:31 crc kubenswrapper[4810]: I1201 15:38:31.491506 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:38:31 crc kubenswrapper[4810]: E1201 15:38:31.492313 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:38:43 crc kubenswrapper[4810]: I1201 15:38:43.491036 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:38:43 crc kubenswrapper[4810]: E1201 15:38:43.491816 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:38:54 crc kubenswrapper[4810]: I1201 15:38:54.494783 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:38:54 crc kubenswrapper[4810]: E1201 15:38:54.495613 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:39:08 crc kubenswrapper[4810]: I1201 15:39:08.491495 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:39:08 crc kubenswrapper[4810]: E1201 15:39:08.492339 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:39:20 crc kubenswrapper[4810]: I1201 15:39:20.491241 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:39:20 crc kubenswrapper[4810]: E1201 15:39:20.492017 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:39:32 crc kubenswrapper[4810]: I1201 15:39:32.491411 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:39:32 crc kubenswrapper[4810]: E1201 15:39:32.492088 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:39:45 crc kubenswrapper[4810]: I1201 15:39:45.490435 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:39:45 crc kubenswrapper[4810]: E1201 15:39:45.491127 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:39:49 crc kubenswrapper[4810]: I1201 15:39:49.762281 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-44w9m"] Dec 01 15:39:49 crc kubenswrapper[4810]: E1201 15:39:49.763167 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9df5eda5-4d25-48e2-8648-65d61297dd42" containerName="extract-content" Dec 01 15:39:49 crc kubenswrapper[4810]: I1201 15:39:49.763180 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9df5eda5-4d25-48e2-8648-65d61297dd42" containerName="extract-content" Dec 01 15:39:49 crc kubenswrapper[4810]: E1201 15:39:49.763200 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9df5eda5-4d25-48e2-8648-65d61297dd42" containerName="extract-utilities" Dec 01 15:39:49 crc kubenswrapper[4810]: I1201 15:39:49.763208 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9df5eda5-4d25-48e2-8648-65d61297dd42" containerName="extract-utilities" Dec 01 15:39:49 crc kubenswrapper[4810]: E1201 15:39:49.763221 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9df5eda5-4d25-48e2-8648-65d61297dd42" containerName="registry-server" Dec 01 15:39:49 crc kubenswrapper[4810]: I1201 15:39:49.763227 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9df5eda5-4d25-48e2-8648-65d61297dd42" containerName="registry-server" Dec 01 15:39:49 crc kubenswrapper[4810]: I1201 15:39:49.763389 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9df5eda5-4d25-48e2-8648-65d61297dd42" containerName="registry-server" Dec 01 15:39:49 crc kubenswrapper[4810]: I1201 15:39:49.766161 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-44w9m" Dec 01 15:39:49 crc kubenswrapper[4810]: I1201 15:39:49.774133 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-44w9m"] Dec 01 15:39:49 crc kubenswrapper[4810]: I1201 15:39:49.843226 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5467\" (UniqueName: \"kubernetes.io/projected/d659701f-0fad-46fd-a451-d20c9feca724-kube-api-access-w5467\") pod \"redhat-marketplace-44w9m\" (UID: \"d659701f-0fad-46fd-a451-d20c9feca724\") " pod="openshift-marketplace/redhat-marketplace-44w9m" Dec 01 15:39:49 crc kubenswrapper[4810]: I1201 15:39:49.843285 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d659701f-0fad-46fd-a451-d20c9feca724-catalog-content\") pod \"redhat-marketplace-44w9m\" (UID: \"d659701f-0fad-46fd-a451-d20c9feca724\") " pod="openshift-marketplace/redhat-marketplace-44w9m" Dec 01 15:39:49 crc kubenswrapper[4810]: I1201 15:39:49.843312 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d659701f-0fad-46fd-a451-d20c9feca724-utilities\") pod \"redhat-marketplace-44w9m\" (UID: \"d659701f-0fad-46fd-a451-d20c9feca724\") " pod="openshift-marketplace/redhat-marketplace-44w9m" Dec 01 15:39:49 crc kubenswrapper[4810]: I1201 15:39:49.944226 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5467\" (UniqueName: \"kubernetes.io/projected/d659701f-0fad-46fd-a451-d20c9feca724-kube-api-access-w5467\") pod \"redhat-marketplace-44w9m\" (UID: \"d659701f-0fad-46fd-a451-d20c9feca724\") " pod="openshift-marketplace/redhat-marketplace-44w9m" Dec 01 15:39:49 crc kubenswrapper[4810]: I1201 15:39:49.944287 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d659701f-0fad-46fd-a451-d20c9feca724-catalog-content\") pod \"redhat-marketplace-44w9m\" (UID: \"d659701f-0fad-46fd-a451-d20c9feca724\") " pod="openshift-marketplace/redhat-marketplace-44w9m" Dec 01 15:39:49 crc kubenswrapper[4810]: I1201 15:39:49.944312 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d659701f-0fad-46fd-a451-d20c9feca724-utilities\") pod \"redhat-marketplace-44w9m\" (UID: \"d659701f-0fad-46fd-a451-d20c9feca724\") " pod="openshift-marketplace/redhat-marketplace-44w9m" Dec 01 15:39:49 crc kubenswrapper[4810]: I1201 15:39:49.944914 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d659701f-0fad-46fd-a451-d20c9feca724-catalog-content\") pod \"redhat-marketplace-44w9m\" (UID: \"d659701f-0fad-46fd-a451-d20c9feca724\") " pod="openshift-marketplace/redhat-marketplace-44w9m" Dec 01 15:39:49 crc kubenswrapper[4810]: I1201 15:39:49.944977 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d659701f-0fad-46fd-a451-d20c9feca724-utilities\") pod \"redhat-marketplace-44w9m\" (UID: \"d659701f-0fad-46fd-a451-d20c9feca724\") " pod="openshift-marketplace/redhat-marketplace-44w9m" Dec 01 15:39:49 crc kubenswrapper[4810]: I1201 15:39:49.963524 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5467\" (UniqueName: \"kubernetes.io/projected/d659701f-0fad-46fd-a451-d20c9feca724-kube-api-access-w5467\") pod \"redhat-marketplace-44w9m\" (UID: \"d659701f-0fad-46fd-a451-d20c9feca724\") " pod="openshift-marketplace/redhat-marketplace-44w9m" Dec 01 15:39:50 crc kubenswrapper[4810]: I1201 15:39:50.084839 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-44w9m" Dec 01 15:39:50 crc kubenswrapper[4810]: I1201 15:39:50.500226 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-44w9m"] Dec 01 15:39:51 crc kubenswrapper[4810]: I1201 15:39:51.019955 4810 generic.go:334] "Generic (PLEG): container finished" podID="d659701f-0fad-46fd-a451-d20c9feca724" containerID="14c1c48c7e724d5d5d3283aa1df361726d1e6fa17e2d968fdf3949465ea2b05d" exitCode=0 Dec 01 15:39:51 crc kubenswrapper[4810]: I1201 15:39:51.020002 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44w9m" event={"ID":"d659701f-0fad-46fd-a451-d20c9feca724","Type":"ContainerDied","Data":"14c1c48c7e724d5d5d3283aa1df361726d1e6fa17e2d968fdf3949465ea2b05d"} Dec 01 15:39:51 crc kubenswrapper[4810]: I1201 15:39:51.020037 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44w9m" event={"ID":"d659701f-0fad-46fd-a451-d20c9feca724","Type":"ContainerStarted","Data":"bef3e8dd23c968b3a27ee5412266c453818765e6991211f62a208b32c6f21927"} Dec 01 15:39:51 crc kubenswrapper[4810]: I1201 15:39:51.021806 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 15:39:53 crc kubenswrapper[4810]: I1201 15:39:53.047108 4810 generic.go:334] "Generic (PLEG): container finished" podID="d659701f-0fad-46fd-a451-d20c9feca724" containerID="33070261b5a551f7914124f699f021785c2dd0416a9a5607c6300be7b7808ebb" exitCode=0 Dec 01 15:39:53 crc kubenswrapper[4810]: I1201 15:39:53.047158 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44w9m" event={"ID":"d659701f-0fad-46fd-a451-d20c9feca724","Type":"ContainerDied","Data":"33070261b5a551f7914124f699f021785c2dd0416a9a5607c6300be7b7808ebb"} Dec 01 15:39:54 crc kubenswrapper[4810]: I1201 15:39:54.056282 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44w9m" event={"ID":"d659701f-0fad-46fd-a451-d20c9feca724","Type":"ContainerStarted","Data":"6e786369dd58ec41ef501868f2c78eaf58da538fcd4376c54eddc572d95c9c1c"} Dec 01 15:39:54 crc kubenswrapper[4810]: I1201 15:39:54.073519 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-44w9m" podStartSLOduration=2.376058974 podStartE2EDuration="5.073500733s" podCreationTimestamp="2025-12-01 15:39:49 +0000 UTC" firstStartedPulling="2025-12-01 15:39:51.021557568 +0000 UTC m=+3956.785067171" lastFinishedPulling="2025-12-01 15:39:53.718999327 +0000 UTC m=+3959.482508930" observedRunningTime="2025-12-01 15:39:54.071820818 +0000 UTC m=+3959.835330421" watchObservedRunningTime="2025-12-01 15:39:54.073500733 +0000 UTC m=+3959.837010336" Dec 01 15:39:58 crc kubenswrapper[4810]: I1201 15:39:58.491867 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:39:58 crc kubenswrapper[4810]: E1201 15:39:58.492700 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:40:00 crc kubenswrapper[4810]: I1201 15:40:00.085683 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-44w9m" Dec 01 15:40:00 crc kubenswrapper[4810]: I1201 15:40:00.086124 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-44w9m" Dec 01 15:40:00 crc kubenswrapper[4810]: I1201 15:40:00.142449 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-44w9m" Dec 01 15:40:00 crc kubenswrapper[4810]: I1201 15:40:00.185890 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-44w9m" Dec 01 15:40:00 crc kubenswrapper[4810]: I1201 15:40:00.379988 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-44w9m"] Dec 01 15:40:02 crc kubenswrapper[4810]: I1201 15:40:02.118922 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-44w9m" podUID="d659701f-0fad-46fd-a451-d20c9feca724" containerName="registry-server" containerID="cri-o://6e786369dd58ec41ef501868f2c78eaf58da538fcd4376c54eddc572d95c9c1c" gracePeriod=2 Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.039688 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-44w9m" Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.126962 4810 generic.go:334] "Generic (PLEG): container finished" podID="d659701f-0fad-46fd-a451-d20c9feca724" containerID="6e786369dd58ec41ef501868f2c78eaf58da538fcd4376c54eddc572d95c9c1c" exitCode=0 Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.127010 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-44w9m" Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.127012 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44w9m" event={"ID":"d659701f-0fad-46fd-a451-d20c9feca724","Type":"ContainerDied","Data":"6e786369dd58ec41ef501868f2c78eaf58da538fcd4376c54eddc572d95c9c1c"} Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.127151 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44w9m" event={"ID":"d659701f-0fad-46fd-a451-d20c9feca724","Type":"ContainerDied","Data":"bef3e8dd23c968b3a27ee5412266c453818765e6991211f62a208b32c6f21927"} Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.127181 4810 scope.go:117] "RemoveContainer" containerID="6e786369dd58ec41ef501868f2c78eaf58da538fcd4376c54eddc572d95c9c1c" Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.129802 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d659701f-0fad-46fd-a451-d20c9feca724-catalog-content\") pod \"d659701f-0fad-46fd-a451-d20c9feca724\" (UID: \"d659701f-0fad-46fd-a451-d20c9feca724\") " Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.130002 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5467\" (UniqueName: \"kubernetes.io/projected/d659701f-0fad-46fd-a451-d20c9feca724-kube-api-access-w5467\") pod \"d659701f-0fad-46fd-a451-d20c9feca724\" (UID: \"d659701f-0fad-46fd-a451-d20c9feca724\") " Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.130193 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d659701f-0fad-46fd-a451-d20c9feca724-utilities\") pod \"d659701f-0fad-46fd-a451-d20c9feca724\" (UID: \"d659701f-0fad-46fd-a451-d20c9feca724\") " Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.131346 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d659701f-0fad-46fd-a451-d20c9feca724-utilities" (OuterVolumeSpecName: "utilities") pod "d659701f-0fad-46fd-a451-d20c9feca724" (UID: "d659701f-0fad-46fd-a451-d20c9feca724"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.152548 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d659701f-0fad-46fd-a451-d20c9feca724-kube-api-access-w5467" (OuterVolumeSpecName: "kube-api-access-w5467") pod "d659701f-0fad-46fd-a451-d20c9feca724" (UID: "d659701f-0fad-46fd-a451-d20c9feca724"). InnerVolumeSpecName "kube-api-access-w5467". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.155251 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d659701f-0fad-46fd-a451-d20c9feca724-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d659701f-0fad-46fd-a451-d20c9feca724" (UID: "d659701f-0fad-46fd-a451-d20c9feca724"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.165375 4810 scope.go:117] "RemoveContainer" containerID="33070261b5a551f7914124f699f021785c2dd0416a9a5607c6300be7b7808ebb" Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.187125 4810 scope.go:117] "RemoveContainer" containerID="14c1c48c7e724d5d5d3283aa1df361726d1e6fa17e2d968fdf3949465ea2b05d" Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.213100 4810 scope.go:117] "RemoveContainer" containerID="6e786369dd58ec41ef501868f2c78eaf58da538fcd4376c54eddc572d95c9c1c" Dec 01 15:40:03 crc kubenswrapper[4810]: E1201 15:40:03.213534 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e786369dd58ec41ef501868f2c78eaf58da538fcd4376c54eddc572d95c9c1c\": container with ID starting with 6e786369dd58ec41ef501868f2c78eaf58da538fcd4376c54eddc572d95c9c1c not found: ID does not exist" containerID="6e786369dd58ec41ef501868f2c78eaf58da538fcd4376c54eddc572d95c9c1c" Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.213592 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e786369dd58ec41ef501868f2c78eaf58da538fcd4376c54eddc572d95c9c1c"} err="failed to get container status \"6e786369dd58ec41ef501868f2c78eaf58da538fcd4376c54eddc572d95c9c1c\": rpc error: code = NotFound desc = could not find container \"6e786369dd58ec41ef501868f2c78eaf58da538fcd4376c54eddc572d95c9c1c\": container with ID starting with 6e786369dd58ec41ef501868f2c78eaf58da538fcd4376c54eddc572d95c9c1c not found: ID does not exist" Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.213625 4810 scope.go:117] "RemoveContainer" containerID="33070261b5a551f7914124f699f021785c2dd0416a9a5607c6300be7b7808ebb" Dec 01 15:40:03 crc kubenswrapper[4810]: E1201 15:40:03.213911 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33070261b5a551f7914124f699f021785c2dd0416a9a5607c6300be7b7808ebb\": container with ID starting with 33070261b5a551f7914124f699f021785c2dd0416a9a5607c6300be7b7808ebb not found: ID does not exist" containerID="33070261b5a551f7914124f699f021785c2dd0416a9a5607c6300be7b7808ebb" Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.213954 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33070261b5a551f7914124f699f021785c2dd0416a9a5607c6300be7b7808ebb"} err="failed to get container status \"33070261b5a551f7914124f699f021785c2dd0416a9a5607c6300be7b7808ebb\": rpc error: code = NotFound desc = could not find container \"33070261b5a551f7914124f699f021785c2dd0416a9a5607c6300be7b7808ebb\": container with ID starting with 33070261b5a551f7914124f699f021785c2dd0416a9a5607c6300be7b7808ebb not found: ID does not exist" Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.213980 4810 scope.go:117] "RemoveContainer" containerID="14c1c48c7e724d5d5d3283aa1df361726d1e6fa17e2d968fdf3949465ea2b05d" Dec 01 15:40:03 crc kubenswrapper[4810]: E1201 15:40:03.214207 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14c1c48c7e724d5d5d3283aa1df361726d1e6fa17e2d968fdf3949465ea2b05d\": container with ID starting with 14c1c48c7e724d5d5d3283aa1df361726d1e6fa17e2d968fdf3949465ea2b05d not found: ID does not exist" containerID="14c1c48c7e724d5d5d3283aa1df361726d1e6fa17e2d968fdf3949465ea2b05d" Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.214248 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14c1c48c7e724d5d5d3283aa1df361726d1e6fa17e2d968fdf3949465ea2b05d"} err="failed to get container status \"14c1c48c7e724d5d5d3283aa1df361726d1e6fa17e2d968fdf3949465ea2b05d\": rpc error: code = NotFound desc = could not find container \"14c1c48c7e724d5d5d3283aa1df361726d1e6fa17e2d968fdf3949465ea2b05d\": container with ID starting with 14c1c48c7e724d5d5d3283aa1df361726d1e6fa17e2d968fdf3949465ea2b05d not found: ID does not exist" Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.232389 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d659701f-0fad-46fd-a451-d20c9feca724-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.232437 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5467\" (UniqueName: \"kubernetes.io/projected/d659701f-0fad-46fd-a451-d20c9feca724-kube-api-access-w5467\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.232450 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d659701f-0fad-46fd-a451-d20c9feca724-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.457143 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-44w9m"] Dec 01 15:40:03 crc kubenswrapper[4810]: I1201 15:40:03.462225 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-44w9m"] Dec 01 15:40:04 crc kubenswrapper[4810]: I1201 15:40:04.503872 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d659701f-0fad-46fd-a451-d20c9feca724" path="/var/lib/kubelet/pods/d659701f-0fad-46fd-a451-d20c9feca724/volumes" Dec 01 15:40:12 crc kubenswrapper[4810]: I1201 15:40:12.491690 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:40:12 crc kubenswrapper[4810]: E1201 15:40:12.492622 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:40:23 crc kubenswrapper[4810]: I1201 15:40:23.491107 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:40:23 crc kubenswrapper[4810]: E1201 15:40:23.491857 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:40:38 crc kubenswrapper[4810]: I1201 15:40:38.491312 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:40:38 crc kubenswrapper[4810]: E1201 15:40:38.492107 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:40:51 crc kubenswrapper[4810]: I1201 15:40:51.491792 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:40:51 crc kubenswrapper[4810]: E1201 15:40:51.492793 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:41:03 crc kubenswrapper[4810]: I1201 15:41:03.492338 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:41:03 crc kubenswrapper[4810]: E1201 15:41:03.493142 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:41:15 crc kubenswrapper[4810]: I1201 15:41:15.491507 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:41:15 crc kubenswrapper[4810]: E1201 15:41:15.492220 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:41:28 crc kubenswrapper[4810]: I1201 15:41:28.491599 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:41:28 crc kubenswrapper[4810]: E1201 15:41:28.494357 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:41:41 crc kubenswrapper[4810]: I1201 15:41:41.490724 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:41:41 crc kubenswrapper[4810]: E1201 15:41:41.491419 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:41:53 crc kubenswrapper[4810]: I1201 15:41:53.491497 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:41:53 crc kubenswrapper[4810]: E1201 15:41:53.492272 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:42:06 crc kubenswrapper[4810]: I1201 15:42:06.491231 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:42:06 crc kubenswrapper[4810]: E1201 15:42:06.492455 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:42:18 crc kubenswrapper[4810]: I1201 15:42:18.148951 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-n9cjl"] Dec 01 15:42:18 crc kubenswrapper[4810]: E1201 15:42:18.149896 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d659701f-0fad-46fd-a451-d20c9feca724" containerName="extract-utilities" Dec 01 15:42:18 crc kubenswrapper[4810]: I1201 15:42:18.149993 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d659701f-0fad-46fd-a451-d20c9feca724" containerName="extract-utilities" Dec 01 15:42:18 crc kubenswrapper[4810]: E1201 15:42:18.150039 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d659701f-0fad-46fd-a451-d20c9feca724" containerName="registry-server" Dec 01 15:42:18 crc kubenswrapper[4810]: I1201 15:42:18.150049 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d659701f-0fad-46fd-a451-d20c9feca724" containerName="registry-server" Dec 01 15:42:18 crc kubenswrapper[4810]: E1201 15:42:18.150092 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d659701f-0fad-46fd-a451-d20c9feca724" containerName="extract-content" Dec 01 15:42:18 crc kubenswrapper[4810]: I1201 15:42:18.150103 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d659701f-0fad-46fd-a451-d20c9feca724" containerName="extract-content" Dec 01 15:42:18 crc kubenswrapper[4810]: I1201 15:42:18.150297 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d659701f-0fad-46fd-a451-d20c9feca724" containerName="registry-server" Dec 01 15:42:18 crc kubenswrapper[4810]: I1201 15:42:18.152822 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n9cjl" Dec 01 15:42:18 crc kubenswrapper[4810]: I1201 15:42:18.158836 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n9cjl"] Dec 01 15:42:18 crc kubenswrapper[4810]: I1201 15:42:18.334511 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hqwb\" (UniqueName: \"kubernetes.io/projected/0720be15-ff6d-42fb-82fc-cda019a726db-kube-api-access-8hqwb\") pod \"certified-operators-n9cjl\" (UID: \"0720be15-ff6d-42fb-82fc-cda019a726db\") " pod="openshift-marketplace/certified-operators-n9cjl" Dec 01 15:42:18 crc kubenswrapper[4810]: I1201 15:42:18.334608 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0720be15-ff6d-42fb-82fc-cda019a726db-utilities\") pod \"certified-operators-n9cjl\" (UID: \"0720be15-ff6d-42fb-82fc-cda019a726db\") " pod="openshift-marketplace/certified-operators-n9cjl" Dec 01 15:42:18 crc kubenswrapper[4810]: I1201 15:42:18.334646 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0720be15-ff6d-42fb-82fc-cda019a726db-catalog-content\") pod \"certified-operators-n9cjl\" (UID: \"0720be15-ff6d-42fb-82fc-cda019a726db\") " pod="openshift-marketplace/certified-operators-n9cjl" Dec 01 15:42:18 crc kubenswrapper[4810]: I1201 15:42:18.436244 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hqwb\" (UniqueName: \"kubernetes.io/projected/0720be15-ff6d-42fb-82fc-cda019a726db-kube-api-access-8hqwb\") pod \"certified-operators-n9cjl\" (UID: \"0720be15-ff6d-42fb-82fc-cda019a726db\") " pod="openshift-marketplace/certified-operators-n9cjl" Dec 01 15:42:18 crc kubenswrapper[4810]: I1201 15:42:18.436327 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0720be15-ff6d-42fb-82fc-cda019a726db-utilities\") pod \"certified-operators-n9cjl\" (UID: \"0720be15-ff6d-42fb-82fc-cda019a726db\") " pod="openshift-marketplace/certified-operators-n9cjl" Dec 01 15:42:18 crc kubenswrapper[4810]: I1201 15:42:18.436351 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0720be15-ff6d-42fb-82fc-cda019a726db-catalog-content\") pod \"certified-operators-n9cjl\" (UID: \"0720be15-ff6d-42fb-82fc-cda019a726db\") " pod="openshift-marketplace/certified-operators-n9cjl" Dec 01 15:42:18 crc kubenswrapper[4810]: I1201 15:42:18.436878 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0720be15-ff6d-42fb-82fc-cda019a726db-catalog-content\") pod \"certified-operators-n9cjl\" (UID: \"0720be15-ff6d-42fb-82fc-cda019a726db\") " pod="openshift-marketplace/certified-operators-n9cjl" Dec 01 15:42:18 crc kubenswrapper[4810]: I1201 15:42:18.437195 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0720be15-ff6d-42fb-82fc-cda019a726db-utilities\") pod \"certified-operators-n9cjl\" (UID: \"0720be15-ff6d-42fb-82fc-cda019a726db\") " pod="openshift-marketplace/certified-operators-n9cjl" Dec 01 15:42:18 crc kubenswrapper[4810]: I1201 15:42:18.458513 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hqwb\" (UniqueName: \"kubernetes.io/projected/0720be15-ff6d-42fb-82fc-cda019a726db-kube-api-access-8hqwb\") pod \"certified-operators-n9cjl\" (UID: \"0720be15-ff6d-42fb-82fc-cda019a726db\") " pod="openshift-marketplace/certified-operators-n9cjl" Dec 01 15:42:18 crc kubenswrapper[4810]: I1201 15:42:18.475040 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n9cjl" Dec 01 15:42:18 crc kubenswrapper[4810]: I1201 15:42:18.908408 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-n9cjl"] Dec 01 15:42:19 crc kubenswrapper[4810]: I1201 15:42:19.182074 4810 generic.go:334] "Generic (PLEG): container finished" podID="0720be15-ff6d-42fb-82fc-cda019a726db" containerID="05ce29b38177589a975fb27859e6ac5d9eb4283af238e207074e2193cff0494b" exitCode=0 Dec 01 15:42:19 crc kubenswrapper[4810]: I1201 15:42:19.182134 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n9cjl" event={"ID":"0720be15-ff6d-42fb-82fc-cda019a726db","Type":"ContainerDied","Data":"05ce29b38177589a975fb27859e6ac5d9eb4283af238e207074e2193cff0494b"} Dec 01 15:42:19 crc kubenswrapper[4810]: I1201 15:42:19.182159 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n9cjl" event={"ID":"0720be15-ff6d-42fb-82fc-cda019a726db","Type":"ContainerStarted","Data":"7ed1b2f2d5e969e3aa8e7fad52842188f54ee483303e22160378469799cb8d28"} Dec 01 15:42:19 crc kubenswrapper[4810]: I1201 15:42:19.490985 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:42:19 crc kubenswrapper[4810]: E1201 15:42:19.491526 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:42:21 crc kubenswrapper[4810]: I1201 15:42:21.197439 4810 generic.go:334] "Generic (PLEG): container finished" podID="0720be15-ff6d-42fb-82fc-cda019a726db" containerID="bd4ff156f6dc9a428a65baf6ce12eae712ce9a5de7593e69175ed3e068e63971" exitCode=0 Dec 01 15:42:21 crc kubenswrapper[4810]: I1201 15:42:21.197569 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n9cjl" event={"ID":"0720be15-ff6d-42fb-82fc-cda019a726db","Type":"ContainerDied","Data":"bd4ff156f6dc9a428a65baf6ce12eae712ce9a5de7593e69175ed3e068e63971"} Dec 01 15:42:22 crc kubenswrapper[4810]: I1201 15:42:22.208014 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n9cjl" event={"ID":"0720be15-ff6d-42fb-82fc-cda019a726db","Type":"ContainerStarted","Data":"de892feb0afee48508caa392762d1f7d6d94f3b6364d4c87b89cc8832564c429"} Dec 01 15:42:28 crc kubenswrapper[4810]: I1201 15:42:28.476286 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-n9cjl" Dec 01 15:42:28 crc kubenswrapper[4810]: I1201 15:42:28.476914 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-n9cjl" Dec 01 15:42:28 crc kubenswrapper[4810]: I1201 15:42:28.533919 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-n9cjl" Dec 01 15:42:28 crc kubenswrapper[4810]: I1201 15:42:28.553951 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-n9cjl" podStartSLOduration=8.118948017 podStartE2EDuration="10.553924764s" podCreationTimestamp="2025-12-01 15:42:18 +0000 UTC" firstStartedPulling="2025-12-01 15:42:19.18389093 +0000 UTC m=+4104.947400533" lastFinishedPulling="2025-12-01 15:42:21.618867677 +0000 UTC m=+4107.382377280" observedRunningTime="2025-12-01 15:42:22.225285761 +0000 UTC m=+4107.988795374" watchObservedRunningTime="2025-12-01 15:42:28.553924764 +0000 UTC m=+4114.317434377" Dec 01 15:42:29 crc kubenswrapper[4810]: I1201 15:42:29.334522 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-n9cjl" Dec 01 15:42:29 crc kubenswrapper[4810]: I1201 15:42:29.385837 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n9cjl"] Dec 01 15:42:31 crc kubenswrapper[4810]: I1201 15:42:31.268176 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-n9cjl" podUID="0720be15-ff6d-42fb-82fc-cda019a726db" containerName="registry-server" containerID="cri-o://de892feb0afee48508caa392762d1f7d6d94f3b6364d4c87b89cc8832564c429" gracePeriod=2 Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.120719 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n9cjl" Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.228506 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hqwb\" (UniqueName: \"kubernetes.io/projected/0720be15-ff6d-42fb-82fc-cda019a726db-kube-api-access-8hqwb\") pod \"0720be15-ff6d-42fb-82fc-cda019a726db\" (UID: \"0720be15-ff6d-42fb-82fc-cda019a726db\") " Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.228571 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0720be15-ff6d-42fb-82fc-cda019a726db-utilities\") pod \"0720be15-ff6d-42fb-82fc-cda019a726db\" (UID: \"0720be15-ff6d-42fb-82fc-cda019a726db\") " Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.228680 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0720be15-ff6d-42fb-82fc-cda019a726db-catalog-content\") pod \"0720be15-ff6d-42fb-82fc-cda019a726db\" (UID: \"0720be15-ff6d-42fb-82fc-cda019a726db\") " Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.230074 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0720be15-ff6d-42fb-82fc-cda019a726db-utilities" (OuterVolumeSpecName: "utilities") pod "0720be15-ff6d-42fb-82fc-cda019a726db" (UID: "0720be15-ff6d-42fb-82fc-cda019a726db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.237760 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0720be15-ff6d-42fb-82fc-cda019a726db-kube-api-access-8hqwb" (OuterVolumeSpecName: "kube-api-access-8hqwb") pod "0720be15-ff6d-42fb-82fc-cda019a726db" (UID: "0720be15-ff6d-42fb-82fc-cda019a726db"). InnerVolumeSpecName "kube-api-access-8hqwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.275128 4810 generic.go:334] "Generic (PLEG): container finished" podID="0720be15-ff6d-42fb-82fc-cda019a726db" containerID="de892feb0afee48508caa392762d1f7d6d94f3b6364d4c87b89cc8832564c429" exitCode=0 Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.275167 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n9cjl" event={"ID":"0720be15-ff6d-42fb-82fc-cda019a726db","Type":"ContainerDied","Data":"de892feb0afee48508caa392762d1f7d6d94f3b6364d4c87b89cc8832564c429"} Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.275194 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-n9cjl" event={"ID":"0720be15-ff6d-42fb-82fc-cda019a726db","Type":"ContainerDied","Data":"7ed1b2f2d5e969e3aa8e7fad52842188f54ee483303e22160378469799cb8d28"} Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.275211 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-n9cjl" Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.275217 4810 scope.go:117] "RemoveContainer" containerID="de892feb0afee48508caa392762d1f7d6d94f3b6364d4c87b89cc8832564c429" Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.279596 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0720be15-ff6d-42fb-82fc-cda019a726db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0720be15-ff6d-42fb-82fc-cda019a726db" (UID: "0720be15-ff6d-42fb-82fc-cda019a726db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.300143 4810 scope.go:117] "RemoveContainer" containerID="bd4ff156f6dc9a428a65baf6ce12eae712ce9a5de7593e69175ed3e068e63971" Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.322036 4810 scope.go:117] "RemoveContainer" containerID="05ce29b38177589a975fb27859e6ac5d9eb4283af238e207074e2193cff0494b" Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.330075 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0720be15-ff6d-42fb-82fc-cda019a726db-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.330119 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hqwb\" (UniqueName: \"kubernetes.io/projected/0720be15-ff6d-42fb-82fc-cda019a726db-kube-api-access-8hqwb\") on node \"crc\" DevicePath \"\"" Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.330131 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0720be15-ff6d-42fb-82fc-cda019a726db-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.347102 4810 scope.go:117] "RemoveContainer" containerID="de892feb0afee48508caa392762d1f7d6d94f3b6364d4c87b89cc8832564c429" Dec 01 15:42:32 crc kubenswrapper[4810]: E1201 15:42:32.347654 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de892feb0afee48508caa392762d1f7d6d94f3b6364d4c87b89cc8832564c429\": container with ID starting with de892feb0afee48508caa392762d1f7d6d94f3b6364d4c87b89cc8832564c429 not found: ID does not exist" containerID="de892feb0afee48508caa392762d1f7d6d94f3b6364d4c87b89cc8832564c429" Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.347698 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de892feb0afee48508caa392762d1f7d6d94f3b6364d4c87b89cc8832564c429"} err="failed to get container status \"de892feb0afee48508caa392762d1f7d6d94f3b6364d4c87b89cc8832564c429\": rpc error: code = NotFound desc = could not find container \"de892feb0afee48508caa392762d1f7d6d94f3b6364d4c87b89cc8832564c429\": container with ID starting with de892feb0afee48508caa392762d1f7d6d94f3b6364d4c87b89cc8832564c429 not found: ID does not exist" Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.347727 4810 scope.go:117] "RemoveContainer" containerID="bd4ff156f6dc9a428a65baf6ce12eae712ce9a5de7593e69175ed3e068e63971" Dec 01 15:42:32 crc kubenswrapper[4810]: E1201 15:42:32.348084 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd4ff156f6dc9a428a65baf6ce12eae712ce9a5de7593e69175ed3e068e63971\": container with ID starting with bd4ff156f6dc9a428a65baf6ce12eae712ce9a5de7593e69175ed3e068e63971 not found: ID does not exist" containerID="bd4ff156f6dc9a428a65baf6ce12eae712ce9a5de7593e69175ed3e068e63971" Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.348112 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd4ff156f6dc9a428a65baf6ce12eae712ce9a5de7593e69175ed3e068e63971"} err="failed to get container status \"bd4ff156f6dc9a428a65baf6ce12eae712ce9a5de7593e69175ed3e068e63971\": rpc error: code = NotFound desc = could not find container \"bd4ff156f6dc9a428a65baf6ce12eae712ce9a5de7593e69175ed3e068e63971\": container with ID starting with bd4ff156f6dc9a428a65baf6ce12eae712ce9a5de7593e69175ed3e068e63971 not found: ID does not exist" Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.348128 4810 scope.go:117] "RemoveContainer" containerID="05ce29b38177589a975fb27859e6ac5d9eb4283af238e207074e2193cff0494b" Dec 01 15:42:32 crc kubenswrapper[4810]: E1201 15:42:32.348422 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05ce29b38177589a975fb27859e6ac5d9eb4283af238e207074e2193cff0494b\": container with ID starting with 05ce29b38177589a975fb27859e6ac5d9eb4283af238e207074e2193cff0494b not found: ID does not exist" containerID="05ce29b38177589a975fb27859e6ac5d9eb4283af238e207074e2193cff0494b" Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.348511 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05ce29b38177589a975fb27859e6ac5d9eb4283af238e207074e2193cff0494b"} err="failed to get container status \"05ce29b38177589a975fb27859e6ac5d9eb4283af238e207074e2193cff0494b\": rpc error: code = NotFound desc = could not find container \"05ce29b38177589a975fb27859e6ac5d9eb4283af238e207074e2193cff0494b\": container with ID starting with 05ce29b38177589a975fb27859e6ac5d9eb4283af238e207074e2193cff0494b not found: ID does not exist" Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.601531 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-n9cjl"] Dec 01 15:42:32 crc kubenswrapper[4810]: I1201 15:42:32.608350 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-n9cjl"] Dec 01 15:42:34 crc kubenswrapper[4810]: I1201 15:42:34.497626 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:42:34 crc kubenswrapper[4810]: E1201 15:42:34.497847 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:42:34 crc kubenswrapper[4810]: I1201 15:42:34.500250 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0720be15-ff6d-42fb-82fc-cda019a726db" path="/var/lib/kubelet/pods/0720be15-ff6d-42fb-82fc-cda019a726db/volumes" Dec 01 15:42:47 crc kubenswrapper[4810]: I1201 15:42:47.490615 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:42:47 crc kubenswrapper[4810]: E1201 15:42:47.491313 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:43:01 crc kubenswrapper[4810]: I1201 15:43:01.611548 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:43:01 crc kubenswrapper[4810]: E1201 15:43:01.614798 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:43:16 crc kubenswrapper[4810]: I1201 15:43:16.490786 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:43:16 crc kubenswrapper[4810]: I1201 15:43:16.716537 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"adfbc88df5691584126b7883ecbdc4c9c6160917ae2024226dfaa689c64d123b"} Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.170270 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv"] Dec 01 15:45:00 crc kubenswrapper[4810]: E1201 15:45:00.171981 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0720be15-ff6d-42fb-82fc-cda019a726db" containerName="extract-utilities" Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.172010 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0720be15-ff6d-42fb-82fc-cda019a726db" containerName="extract-utilities" Dec 01 15:45:00 crc kubenswrapper[4810]: E1201 15:45:00.172035 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0720be15-ff6d-42fb-82fc-cda019a726db" containerName="extract-content" Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.172042 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0720be15-ff6d-42fb-82fc-cda019a726db" containerName="extract-content" Dec 01 15:45:00 crc kubenswrapper[4810]: E1201 15:45:00.172057 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0720be15-ff6d-42fb-82fc-cda019a726db" containerName="registry-server" Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.172064 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0720be15-ff6d-42fb-82fc-cda019a726db" containerName="registry-server" Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.172204 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0720be15-ff6d-42fb-82fc-cda019a726db" containerName="registry-server" Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.172731 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv" Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.175092 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.175221 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.188465 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv"] Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.347252 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7a3c1b8-3d1c-4884-b887-c13edcaacd53-secret-volume\") pod \"collect-profiles-29410065-knjcv\" (UID: \"a7a3c1b8-3d1c-4884-b887-c13edcaacd53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv" Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.347295 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm5xn\" (UniqueName: \"kubernetes.io/projected/a7a3c1b8-3d1c-4884-b887-c13edcaacd53-kube-api-access-fm5xn\") pod \"collect-profiles-29410065-knjcv\" (UID: \"a7a3c1b8-3d1c-4884-b887-c13edcaacd53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv" Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.347712 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7a3c1b8-3d1c-4884-b887-c13edcaacd53-config-volume\") pod \"collect-profiles-29410065-knjcv\" (UID: \"a7a3c1b8-3d1c-4884-b887-c13edcaacd53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv" Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.449102 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7a3c1b8-3d1c-4884-b887-c13edcaacd53-secret-volume\") pod \"collect-profiles-29410065-knjcv\" (UID: \"a7a3c1b8-3d1c-4884-b887-c13edcaacd53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv" Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.449147 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm5xn\" (UniqueName: \"kubernetes.io/projected/a7a3c1b8-3d1c-4884-b887-c13edcaacd53-kube-api-access-fm5xn\") pod \"collect-profiles-29410065-knjcv\" (UID: \"a7a3c1b8-3d1c-4884-b887-c13edcaacd53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv" Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.449200 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7a3c1b8-3d1c-4884-b887-c13edcaacd53-config-volume\") pod \"collect-profiles-29410065-knjcv\" (UID: \"a7a3c1b8-3d1c-4884-b887-c13edcaacd53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv" Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.449957 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7a3c1b8-3d1c-4884-b887-c13edcaacd53-config-volume\") pod \"collect-profiles-29410065-knjcv\" (UID: \"a7a3c1b8-3d1c-4884-b887-c13edcaacd53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv" Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.458912 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7a3c1b8-3d1c-4884-b887-c13edcaacd53-secret-volume\") pod \"collect-profiles-29410065-knjcv\" (UID: \"a7a3c1b8-3d1c-4884-b887-c13edcaacd53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv" Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.468976 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm5xn\" (UniqueName: \"kubernetes.io/projected/a7a3c1b8-3d1c-4884-b887-c13edcaacd53-kube-api-access-fm5xn\") pod \"collect-profiles-29410065-knjcv\" (UID: \"a7a3c1b8-3d1c-4884-b887-c13edcaacd53\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv" Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.493937 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv" Dec 01 15:45:00 crc kubenswrapper[4810]: I1201 15:45:00.932069 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv"] Dec 01 15:45:01 crc kubenswrapper[4810]: I1201 15:45:01.531898 4810 generic.go:334] "Generic (PLEG): container finished" podID="a7a3c1b8-3d1c-4884-b887-c13edcaacd53" containerID="d7e73dcfe83a0bbd79d836bdea642db498fc3588517c95a2e7be6b62f3b1bd91" exitCode=0 Dec 01 15:45:01 crc kubenswrapper[4810]: I1201 15:45:01.531993 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv" event={"ID":"a7a3c1b8-3d1c-4884-b887-c13edcaacd53","Type":"ContainerDied","Data":"d7e73dcfe83a0bbd79d836bdea642db498fc3588517c95a2e7be6b62f3b1bd91"} Dec 01 15:45:01 crc kubenswrapper[4810]: I1201 15:45:01.532267 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv" event={"ID":"a7a3c1b8-3d1c-4884-b887-c13edcaacd53","Type":"ContainerStarted","Data":"100e8ffd0410fac536d46bbb16651f199783654248c369c606c58cb33eccbea5"} Dec 01 15:45:02 crc kubenswrapper[4810]: I1201 15:45:02.787576 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv" Dec 01 15:45:02 crc kubenswrapper[4810]: I1201 15:45:02.880749 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7a3c1b8-3d1c-4884-b887-c13edcaacd53-secret-volume\") pod \"a7a3c1b8-3d1c-4884-b887-c13edcaacd53\" (UID: \"a7a3c1b8-3d1c-4884-b887-c13edcaacd53\") " Dec 01 15:45:02 crc kubenswrapper[4810]: I1201 15:45:02.880876 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fm5xn\" (UniqueName: \"kubernetes.io/projected/a7a3c1b8-3d1c-4884-b887-c13edcaacd53-kube-api-access-fm5xn\") pod \"a7a3c1b8-3d1c-4884-b887-c13edcaacd53\" (UID: \"a7a3c1b8-3d1c-4884-b887-c13edcaacd53\") " Dec 01 15:45:02 crc kubenswrapper[4810]: I1201 15:45:02.880933 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7a3c1b8-3d1c-4884-b887-c13edcaacd53-config-volume\") pod \"a7a3c1b8-3d1c-4884-b887-c13edcaacd53\" (UID: \"a7a3c1b8-3d1c-4884-b887-c13edcaacd53\") " Dec 01 15:45:02 crc kubenswrapper[4810]: I1201 15:45:02.882256 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7a3c1b8-3d1c-4884-b887-c13edcaacd53-config-volume" (OuterVolumeSpecName: "config-volume") pod "a7a3c1b8-3d1c-4884-b887-c13edcaacd53" (UID: "a7a3c1b8-3d1c-4884-b887-c13edcaacd53"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:45:02 crc kubenswrapper[4810]: I1201 15:45:02.887861 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7a3c1b8-3d1c-4884-b887-c13edcaacd53-kube-api-access-fm5xn" (OuterVolumeSpecName: "kube-api-access-fm5xn") pod "a7a3c1b8-3d1c-4884-b887-c13edcaacd53" (UID: "a7a3c1b8-3d1c-4884-b887-c13edcaacd53"). InnerVolumeSpecName "kube-api-access-fm5xn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:45:02 crc kubenswrapper[4810]: I1201 15:45:02.887892 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7a3c1b8-3d1c-4884-b887-c13edcaacd53-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a7a3c1b8-3d1c-4884-b887-c13edcaacd53" (UID: "a7a3c1b8-3d1c-4884-b887-c13edcaacd53"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 15:45:02 crc kubenswrapper[4810]: I1201 15:45:02.982775 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a7a3c1b8-3d1c-4884-b887-c13edcaacd53-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:02 crc kubenswrapper[4810]: I1201 15:45:02.982843 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fm5xn\" (UniqueName: \"kubernetes.io/projected/a7a3c1b8-3d1c-4884-b887-c13edcaacd53-kube-api-access-fm5xn\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:02 crc kubenswrapper[4810]: I1201 15:45:02.982862 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a7a3c1b8-3d1c-4884-b887-c13edcaacd53-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 15:45:03 crc kubenswrapper[4810]: I1201 15:45:03.548234 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv" event={"ID":"a7a3c1b8-3d1c-4884-b887-c13edcaacd53","Type":"ContainerDied","Data":"100e8ffd0410fac536d46bbb16651f199783654248c369c606c58cb33eccbea5"} Dec 01 15:45:03 crc kubenswrapper[4810]: I1201 15:45:03.548279 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="100e8ffd0410fac536d46bbb16651f199783654248c369c606c58cb33eccbea5" Dec 01 15:45:03 crc kubenswrapper[4810]: I1201 15:45:03.548302 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv" Dec 01 15:45:03 crc kubenswrapper[4810]: I1201 15:45:03.858556 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r"] Dec 01 15:45:03 crc kubenswrapper[4810]: I1201 15:45:03.866363 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410020-zvw8r"] Dec 01 15:45:04 crc kubenswrapper[4810]: I1201 15:45:04.500185 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec7ca8a4-04cb-4b40-be20-6f2d872f8b25" path="/var/lib/kubelet/pods/ec7ca8a4-04cb-4b40-be20-6f2d872f8b25/volumes" Dec 01 15:45:08 crc kubenswrapper[4810]: I1201 15:45:08.373201 4810 scope.go:117] "RemoveContainer" containerID="16c524e58e80b9b3ae10fd0e7efafe0e4568dc81b9a83ba564662a50e1693504" Dec 01 15:45:32 crc kubenswrapper[4810]: I1201 15:45:32.972106 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:45:32 crc kubenswrapper[4810]: I1201 15:45:32.973164 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:46:02 crc kubenswrapper[4810]: I1201 15:46:02.971684 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:46:02 crc kubenswrapper[4810]: I1201 15:46:02.972248 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:46:11 crc kubenswrapper[4810]: I1201 15:46:11.696057 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7hxnx"] Dec 01 15:46:11 crc kubenswrapper[4810]: E1201 15:46:11.696872 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7a3c1b8-3d1c-4884-b887-c13edcaacd53" containerName="collect-profiles" Dec 01 15:46:11 crc kubenswrapper[4810]: I1201 15:46:11.696884 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7a3c1b8-3d1c-4884-b887-c13edcaacd53" containerName="collect-profiles" Dec 01 15:46:11 crc kubenswrapper[4810]: I1201 15:46:11.697024 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7a3c1b8-3d1c-4884-b887-c13edcaacd53" containerName="collect-profiles" Dec 01 15:46:11 crc kubenswrapper[4810]: I1201 15:46:11.698111 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7hxnx" Dec 01 15:46:11 crc kubenswrapper[4810]: I1201 15:46:11.717122 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7hxnx"] Dec 01 15:46:11 crc kubenswrapper[4810]: I1201 15:46:11.753398 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7bb8807-1be6-4fc7-8eef-095c80d67cd6-catalog-content\") pod \"community-operators-7hxnx\" (UID: \"f7bb8807-1be6-4fc7-8eef-095c80d67cd6\") " pod="openshift-marketplace/community-operators-7hxnx" Dec 01 15:46:11 crc kubenswrapper[4810]: I1201 15:46:11.753569 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlnfb\" (UniqueName: \"kubernetes.io/projected/f7bb8807-1be6-4fc7-8eef-095c80d67cd6-kube-api-access-zlnfb\") pod \"community-operators-7hxnx\" (UID: \"f7bb8807-1be6-4fc7-8eef-095c80d67cd6\") " pod="openshift-marketplace/community-operators-7hxnx" Dec 01 15:46:11 crc kubenswrapper[4810]: I1201 15:46:11.753627 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7bb8807-1be6-4fc7-8eef-095c80d67cd6-utilities\") pod \"community-operators-7hxnx\" (UID: \"f7bb8807-1be6-4fc7-8eef-095c80d67cd6\") " pod="openshift-marketplace/community-operators-7hxnx" Dec 01 15:46:11 crc kubenswrapper[4810]: I1201 15:46:11.854708 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlnfb\" (UniqueName: \"kubernetes.io/projected/f7bb8807-1be6-4fc7-8eef-095c80d67cd6-kube-api-access-zlnfb\") pod \"community-operators-7hxnx\" (UID: \"f7bb8807-1be6-4fc7-8eef-095c80d67cd6\") " pod="openshift-marketplace/community-operators-7hxnx" Dec 01 15:46:11 crc kubenswrapper[4810]: I1201 15:46:11.854783 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7bb8807-1be6-4fc7-8eef-095c80d67cd6-utilities\") pod \"community-operators-7hxnx\" (UID: \"f7bb8807-1be6-4fc7-8eef-095c80d67cd6\") " pod="openshift-marketplace/community-operators-7hxnx" Dec 01 15:46:11 crc kubenswrapper[4810]: I1201 15:46:11.854833 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7bb8807-1be6-4fc7-8eef-095c80d67cd6-catalog-content\") pod \"community-operators-7hxnx\" (UID: \"f7bb8807-1be6-4fc7-8eef-095c80d67cd6\") " pod="openshift-marketplace/community-operators-7hxnx" Dec 01 15:46:11 crc kubenswrapper[4810]: I1201 15:46:11.855307 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7bb8807-1be6-4fc7-8eef-095c80d67cd6-catalog-content\") pod \"community-operators-7hxnx\" (UID: \"f7bb8807-1be6-4fc7-8eef-095c80d67cd6\") " pod="openshift-marketplace/community-operators-7hxnx" Dec 01 15:46:11 crc kubenswrapper[4810]: I1201 15:46:11.855410 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7bb8807-1be6-4fc7-8eef-095c80d67cd6-utilities\") pod \"community-operators-7hxnx\" (UID: \"f7bb8807-1be6-4fc7-8eef-095c80d67cd6\") " pod="openshift-marketplace/community-operators-7hxnx" Dec 01 15:46:11 crc kubenswrapper[4810]: I1201 15:46:11.873537 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlnfb\" (UniqueName: \"kubernetes.io/projected/f7bb8807-1be6-4fc7-8eef-095c80d67cd6-kube-api-access-zlnfb\") pod \"community-operators-7hxnx\" (UID: \"f7bb8807-1be6-4fc7-8eef-095c80d67cd6\") " pod="openshift-marketplace/community-operators-7hxnx" Dec 01 15:46:12 crc kubenswrapper[4810]: I1201 15:46:12.042915 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7hxnx" Dec 01 15:46:12 crc kubenswrapper[4810]: I1201 15:46:12.370139 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7hxnx"] Dec 01 15:46:13 crc kubenswrapper[4810]: I1201 15:46:13.105876 4810 generic.go:334] "Generic (PLEG): container finished" podID="f7bb8807-1be6-4fc7-8eef-095c80d67cd6" containerID="bfd28a94abbf0c734f31fe59939c624c92f1e15e402651999e993ba10adc7c01" exitCode=0 Dec 01 15:46:13 crc kubenswrapper[4810]: I1201 15:46:13.105939 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hxnx" event={"ID":"f7bb8807-1be6-4fc7-8eef-095c80d67cd6","Type":"ContainerDied","Data":"bfd28a94abbf0c734f31fe59939c624c92f1e15e402651999e993ba10adc7c01"} Dec 01 15:46:13 crc kubenswrapper[4810]: I1201 15:46:13.106207 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hxnx" event={"ID":"f7bb8807-1be6-4fc7-8eef-095c80d67cd6","Type":"ContainerStarted","Data":"8b1fa9a2c78c403aa25e6b1303df967ff3c7f0faa16c2444cae431e255b14b03"} Dec 01 15:46:13 crc kubenswrapper[4810]: I1201 15:46:13.109704 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 15:46:14 crc kubenswrapper[4810]: I1201 15:46:14.117776 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hxnx" event={"ID":"f7bb8807-1be6-4fc7-8eef-095c80d67cd6","Type":"ContainerStarted","Data":"a65862ef0c2f4ee327d589d9718c9446701d6645b97a5e3add90f2d71ca543d9"} Dec 01 15:46:15 crc kubenswrapper[4810]: I1201 15:46:15.127117 4810 generic.go:334] "Generic (PLEG): container finished" podID="f7bb8807-1be6-4fc7-8eef-095c80d67cd6" containerID="a65862ef0c2f4ee327d589d9718c9446701d6645b97a5e3add90f2d71ca543d9" exitCode=0 Dec 01 15:46:15 crc kubenswrapper[4810]: I1201 15:46:15.127221 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hxnx" event={"ID":"f7bb8807-1be6-4fc7-8eef-095c80d67cd6","Type":"ContainerDied","Data":"a65862ef0c2f4ee327d589d9718c9446701d6645b97a5e3add90f2d71ca543d9"} Dec 01 15:46:15 crc kubenswrapper[4810]: I1201 15:46:15.127470 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hxnx" event={"ID":"f7bb8807-1be6-4fc7-8eef-095c80d67cd6","Type":"ContainerStarted","Data":"28f2f9421d791918742e7e8c36db36a61dd45a1c90cc89b1a863b69a000636fd"} Dec 01 15:46:15 crc kubenswrapper[4810]: I1201 15:46:15.148710 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7hxnx" podStartSLOduration=2.614172592 podStartE2EDuration="4.148687386s" podCreationTimestamp="2025-12-01 15:46:11 +0000 UTC" firstStartedPulling="2025-12-01 15:46:13.109357851 +0000 UTC m=+4338.872867454" lastFinishedPulling="2025-12-01 15:46:14.643872645 +0000 UTC m=+4340.407382248" observedRunningTime="2025-12-01 15:46:15.144056531 +0000 UTC m=+4340.907566134" watchObservedRunningTime="2025-12-01 15:46:15.148687386 +0000 UTC m=+4340.912196989" Dec 01 15:46:22 crc kubenswrapper[4810]: I1201 15:46:22.043173 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7hxnx" Dec 01 15:46:22 crc kubenswrapper[4810]: I1201 15:46:22.043771 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7hxnx" Dec 01 15:46:22 crc kubenswrapper[4810]: I1201 15:46:22.391252 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7hxnx" Dec 01 15:46:22 crc kubenswrapper[4810]: I1201 15:46:22.436674 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7hxnx" Dec 01 15:46:22 crc kubenswrapper[4810]: I1201 15:46:22.626261 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7hxnx"] Dec 01 15:46:24 crc kubenswrapper[4810]: I1201 15:46:24.202428 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7hxnx" podUID="f7bb8807-1be6-4fc7-8eef-095c80d67cd6" containerName="registry-server" containerID="cri-o://28f2f9421d791918742e7e8c36db36a61dd45a1c90cc89b1a863b69a000636fd" gracePeriod=2 Dec 01 15:46:24 crc kubenswrapper[4810]: I1201 15:46:24.561522 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7hxnx" Dec 01 15:46:24 crc kubenswrapper[4810]: I1201 15:46:24.721597 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7bb8807-1be6-4fc7-8eef-095c80d67cd6-utilities\") pod \"f7bb8807-1be6-4fc7-8eef-095c80d67cd6\" (UID: \"f7bb8807-1be6-4fc7-8eef-095c80d67cd6\") " Dec 01 15:46:24 crc kubenswrapper[4810]: I1201 15:46:24.721673 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7bb8807-1be6-4fc7-8eef-095c80d67cd6-catalog-content\") pod \"f7bb8807-1be6-4fc7-8eef-095c80d67cd6\" (UID: \"f7bb8807-1be6-4fc7-8eef-095c80d67cd6\") " Dec 01 15:46:24 crc kubenswrapper[4810]: I1201 15:46:24.721785 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlnfb\" (UniqueName: \"kubernetes.io/projected/f7bb8807-1be6-4fc7-8eef-095c80d67cd6-kube-api-access-zlnfb\") pod \"f7bb8807-1be6-4fc7-8eef-095c80d67cd6\" (UID: \"f7bb8807-1be6-4fc7-8eef-095c80d67cd6\") " Dec 01 15:46:24 crc kubenswrapper[4810]: I1201 15:46:24.722653 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7bb8807-1be6-4fc7-8eef-095c80d67cd6-utilities" (OuterVolumeSpecName: "utilities") pod "f7bb8807-1be6-4fc7-8eef-095c80d67cd6" (UID: "f7bb8807-1be6-4fc7-8eef-095c80d67cd6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:46:24 crc kubenswrapper[4810]: I1201 15:46:24.728452 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7bb8807-1be6-4fc7-8eef-095c80d67cd6-kube-api-access-zlnfb" (OuterVolumeSpecName: "kube-api-access-zlnfb") pod "f7bb8807-1be6-4fc7-8eef-095c80d67cd6" (UID: "f7bb8807-1be6-4fc7-8eef-095c80d67cd6"). InnerVolumeSpecName "kube-api-access-zlnfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:46:24 crc kubenswrapper[4810]: I1201 15:46:24.770347 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7bb8807-1be6-4fc7-8eef-095c80d67cd6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7bb8807-1be6-4fc7-8eef-095c80d67cd6" (UID: "f7bb8807-1be6-4fc7-8eef-095c80d67cd6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:46:24 crc kubenswrapper[4810]: I1201 15:46:24.823845 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7bb8807-1be6-4fc7-8eef-095c80d67cd6-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:46:24 crc kubenswrapper[4810]: I1201 15:46:24.823904 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7bb8807-1be6-4fc7-8eef-095c80d67cd6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:46:24 crc kubenswrapper[4810]: I1201 15:46:24.823916 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlnfb\" (UniqueName: \"kubernetes.io/projected/f7bb8807-1be6-4fc7-8eef-095c80d67cd6-kube-api-access-zlnfb\") on node \"crc\" DevicePath \"\"" Dec 01 15:46:25 crc kubenswrapper[4810]: I1201 15:46:25.211204 4810 generic.go:334] "Generic (PLEG): container finished" podID="f7bb8807-1be6-4fc7-8eef-095c80d67cd6" containerID="28f2f9421d791918742e7e8c36db36a61dd45a1c90cc89b1a863b69a000636fd" exitCode=0 Dec 01 15:46:25 crc kubenswrapper[4810]: I1201 15:46:25.211239 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hxnx" event={"ID":"f7bb8807-1be6-4fc7-8eef-095c80d67cd6","Type":"ContainerDied","Data":"28f2f9421d791918742e7e8c36db36a61dd45a1c90cc89b1a863b69a000636fd"} Dec 01 15:46:25 crc kubenswrapper[4810]: I1201 15:46:25.211265 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7hxnx" event={"ID":"f7bb8807-1be6-4fc7-8eef-095c80d67cd6","Type":"ContainerDied","Data":"8b1fa9a2c78c403aa25e6b1303df967ff3c7f0faa16c2444cae431e255b14b03"} Dec 01 15:46:25 crc kubenswrapper[4810]: I1201 15:46:25.211282 4810 scope.go:117] "RemoveContainer" containerID="28f2f9421d791918742e7e8c36db36a61dd45a1c90cc89b1a863b69a000636fd" Dec 01 15:46:25 crc kubenswrapper[4810]: I1201 15:46:25.211331 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7hxnx" Dec 01 15:46:25 crc kubenswrapper[4810]: I1201 15:46:25.229270 4810 scope.go:117] "RemoveContainer" containerID="a65862ef0c2f4ee327d589d9718c9446701d6645b97a5e3add90f2d71ca543d9" Dec 01 15:46:25 crc kubenswrapper[4810]: I1201 15:46:25.248893 4810 scope.go:117] "RemoveContainer" containerID="bfd28a94abbf0c734f31fe59939c624c92f1e15e402651999e993ba10adc7c01" Dec 01 15:46:25 crc kubenswrapper[4810]: I1201 15:46:25.287881 4810 scope.go:117] "RemoveContainer" containerID="28f2f9421d791918742e7e8c36db36a61dd45a1c90cc89b1a863b69a000636fd" Dec 01 15:46:25 crc kubenswrapper[4810]: E1201 15:46:25.288240 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28f2f9421d791918742e7e8c36db36a61dd45a1c90cc89b1a863b69a000636fd\": container with ID starting with 28f2f9421d791918742e7e8c36db36a61dd45a1c90cc89b1a863b69a000636fd not found: ID does not exist" containerID="28f2f9421d791918742e7e8c36db36a61dd45a1c90cc89b1a863b69a000636fd" Dec 01 15:46:25 crc kubenswrapper[4810]: I1201 15:46:25.288275 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28f2f9421d791918742e7e8c36db36a61dd45a1c90cc89b1a863b69a000636fd"} err="failed to get container status \"28f2f9421d791918742e7e8c36db36a61dd45a1c90cc89b1a863b69a000636fd\": rpc error: code = NotFound desc = could not find container \"28f2f9421d791918742e7e8c36db36a61dd45a1c90cc89b1a863b69a000636fd\": container with ID starting with 28f2f9421d791918742e7e8c36db36a61dd45a1c90cc89b1a863b69a000636fd not found: ID does not exist" Dec 01 15:46:25 crc kubenswrapper[4810]: I1201 15:46:25.288302 4810 scope.go:117] "RemoveContainer" containerID="a65862ef0c2f4ee327d589d9718c9446701d6645b97a5e3add90f2d71ca543d9" Dec 01 15:46:25 crc kubenswrapper[4810]: E1201 15:46:25.288619 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a65862ef0c2f4ee327d589d9718c9446701d6645b97a5e3add90f2d71ca543d9\": container with ID starting with a65862ef0c2f4ee327d589d9718c9446701d6645b97a5e3add90f2d71ca543d9 not found: ID does not exist" containerID="a65862ef0c2f4ee327d589d9718c9446701d6645b97a5e3add90f2d71ca543d9" Dec 01 15:46:25 crc kubenswrapper[4810]: I1201 15:46:25.288643 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a65862ef0c2f4ee327d589d9718c9446701d6645b97a5e3add90f2d71ca543d9"} err="failed to get container status \"a65862ef0c2f4ee327d589d9718c9446701d6645b97a5e3add90f2d71ca543d9\": rpc error: code = NotFound desc = could not find container \"a65862ef0c2f4ee327d589d9718c9446701d6645b97a5e3add90f2d71ca543d9\": container with ID starting with a65862ef0c2f4ee327d589d9718c9446701d6645b97a5e3add90f2d71ca543d9 not found: ID does not exist" Dec 01 15:46:25 crc kubenswrapper[4810]: I1201 15:46:25.288661 4810 scope.go:117] "RemoveContainer" containerID="bfd28a94abbf0c734f31fe59939c624c92f1e15e402651999e993ba10adc7c01" Dec 01 15:46:25 crc kubenswrapper[4810]: E1201 15:46:25.289006 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfd28a94abbf0c734f31fe59939c624c92f1e15e402651999e993ba10adc7c01\": container with ID starting with bfd28a94abbf0c734f31fe59939c624c92f1e15e402651999e993ba10adc7c01 not found: ID does not exist" containerID="bfd28a94abbf0c734f31fe59939c624c92f1e15e402651999e993ba10adc7c01" Dec 01 15:46:25 crc kubenswrapper[4810]: I1201 15:46:25.289027 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfd28a94abbf0c734f31fe59939c624c92f1e15e402651999e993ba10adc7c01"} err="failed to get container status \"bfd28a94abbf0c734f31fe59939c624c92f1e15e402651999e993ba10adc7c01\": rpc error: code = NotFound desc = could not find container \"bfd28a94abbf0c734f31fe59939c624c92f1e15e402651999e993ba10adc7c01\": container with ID starting with bfd28a94abbf0c734f31fe59939c624c92f1e15e402651999e993ba10adc7c01 not found: ID does not exist" Dec 01 15:46:25 crc kubenswrapper[4810]: I1201 15:46:25.292560 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7hxnx"] Dec 01 15:46:25 crc kubenswrapper[4810]: I1201 15:46:25.297391 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7hxnx"] Dec 01 15:46:26 crc kubenswrapper[4810]: I1201 15:46:26.505837 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7bb8807-1be6-4fc7-8eef-095c80d67cd6" path="/var/lib/kubelet/pods/f7bb8807-1be6-4fc7-8eef-095c80d67cd6/volumes" Dec 01 15:46:32 crc kubenswrapper[4810]: I1201 15:46:32.972287 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:46:32 crc kubenswrapper[4810]: I1201 15:46:32.972921 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:46:32 crc kubenswrapper[4810]: I1201 15:46:32.972968 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 15:46:32 crc kubenswrapper[4810]: I1201 15:46:32.973717 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"adfbc88df5691584126b7883ecbdc4c9c6160917ae2024226dfaa689c64d123b"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 15:46:32 crc kubenswrapper[4810]: I1201 15:46:32.973794 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://adfbc88df5691584126b7883ecbdc4c9c6160917ae2024226dfaa689c64d123b" gracePeriod=600 Dec 01 15:46:33 crc kubenswrapper[4810]: I1201 15:46:33.269935 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="adfbc88df5691584126b7883ecbdc4c9c6160917ae2024226dfaa689c64d123b" exitCode=0 Dec 01 15:46:33 crc kubenswrapper[4810]: I1201 15:46:33.269975 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"adfbc88df5691584126b7883ecbdc4c9c6160917ae2024226dfaa689c64d123b"} Dec 01 15:46:33 crc kubenswrapper[4810]: I1201 15:46:33.270005 4810 scope.go:117] "RemoveContainer" containerID="147e29d72ebddd0193a9d6ec0345bfae1923e30d9a4d4b81296f350c399e8096" Dec 01 15:46:34 crc kubenswrapper[4810]: I1201 15:46:34.278052 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889"} Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.048022 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lhtl2"] Dec 01 15:48:09 crc kubenswrapper[4810]: E1201 15:48:09.048966 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7bb8807-1be6-4fc7-8eef-095c80d67cd6" containerName="registry-server" Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.048984 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7bb8807-1be6-4fc7-8eef-095c80d67cd6" containerName="registry-server" Dec 01 15:48:09 crc kubenswrapper[4810]: E1201 15:48:09.049014 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7bb8807-1be6-4fc7-8eef-095c80d67cd6" containerName="extract-utilities" Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.049023 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7bb8807-1be6-4fc7-8eef-095c80d67cd6" containerName="extract-utilities" Dec 01 15:48:09 crc kubenswrapper[4810]: E1201 15:48:09.049036 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7bb8807-1be6-4fc7-8eef-095c80d67cd6" containerName="extract-content" Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.049043 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7bb8807-1be6-4fc7-8eef-095c80d67cd6" containerName="extract-content" Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.049223 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7bb8807-1be6-4fc7-8eef-095c80d67cd6" containerName="registry-server" Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.050400 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lhtl2" Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.057017 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lhtl2"] Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.126802 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/781e228c-1722-4ded-b152-1c0fc69dd875-utilities\") pod \"redhat-operators-lhtl2\" (UID: \"781e228c-1722-4ded-b152-1c0fc69dd875\") " pod="openshift-marketplace/redhat-operators-lhtl2" Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.126908 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4qwg\" (UniqueName: \"kubernetes.io/projected/781e228c-1722-4ded-b152-1c0fc69dd875-kube-api-access-z4qwg\") pod \"redhat-operators-lhtl2\" (UID: \"781e228c-1722-4ded-b152-1c0fc69dd875\") " pod="openshift-marketplace/redhat-operators-lhtl2" Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.126967 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/781e228c-1722-4ded-b152-1c0fc69dd875-catalog-content\") pod \"redhat-operators-lhtl2\" (UID: \"781e228c-1722-4ded-b152-1c0fc69dd875\") " pod="openshift-marketplace/redhat-operators-lhtl2" Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.228424 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/781e228c-1722-4ded-b152-1c0fc69dd875-catalog-content\") pod \"redhat-operators-lhtl2\" (UID: \"781e228c-1722-4ded-b152-1c0fc69dd875\") " pod="openshift-marketplace/redhat-operators-lhtl2" Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.228584 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/781e228c-1722-4ded-b152-1c0fc69dd875-utilities\") pod \"redhat-operators-lhtl2\" (UID: \"781e228c-1722-4ded-b152-1c0fc69dd875\") " pod="openshift-marketplace/redhat-operators-lhtl2" Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.228654 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4qwg\" (UniqueName: \"kubernetes.io/projected/781e228c-1722-4ded-b152-1c0fc69dd875-kube-api-access-z4qwg\") pod \"redhat-operators-lhtl2\" (UID: \"781e228c-1722-4ded-b152-1c0fc69dd875\") " pod="openshift-marketplace/redhat-operators-lhtl2" Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.228924 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/781e228c-1722-4ded-b152-1c0fc69dd875-catalog-content\") pod \"redhat-operators-lhtl2\" (UID: \"781e228c-1722-4ded-b152-1c0fc69dd875\") " pod="openshift-marketplace/redhat-operators-lhtl2" Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.229005 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/781e228c-1722-4ded-b152-1c0fc69dd875-utilities\") pod \"redhat-operators-lhtl2\" (UID: \"781e228c-1722-4ded-b152-1c0fc69dd875\") " pod="openshift-marketplace/redhat-operators-lhtl2" Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.254071 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4qwg\" (UniqueName: \"kubernetes.io/projected/781e228c-1722-4ded-b152-1c0fc69dd875-kube-api-access-z4qwg\") pod \"redhat-operators-lhtl2\" (UID: \"781e228c-1722-4ded-b152-1c0fc69dd875\") " pod="openshift-marketplace/redhat-operators-lhtl2" Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.373807 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lhtl2" Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.871283 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lhtl2"] Dec 01 15:48:09 crc kubenswrapper[4810]: I1201 15:48:09.979511 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lhtl2" event={"ID":"781e228c-1722-4ded-b152-1c0fc69dd875","Type":"ContainerStarted","Data":"59802abe364743303a82ca31a4b67d78560a218b04f45d5f12aaa92d8e148a1e"} Dec 01 15:48:10 crc kubenswrapper[4810]: I1201 15:48:10.988312 4810 generic.go:334] "Generic (PLEG): container finished" podID="781e228c-1722-4ded-b152-1c0fc69dd875" containerID="007255f307e9819db022976fe3f34d48158487cbaea6c8a5a2cd124b8ee32d64" exitCode=0 Dec 01 15:48:10 crc kubenswrapper[4810]: I1201 15:48:10.988370 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lhtl2" event={"ID":"781e228c-1722-4ded-b152-1c0fc69dd875","Type":"ContainerDied","Data":"007255f307e9819db022976fe3f34d48158487cbaea6c8a5a2cd124b8ee32d64"} Dec 01 15:48:11 crc kubenswrapper[4810]: I1201 15:48:11.996652 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lhtl2" event={"ID":"781e228c-1722-4ded-b152-1c0fc69dd875","Type":"ContainerStarted","Data":"98bcab58d1bb8df9f5d07010563abf6a997edd203b01c1c28273ee9810559baa"} Dec 01 15:48:13 crc kubenswrapper[4810]: I1201 15:48:13.012398 4810 generic.go:334] "Generic (PLEG): container finished" podID="781e228c-1722-4ded-b152-1c0fc69dd875" containerID="98bcab58d1bb8df9f5d07010563abf6a997edd203b01c1c28273ee9810559baa" exitCode=0 Dec 01 15:48:13 crc kubenswrapper[4810]: I1201 15:48:13.012442 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lhtl2" event={"ID":"781e228c-1722-4ded-b152-1c0fc69dd875","Type":"ContainerDied","Data":"98bcab58d1bb8df9f5d07010563abf6a997edd203b01c1c28273ee9810559baa"} Dec 01 15:48:14 crc kubenswrapper[4810]: I1201 15:48:14.021238 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lhtl2" event={"ID":"781e228c-1722-4ded-b152-1c0fc69dd875","Type":"ContainerStarted","Data":"e474953aef10c3383be2d3d4f5e4a598ec6211736c5dc382b22303e40e2bc155"} Dec 01 15:48:14 crc kubenswrapper[4810]: I1201 15:48:14.051023 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lhtl2" podStartSLOduration=2.625843501 podStartE2EDuration="5.051004637s" podCreationTimestamp="2025-12-01 15:48:09 +0000 UTC" firstStartedPulling="2025-12-01 15:48:10.989783103 +0000 UTC m=+4456.753292706" lastFinishedPulling="2025-12-01 15:48:13.414944239 +0000 UTC m=+4459.178453842" observedRunningTime="2025-12-01 15:48:14.049390203 +0000 UTC m=+4459.812899806" watchObservedRunningTime="2025-12-01 15:48:14.051004637 +0000 UTC m=+4459.814514240" Dec 01 15:48:19 crc kubenswrapper[4810]: I1201 15:48:19.373950 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lhtl2" Dec 01 15:48:19 crc kubenswrapper[4810]: I1201 15:48:19.374525 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lhtl2" Dec 01 15:48:19 crc kubenswrapper[4810]: I1201 15:48:19.418787 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lhtl2" Dec 01 15:48:20 crc kubenswrapper[4810]: I1201 15:48:20.098878 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lhtl2" Dec 01 15:48:20 crc kubenswrapper[4810]: I1201 15:48:20.146695 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lhtl2"] Dec 01 15:48:22 crc kubenswrapper[4810]: I1201 15:48:22.072539 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lhtl2" podUID="781e228c-1722-4ded-b152-1c0fc69dd875" containerName="registry-server" containerID="cri-o://e474953aef10c3383be2d3d4f5e4a598ec6211736c5dc382b22303e40e2bc155" gracePeriod=2 Dec 01 15:48:24 crc kubenswrapper[4810]: I1201 15:48:24.087523 4810 generic.go:334] "Generic (PLEG): container finished" podID="781e228c-1722-4ded-b152-1c0fc69dd875" containerID="e474953aef10c3383be2d3d4f5e4a598ec6211736c5dc382b22303e40e2bc155" exitCode=0 Dec 01 15:48:24 crc kubenswrapper[4810]: I1201 15:48:24.087857 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lhtl2" event={"ID":"781e228c-1722-4ded-b152-1c0fc69dd875","Type":"ContainerDied","Data":"e474953aef10c3383be2d3d4f5e4a598ec6211736c5dc382b22303e40e2bc155"} Dec 01 15:48:24 crc kubenswrapper[4810]: I1201 15:48:24.288855 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lhtl2" Dec 01 15:48:24 crc kubenswrapper[4810]: I1201 15:48:24.338753 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/781e228c-1722-4ded-b152-1c0fc69dd875-utilities\") pod \"781e228c-1722-4ded-b152-1c0fc69dd875\" (UID: \"781e228c-1722-4ded-b152-1c0fc69dd875\") " Dec 01 15:48:24 crc kubenswrapper[4810]: I1201 15:48:24.338824 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4qwg\" (UniqueName: \"kubernetes.io/projected/781e228c-1722-4ded-b152-1c0fc69dd875-kube-api-access-z4qwg\") pod \"781e228c-1722-4ded-b152-1c0fc69dd875\" (UID: \"781e228c-1722-4ded-b152-1c0fc69dd875\") " Dec 01 15:48:24 crc kubenswrapper[4810]: I1201 15:48:24.338861 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/781e228c-1722-4ded-b152-1c0fc69dd875-catalog-content\") pod \"781e228c-1722-4ded-b152-1c0fc69dd875\" (UID: \"781e228c-1722-4ded-b152-1c0fc69dd875\") " Dec 01 15:48:24 crc kubenswrapper[4810]: I1201 15:48:24.339734 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/781e228c-1722-4ded-b152-1c0fc69dd875-utilities" (OuterVolumeSpecName: "utilities") pod "781e228c-1722-4ded-b152-1c0fc69dd875" (UID: "781e228c-1722-4ded-b152-1c0fc69dd875"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:48:24 crc kubenswrapper[4810]: I1201 15:48:24.347822 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/781e228c-1722-4ded-b152-1c0fc69dd875-kube-api-access-z4qwg" (OuterVolumeSpecName: "kube-api-access-z4qwg") pod "781e228c-1722-4ded-b152-1c0fc69dd875" (UID: "781e228c-1722-4ded-b152-1c0fc69dd875"). InnerVolumeSpecName "kube-api-access-z4qwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:48:24 crc kubenswrapper[4810]: I1201 15:48:24.440393 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/781e228c-1722-4ded-b152-1c0fc69dd875-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:48:24 crc kubenswrapper[4810]: I1201 15:48:24.440436 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4qwg\" (UniqueName: \"kubernetes.io/projected/781e228c-1722-4ded-b152-1c0fc69dd875-kube-api-access-z4qwg\") on node \"crc\" DevicePath \"\"" Dec 01 15:48:24 crc kubenswrapper[4810]: I1201 15:48:24.445557 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/781e228c-1722-4ded-b152-1c0fc69dd875-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "781e228c-1722-4ded-b152-1c0fc69dd875" (UID: "781e228c-1722-4ded-b152-1c0fc69dd875"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:48:24 crc kubenswrapper[4810]: I1201 15:48:24.542379 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/781e228c-1722-4ded-b152-1c0fc69dd875-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:48:25 crc kubenswrapper[4810]: I1201 15:48:25.097766 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lhtl2" event={"ID":"781e228c-1722-4ded-b152-1c0fc69dd875","Type":"ContainerDied","Data":"59802abe364743303a82ca31a4b67d78560a218b04f45d5f12aaa92d8e148a1e"} Dec 01 15:48:25 crc kubenswrapper[4810]: I1201 15:48:25.097870 4810 scope.go:117] "RemoveContainer" containerID="e474953aef10c3383be2d3d4f5e4a598ec6211736c5dc382b22303e40e2bc155" Dec 01 15:48:25 crc kubenswrapper[4810]: I1201 15:48:25.097995 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lhtl2" Dec 01 15:48:25 crc kubenswrapper[4810]: I1201 15:48:25.130141 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lhtl2"] Dec 01 15:48:25 crc kubenswrapper[4810]: I1201 15:48:25.133287 4810 scope.go:117] "RemoveContainer" containerID="98bcab58d1bb8df9f5d07010563abf6a997edd203b01c1c28273ee9810559baa" Dec 01 15:48:25 crc kubenswrapper[4810]: I1201 15:48:25.139135 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lhtl2"] Dec 01 15:48:25 crc kubenswrapper[4810]: I1201 15:48:25.152777 4810 scope.go:117] "RemoveContainer" containerID="007255f307e9819db022976fe3f34d48158487cbaea6c8a5a2cd124b8ee32d64" Dec 01 15:48:26 crc kubenswrapper[4810]: I1201 15:48:26.510889 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="781e228c-1722-4ded-b152-1c0fc69dd875" path="/var/lib/kubelet/pods/781e228c-1722-4ded-b152-1c0fc69dd875/volumes" Dec 01 15:49:02 crc kubenswrapper[4810]: I1201 15:49:02.971949 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:49:02 crc kubenswrapper[4810]: I1201 15:49:02.972650 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:49:32 crc kubenswrapper[4810]: I1201 15:49:32.972625 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:49:32 crc kubenswrapper[4810]: I1201 15:49:32.973297 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:49:59 crc kubenswrapper[4810]: I1201 15:49:59.996010 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rnvxq"] Dec 01 15:49:59 crc kubenswrapper[4810]: E1201 15:49:59.997076 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="781e228c-1722-4ded-b152-1c0fc69dd875" containerName="extract-content" Dec 01 15:49:59 crc kubenswrapper[4810]: I1201 15:49:59.997100 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="781e228c-1722-4ded-b152-1c0fc69dd875" containerName="extract-content" Dec 01 15:49:59 crc kubenswrapper[4810]: E1201 15:49:59.997147 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="781e228c-1722-4ded-b152-1c0fc69dd875" containerName="extract-utilities" Dec 01 15:49:59 crc kubenswrapper[4810]: I1201 15:49:59.997161 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="781e228c-1722-4ded-b152-1c0fc69dd875" containerName="extract-utilities" Dec 01 15:49:59 crc kubenswrapper[4810]: E1201 15:49:59.997194 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="781e228c-1722-4ded-b152-1c0fc69dd875" containerName="registry-server" Dec 01 15:49:59 crc kubenswrapper[4810]: I1201 15:49:59.997206 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="781e228c-1722-4ded-b152-1c0fc69dd875" containerName="registry-server" Dec 01 15:49:59 crc kubenswrapper[4810]: I1201 15:49:59.997433 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="781e228c-1722-4ded-b152-1c0fc69dd875" containerName="registry-server" Dec 01 15:49:59 crc kubenswrapper[4810]: I1201 15:49:59.999150 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rnvxq" Dec 01 15:50:00 crc kubenswrapper[4810]: I1201 15:50:00.004686 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rnvxq"] Dec 01 15:50:00 crc kubenswrapper[4810]: I1201 15:50:00.141095 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dfc0170-c937-4ff8-ad53-b4d27482582c-utilities\") pod \"redhat-marketplace-rnvxq\" (UID: \"4dfc0170-c937-4ff8-ad53-b4d27482582c\") " pod="openshift-marketplace/redhat-marketplace-rnvxq" Dec 01 15:50:00 crc kubenswrapper[4810]: I1201 15:50:00.141159 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj6ql\" (UniqueName: \"kubernetes.io/projected/4dfc0170-c937-4ff8-ad53-b4d27482582c-kube-api-access-kj6ql\") pod \"redhat-marketplace-rnvxq\" (UID: \"4dfc0170-c937-4ff8-ad53-b4d27482582c\") " pod="openshift-marketplace/redhat-marketplace-rnvxq" Dec 01 15:50:00 crc kubenswrapper[4810]: I1201 15:50:00.141196 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dfc0170-c937-4ff8-ad53-b4d27482582c-catalog-content\") pod \"redhat-marketplace-rnvxq\" (UID: \"4dfc0170-c937-4ff8-ad53-b4d27482582c\") " pod="openshift-marketplace/redhat-marketplace-rnvxq" Dec 01 15:50:00 crc kubenswrapper[4810]: I1201 15:50:00.242334 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dfc0170-c937-4ff8-ad53-b4d27482582c-utilities\") pod \"redhat-marketplace-rnvxq\" (UID: \"4dfc0170-c937-4ff8-ad53-b4d27482582c\") " pod="openshift-marketplace/redhat-marketplace-rnvxq" Dec 01 15:50:00 crc kubenswrapper[4810]: I1201 15:50:00.242392 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj6ql\" (UniqueName: \"kubernetes.io/projected/4dfc0170-c937-4ff8-ad53-b4d27482582c-kube-api-access-kj6ql\") pod \"redhat-marketplace-rnvxq\" (UID: \"4dfc0170-c937-4ff8-ad53-b4d27482582c\") " pod="openshift-marketplace/redhat-marketplace-rnvxq" Dec 01 15:50:00 crc kubenswrapper[4810]: I1201 15:50:00.242422 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dfc0170-c937-4ff8-ad53-b4d27482582c-catalog-content\") pod \"redhat-marketplace-rnvxq\" (UID: \"4dfc0170-c937-4ff8-ad53-b4d27482582c\") " pod="openshift-marketplace/redhat-marketplace-rnvxq" Dec 01 15:50:00 crc kubenswrapper[4810]: I1201 15:50:00.243037 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dfc0170-c937-4ff8-ad53-b4d27482582c-catalog-content\") pod \"redhat-marketplace-rnvxq\" (UID: \"4dfc0170-c937-4ff8-ad53-b4d27482582c\") " pod="openshift-marketplace/redhat-marketplace-rnvxq" Dec 01 15:50:00 crc kubenswrapper[4810]: I1201 15:50:00.243037 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dfc0170-c937-4ff8-ad53-b4d27482582c-utilities\") pod \"redhat-marketplace-rnvxq\" (UID: \"4dfc0170-c937-4ff8-ad53-b4d27482582c\") " pod="openshift-marketplace/redhat-marketplace-rnvxq" Dec 01 15:50:00 crc kubenswrapper[4810]: I1201 15:50:00.261522 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj6ql\" (UniqueName: \"kubernetes.io/projected/4dfc0170-c937-4ff8-ad53-b4d27482582c-kube-api-access-kj6ql\") pod \"redhat-marketplace-rnvxq\" (UID: \"4dfc0170-c937-4ff8-ad53-b4d27482582c\") " pod="openshift-marketplace/redhat-marketplace-rnvxq" Dec 01 15:50:00 crc kubenswrapper[4810]: I1201 15:50:00.317563 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rnvxq" Dec 01 15:50:00 crc kubenswrapper[4810]: I1201 15:50:00.811187 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rnvxq"] Dec 01 15:50:01 crc kubenswrapper[4810]: I1201 15:50:01.783548 4810 generic.go:334] "Generic (PLEG): container finished" podID="4dfc0170-c937-4ff8-ad53-b4d27482582c" containerID="921ffe3127437a18010ddefae7197fe6d95fc44c5e85aae0dacef1959533f0c4" exitCode=0 Dec 01 15:50:01 crc kubenswrapper[4810]: I1201 15:50:01.783602 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rnvxq" event={"ID":"4dfc0170-c937-4ff8-ad53-b4d27482582c","Type":"ContainerDied","Data":"921ffe3127437a18010ddefae7197fe6d95fc44c5e85aae0dacef1959533f0c4"} Dec 01 15:50:01 crc kubenswrapper[4810]: I1201 15:50:01.783629 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rnvxq" event={"ID":"4dfc0170-c937-4ff8-ad53-b4d27482582c","Type":"ContainerStarted","Data":"e51781c006216323763bd072740b0991a483aefb454a3c5924c15a721585c3bc"} Dec 01 15:50:02 crc kubenswrapper[4810]: I1201 15:50:02.972360 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:50:02 crc kubenswrapper[4810]: I1201 15:50:02.972817 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:50:02 crc kubenswrapper[4810]: I1201 15:50:02.972868 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 15:50:02 crc kubenswrapper[4810]: I1201 15:50:02.973339 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 15:50:02 crc kubenswrapper[4810]: I1201 15:50:02.973389 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" gracePeriod=600 Dec 01 15:50:03 crc kubenswrapper[4810]: E1201 15:50:03.095338 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:50:03 crc kubenswrapper[4810]: I1201 15:50:03.799694 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" exitCode=0 Dec 01 15:50:03 crc kubenswrapper[4810]: I1201 15:50:03.799781 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889"} Dec 01 15:50:03 crc kubenswrapper[4810]: I1201 15:50:03.799861 4810 scope.go:117] "RemoveContainer" containerID="adfbc88df5691584126b7883ecbdc4c9c6160917ae2024226dfaa689c64d123b" Dec 01 15:50:03 crc kubenswrapper[4810]: I1201 15:50:03.800562 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:50:03 crc kubenswrapper[4810]: E1201 15:50:03.800959 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:50:03 crc kubenswrapper[4810]: I1201 15:50:03.802043 4810 generic.go:334] "Generic (PLEG): container finished" podID="4dfc0170-c937-4ff8-ad53-b4d27482582c" containerID="a45dc3951b58386684eb5015f69aea95ac3400022dfede9f45d29a10d2f9c60a" exitCode=0 Dec 01 15:50:03 crc kubenswrapper[4810]: I1201 15:50:03.802073 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rnvxq" event={"ID":"4dfc0170-c937-4ff8-ad53-b4d27482582c","Type":"ContainerDied","Data":"a45dc3951b58386684eb5015f69aea95ac3400022dfede9f45d29a10d2f9c60a"} Dec 01 15:50:04 crc kubenswrapper[4810]: I1201 15:50:04.814208 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rnvxq" event={"ID":"4dfc0170-c937-4ff8-ad53-b4d27482582c","Type":"ContainerStarted","Data":"2eee18b7cc7b645756a80e9029288784cd8ea882f506e29deb2b7513c8eda3b2"} Dec 01 15:50:04 crc kubenswrapper[4810]: I1201 15:50:04.841233 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rnvxq" podStartSLOduration=3.304575567 podStartE2EDuration="5.841211494s" podCreationTimestamp="2025-12-01 15:49:59 +0000 UTC" firstStartedPulling="2025-12-01 15:50:01.785027726 +0000 UTC m=+4567.548537319" lastFinishedPulling="2025-12-01 15:50:04.321663633 +0000 UTC m=+4570.085173246" observedRunningTime="2025-12-01 15:50:04.837244776 +0000 UTC m=+4570.600754379" watchObservedRunningTime="2025-12-01 15:50:04.841211494 +0000 UTC m=+4570.604721097" Dec 01 15:50:10 crc kubenswrapper[4810]: I1201 15:50:10.318862 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rnvxq" Dec 01 15:50:10 crc kubenswrapper[4810]: I1201 15:50:10.320026 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rnvxq" Dec 01 15:50:10 crc kubenswrapper[4810]: I1201 15:50:10.369704 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rnvxq" Dec 01 15:50:10 crc kubenswrapper[4810]: I1201 15:50:10.906369 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rnvxq" Dec 01 15:50:10 crc kubenswrapper[4810]: I1201 15:50:10.947660 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rnvxq"] Dec 01 15:50:12 crc kubenswrapper[4810]: I1201 15:50:12.873783 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rnvxq" podUID="4dfc0170-c937-4ff8-ad53-b4d27482582c" containerName="registry-server" containerID="cri-o://2eee18b7cc7b645756a80e9029288784cd8ea882f506e29deb2b7513c8eda3b2" gracePeriod=2 Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.368074 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rnvxq" Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.430440 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dfc0170-c937-4ff8-ad53-b4d27482582c-catalog-content\") pod \"4dfc0170-c937-4ff8-ad53-b4d27482582c\" (UID: \"4dfc0170-c937-4ff8-ad53-b4d27482582c\") " Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.430551 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dfc0170-c937-4ff8-ad53-b4d27482582c-utilities\") pod \"4dfc0170-c937-4ff8-ad53-b4d27482582c\" (UID: \"4dfc0170-c937-4ff8-ad53-b4d27482582c\") " Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.430608 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kj6ql\" (UniqueName: \"kubernetes.io/projected/4dfc0170-c937-4ff8-ad53-b4d27482582c-kube-api-access-kj6ql\") pod \"4dfc0170-c937-4ff8-ad53-b4d27482582c\" (UID: \"4dfc0170-c937-4ff8-ad53-b4d27482582c\") " Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.431514 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dfc0170-c937-4ff8-ad53-b4d27482582c-utilities" (OuterVolumeSpecName: "utilities") pod "4dfc0170-c937-4ff8-ad53-b4d27482582c" (UID: "4dfc0170-c937-4ff8-ad53-b4d27482582c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.450485 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dfc0170-c937-4ff8-ad53-b4d27482582c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4dfc0170-c937-4ff8-ad53-b4d27482582c" (UID: "4dfc0170-c937-4ff8-ad53-b4d27482582c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.532193 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dfc0170-c937-4ff8-ad53-b4d27482582c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.532233 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dfc0170-c937-4ff8-ad53-b4d27482582c-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.752729 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dfc0170-c937-4ff8-ad53-b4d27482582c-kube-api-access-kj6ql" (OuterVolumeSpecName: "kube-api-access-kj6ql") pod "4dfc0170-c937-4ff8-ad53-b4d27482582c" (UID: "4dfc0170-c937-4ff8-ad53-b4d27482582c"). InnerVolumeSpecName "kube-api-access-kj6ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.836660 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kj6ql\" (UniqueName: \"kubernetes.io/projected/4dfc0170-c937-4ff8-ad53-b4d27482582c-kube-api-access-kj6ql\") on node \"crc\" DevicePath \"\"" Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.882023 4810 generic.go:334] "Generic (PLEG): container finished" podID="4dfc0170-c937-4ff8-ad53-b4d27482582c" containerID="2eee18b7cc7b645756a80e9029288784cd8ea882f506e29deb2b7513c8eda3b2" exitCode=0 Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.882072 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rnvxq" event={"ID":"4dfc0170-c937-4ff8-ad53-b4d27482582c","Type":"ContainerDied","Data":"2eee18b7cc7b645756a80e9029288784cd8ea882f506e29deb2b7513c8eda3b2"} Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.882076 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rnvxq" Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.882101 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rnvxq" event={"ID":"4dfc0170-c937-4ff8-ad53-b4d27482582c","Type":"ContainerDied","Data":"e51781c006216323763bd072740b0991a483aefb454a3c5924c15a721585c3bc"} Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.882124 4810 scope.go:117] "RemoveContainer" containerID="2eee18b7cc7b645756a80e9029288784cd8ea882f506e29deb2b7513c8eda3b2" Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.919610 4810 scope.go:117] "RemoveContainer" containerID="a45dc3951b58386684eb5015f69aea95ac3400022dfede9f45d29a10d2f9c60a" Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.921570 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rnvxq"] Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.929522 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rnvxq"] Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.936496 4810 scope.go:117] "RemoveContainer" containerID="921ffe3127437a18010ddefae7197fe6d95fc44c5e85aae0dacef1959533f0c4" Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.960639 4810 scope.go:117] "RemoveContainer" containerID="2eee18b7cc7b645756a80e9029288784cd8ea882f506e29deb2b7513c8eda3b2" Dec 01 15:50:13 crc kubenswrapper[4810]: E1201 15:50:13.961354 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2eee18b7cc7b645756a80e9029288784cd8ea882f506e29deb2b7513c8eda3b2\": container with ID starting with 2eee18b7cc7b645756a80e9029288784cd8ea882f506e29deb2b7513c8eda3b2 not found: ID does not exist" containerID="2eee18b7cc7b645756a80e9029288784cd8ea882f506e29deb2b7513c8eda3b2" Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.961420 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2eee18b7cc7b645756a80e9029288784cd8ea882f506e29deb2b7513c8eda3b2"} err="failed to get container status \"2eee18b7cc7b645756a80e9029288784cd8ea882f506e29deb2b7513c8eda3b2\": rpc error: code = NotFound desc = could not find container \"2eee18b7cc7b645756a80e9029288784cd8ea882f506e29deb2b7513c8eda3b2\": container with ID starting with 2eee18b7cc7b645756a80e9029288784cd8ea882f506e29deb2b7513c8eda3b2 not found: ID does not exist" Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.961447 4810 scope.go:117] "RemoveContainer" containerID="a45dc3951b58386684eb5015f69aea95ac3400022dfede9f45d29a10d2f9c60a" Dec 01 15:50:13 crc kubenswrapper[4810]: E1201 15:50:13.962003 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a45dc3951b58386684eb5015f69aea95ac3400022dfede9f45d29a10d2f9c60a\": container with ID starting with a45dc3951b58386684eb5015f69aea95ac3400022dfede9f45d29a10d2f9c60a not found: ID does not exist" containerID="a45dc3951b58386684eb5015f69aea95ac3400022dfede9f45d29a10d2f9c60a" Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.962031 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a45dc3951b58386684eb5015f69aea95ac3400022dfede9f45d29a10d2f9c60a"} err="failed to get container status \"a45dc3951b58386684eb5015f69aea95ac3400022dfede9f45d29a10d2f9c60a\": rpc error: code = NotFound desc = could not find container \"a45dc3951b58386684eb5015f69aea95ac3400022dfede9f45d29a10d2f9c60a\": container with ID starting with a45dc3951b58386684eb5015f69aea95ac3400022dfede9f45d29a10d2f9c60a not found: ID does not exist" Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.962047 4810 scope.go:117] "RemoveContainer" containerID="921ffe3127437a18010ddefae7197fe6d95fc44c5e85aae0dacef1959533f0c4" Dec 01 15:50:13 crc kubenswrapper[4810]: E1201 15:50:13.962390 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"921ffe3127437a18010ddefae7197fe6d95fc44c5e85aae0dacef1959533f0c4\": container with ID starting with 921ffe3127437a18010ddefae7197fe6d95fc44c5e85aae0dacef1959533f0c4 not found: ID does not exist" containerID="921ffe3127437a18010ddefae7197fe6d95fc44c5e85aae0dacef1959533f0c4" Dec 01 15:50:13 crc kubenswrapper[4810]: I1201 15:50:13.962430 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"921ffe3127437a18010ddefae7197fe6d95fc44c5e85aae0dacef1959533f0c4"} err="failed to get container status \"921ffe3127437a18010ddefae7197fe6d95fc44c5e85aae0dacef1959533f0c4\": rpc error: code = NotFound desc = could not find container \"921ffe3127437a18010ddefae7197fe6d95fc44c5e85aae0dacef1959533f0c4\": container with ID starting with 921ffe3127437a18010ddefae7197fe6d95fc44c5e85aae0dacef1959533f0c4 not found: ID does not exist" Dec 01 15:50:14 crc kubenswrapper[4810]: I1201 15:50:14.502173 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dfc0170-c937-4ff8-ad53-b4d27482582c" path="/var/lib/kubelet/pods/4dfc0170-c937-4ff8-ad53-b4d27482582c/volumes" Dec 01 15:50:15 crc kubenswrapper[4810]: I1201 15:50:15.491000 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:50:15 crc kubenswrapper[4810]: E1201 15:50:15.491276 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:50:26 crc kubenswrapper[4810]: I1201 15:50:26.490898 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:50:26 crc kubenswrapper[4810]: E1201 15:50:26.491732 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:50:39 crc kubenswrapper[4810]: I1201 15:50:39.491028 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:50:39 crc kubenswrapper[4810]: E1201 15:50:39.491623 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:50:53 crc kubenswrapper[4810]: I1201 15:50:53.491043 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:50:53 crc kubenswrapper[4810]: E1201 15:50:53.491938 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:51:07 crc kubenswrapper[4810]: I1201 15:51:07.490732 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:51:07 crc kubenswrapper[4810]: E1201 15:51:07.491528 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:51:20 crc kubenswrapper[4810]: I1201 15:51:20.491821 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:51:20 crc kubenswrapper[4810]: E1201 15:51:20.492582 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:51:31 crc kubenswrapper[4810]: I1201 15:51:31.490796 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:51:31 crc kubenswrapper[4810]: E1201 15:51:31.491525 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:51:46 crc kubenswrapper[4810]: I1201 15:51:46.490930 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:51:46 crc kubenswrapper[4810]: E1201 15:51:46.491857 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:51:58 crc kubenswrapper[4810]: I1201 15:51:58.491024 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:51:58 crc kubenswrapper[4810]: E1201 15:51:58.492014 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:52:09 crc kubenswrapper[4810]: I1201 15:52:09.490813 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:52:09 crc kubenswrapper[4810]: E1201 15:52:09.491511 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:52:20 crc kubenswrapper[4810]: I1201 15:52:20.491867 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:52:20 crc kubenswrapper[4810]: E1201 15:52:20.494508 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:52:35 crc kubenswrapper[4810]: I1201 15:52:35.491027 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:52:35 crc kubenswrapper[4810]: E1201 15:52:35.491838 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:52:47 crc kubenswrapper[4810]: I1201 15:52:47.490803 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:52:47 crc kubenswrapper[4810]: E1201 15:52:47.491791 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:53:00 crc kubenswrapper[4810]: I1201 15:53:00.491767 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:53:00 crc kubenswrapper[4810]: E1201 15:53:00.492486 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:53:11 crc kubenswrapper[4810]: I1201 15:53:11.491574 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:53:11 crc kubenswrapper[4810]: E1201 15:53:11.492381 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:53:24 crc kubenswrapper[4810]: I1201 15:53:24.496113 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:53:24 crc kubenswrapper[4810]: E1201 15:53:24.497060 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:53:28 crc kubenswrapper[4810]: I1201 15:53:28.957660 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lxcfp"] Dec 01 15:53:28 crc kubenswrapper[4810]: E1201 15:53:28.958423 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dfc0170-c937-4ff8-ad53-b4d27482582c" containerName="registry-server" Dec 01 15:53:28 crc kubenswrapper[4810]: I1201 15:53:28.958440 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dfc0170-c937-4ff8-ad53-b4d27482582c" containerName="registry-server" Dec 01 15:53:28 crc kubenswrapper[4810]: E1201 15:53:28.958460 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dfc0170-c937-4ff8-ad53-b4d27482582c" containerName="extract-utilities" Dec 01 15:53:28 crc kubenswrapper[4810]: I1201 15:53:28.958490 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dfc0170-c937-4ff8-ad53-b4d27482582c" containerName="extract-utilities" Dec 01 15:53:28 crc kubenswrapper[4810]: E1201 15:53:28.958511 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dfc0170-c937-4ff8-ad53-b4d27482582c" containerName="extract-content" Dec 01 15:53:28 crc kubenswrapper[4810]: I1201 15:53:28.958522 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dfc0170-c937-4ff8-ad53-b4d27482582c" containerName="extract-content" Dec 01 15:53:28 crc kubenswrapper[4810]: I1201 15:53:28.958716 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dfc0170-c937-4ff8-ad53-b4d27482582c" containerName="registry-server" Dec 01 15:53:28 crc kubenswrapper[4810]: I1201 15:53:28.960125 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lxcfp" Dec 01 15:53:29 crc kubenswrapper[4810]: I1201 15:53:28.978910 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lxcfp"] Dec 01 15:53:29 crc kubenswrapper[4810]: I1201 15:53:29.115543 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzchj\" (UniqueName: \"kubernetes.io/projected/61fdf03d-257d-4d46-93c4-8b772371cf61-kube-api-access-hzchj\") pod \"certified-operators-lxcfp\" (UID: \"61fdf03d-257d-4d46-93c4-8b772371cf61\") " pod="openshift-marketplace/certified-operators-lxcfp" Dec 01 15:53:29 crc kubenswrapper[4810]: I1201 15:53:29.115607 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61fdf03d-257d-4d46-93c4-8b772371cf61-utilities\") pod \"certified-operators-lxcfp\" (UID: \"61fdf03d-257d-4d46-93c4-8b772371cf61\") " pod="openshift-marketplace/certified-operators-lxcfp" Dec 01 15:53:29 crc kubenswrapper[4810]: I1201 15:53:29.115743 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61fdf03d-257d-4d46-93c4-8b772371cf61-catalog-content\") pod \"certified-operators-lxcfp\" (UID: \"61fdf03d-257d-4d46-93c4-8b772371cf61\") " pod="openshift-marketplace/certified-operators-lxcfp" Dec 01 15:53:29 crc kubenswrapper[4810]: I1201 15:53:29.216780 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61fdf03d-257d-4d46-93c4-8b772371cf61-catalog-content\") pod \"certified-operators-lxcfp\" (UID: \"61fdf03d-257d-4d46-93c4-8b772371cf61\") " pod="openshift-marketplace/certified-operators-lxcfp" Dec 01 15:53:29 crc kubenswrapper[4810]: I1201 15:53:29.217241 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzchj\" (UniqueName: \"kubernetes.io/projected/61fdf03d-257d-4d46-93c4-8b772371cf61-kube-api-access-hzchj\") pod \"certified-operators-lxcfp\" (UID: \"61fdf03d-257d-4d46-93c4-8b772371cf61\") " pod="openshift-marketplace/certified-operators-lxcfp" Dec 01 15:53:29 crc kubenswrapper[4810]: I1201 15:53:29.217277 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61fdf03d-257d-4d46-93c4-8b772371cf61-utilities\") pod \"certified-operators-lxcfp\" (UID: \"61fdf03d-257d-4d46-93c4-8b772371cf61\") " pod="openshift-marketplace/certified-operators-lxcfp" Dec 01 15:53:29 crc kubenswrapper[4810]: I1201 15:53:29.217350 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61fdf03d-257d-4d46-93c4-8b772371cf61-catalog-content\") pod \"certified-operators-lxcfp\" (UID: \"61fdf03d-257d-4d46-93c4-8b772371cf61\") " pod="openshift-marketplace/certified-operators-lxcfp" Dec 01 15:53:29 crc kubenswrapper[4810]: I1201 15:53:29.217865 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61fdf03d-257d-4d46-93c4-8b772371cf61-utilities\") pod \"certified-operators-lxcfp\" (UID: \"61fdf03d-257d-4d46-93c4-8b772371cf61\") " pod="openshift-marketplace/certified-operators-lxcfp" Dec 01 15:53:29 crc kubenswrapper[4810]: I1201 15:53:29.241394 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzchj\" (UniqueName: \"kubernetes.io/projected/61fdf03d-257d-4d46-93c4-8b772371cf61-kube-api-access-hzchj\") pod \"certified-operators-lxcfp\" (UID: \"61fdf03d-257d-4d46-93c4-8b772371cf61\") " pod="openshift-marketplace/certified-operators-lxcfp" Dec 01 15:53:29 crc kubenswrapper[4810]: I1201 15:53:29.320765 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lxcfp" Dec 01 15:53:29 crc kubenswrapper[4810]: I1201 15:53:29.812943 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lxcfp"] Dec 01 15:53:30 crc kubenswrapper[4810]: I1201 15:53:30.332431 4810 generic.go:334] "Generic (PLEG): container finished" podID="61fdf03d-257d-4d46-93c4-8b772371cf61" containerID="782e1c99d910749de0d78a3fb2da995a2e63d505efda016f54c41bf8c7eb5152" exitCode=0 Dec 01 15:53:30 crc kubenswrapper[4810]: I1201 15:53:30.332564 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxcfp" event={"ID":"61fdf03d-257d-4d46-93c4-8b772371cf61","Type":"ContainerDied","Data":"782e1c99d910749de0d78a3fb2da995a2e63d505efda016f54c41bf8c7eb5152"} Dec 01 15:53:30 crc kubenswrapper[4810]: I1201 15:53:30.332875 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxcfp" event={"ID":"61fdf03d-257d-4d46-93c4-8b772371cf61","Type":"ContainerStarted","Data":"7171653cf638f9ab8a14f4bbe3cbb95168145bb484eea8eefc0c2d662ac7fecf"} Dec 01 15:53:30 crc kubenswrapper[4810]: I1201 15:53:30.335521 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 15:53:34 crc kubenswrapper[4810]: I1201 15:53:34.363983 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxcfp" event={"ID":"61fdf03d-257d-4d46-93c4-8b772371cf61","Type":"ContainerStarted","Data":"b2e5141706467610d79bea0fc62e605112035d2280a1cf83a3f9a340827bd720"} Dec 01 15:53:35 crc kubenswrapper[4810]: I1201 15:53:35.373420 4810 generic.go:334] "Generic (PLEG): container finished" podID="61fdf03d-257d-4d46-93c4-8b772371cf61" containerID="b2e5141706467610d79bea0fc62e605112035d2280a1cf83a3f9a340827bd720" exitCode=0 Dec 01 15:53:35 crc kubenswrapper[4810]: I1201 15:53:35.373488 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxcfp" event={"ID":"61fdf03d-257d-4d46-93c4-8b772371cf61","Type":"ContainerDied","Data":"b2e5141706467610d79bea0fc62e605112035d2280a1cf83a3f9a340827bd720"} Dec 01 15:53:36 crc kubenswrapper[4810]: I1201 15:53:36.383498 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxcfp" event={"ID":"61fdf03d-257d-4d46-93c4-8b772371cf61","Type":"ContainerStarted","Data":"16e1a603cc9aa502a52d917fae64c2a401a08e03eea47afdcbc47a89b6271eba"} Dec 01 15:53:36 crc kubenswrapper[4810]: I1201 15:53:36.408337 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lxcfp" podStartSLOduration=2.539057202 podStartE2EDuration="8.408316698s" podCreationTimestamp="2025-12-01 15:53:28 +0000 UTC" firstStartedPulling="2025-12-01 15:53:30.334921389 +0000 UTC m=+4776.098431032" lastFinishedPulling="2025-12-01 15:53:36.204180915 +0000 UTC m=+4781.967690528" observedRunningTime="2025-12-01 15:53:36.400169236 +0000 UTC m=+4782.163678839" watchObservedRunningTime="2025-12-01 15:53:36.408316698 +0000 UTC m=+4782.171826301" Dec 01 15:53:37 crc kubenswrapper[4810]: I1201 15:53:37.491293 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:53:37 crc kubenswrapper[4810]: E1201 15:53:37.491554 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:53:39 crc kubenswrapper[4810]: I1201 15:53:39.321823 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lxcfp" Dec 01 15:53:39 crc kubenswrapper[4810]: I1201 15:53:39.322179 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lxcfp" Dec 01 15:53:39 crc kubenswrapper[4810]: I1201 15:53:39.385799 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lxcfp" Dec 01 15:53:49 crc kubenswrapper[4810]: I1201 15:53:49.367069 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lxcfp" Dec 01 15:53:49 crc kubenswrapper[4810]: I1201 15:53:49.425577 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lxcfp"] Dec 01 15:53:49 crc kubenswrapper[4810]: I1201 15:53:49.478293 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zcdww"] Dec 01 15:53:49 crc kubenswrapper[4810]: I1201 15:53:49.479872 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zcdww" podUID="d0e41ca0-4806-49a1-bb76-55f140d4ad99" containerName="registry-server" containerID="cri-o://351aa4d079c8d533d4855417d5a0ecdb3fa2637be834c9fdd03b0e6ab9646ec6" gracePeriod=2 Dec 01 15:53:49 crc kubenswrapper[4810]: I1201 15:53:49.492502 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:53:49 crc kubenswrapper[4810]: E1201 15:53:49.493096 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:53:50 crc kubenswrapper[4810]: I1201 15:53:50.506649 4810 generic.go:334] "Generic (PLEG): container finished" podID="d0e41ca0-4806-49a1-bb76-55f140d4ad99" containerID="351aa4d079c8d533d4855417d5a0ecdb3fa2637be834c9fdd03b0e6ab9646ec6" exitCode=0 Dec 01 15:53:50 crc kubenswrapper[4810]: I1201 15:53:50.506726 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zcdww" event={"ID":"d0e41ca0-4806-49a1-bb76-55f140d4ad99","Type":"ContainerDied","Data":"351aa4d079c8d533d4855417d5a0ecdb3fa2637be834c9fdd03b0e6ab9646ec6"} Dec 01 15:53:50 crc kubenswrapper[4810]: I1201 15:53:50.957635 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zcdww" Dec 01 15:53:51 crc kubenswrapper[4810]: I1201 15:53:51.039308 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0e41ca0-4806-49a1-bb76-55f140d4ad99-catalog-content\") pod \"d0e41ca0-4806-49a1-bb76-55f140d4ad99\" (UID: \"d0e41ca0-4806-49a1-bb76-55f140d4ad99\") " Dec 01 15:53:51 crc kubenswrapper[4810]: I1201 15:53:51.039433 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8wf2\" (UniqueName: \"kubernetes.io/projected/d0e41ca0-4806-49a1-bb76-55f140d4ad99-kube-api-access-b8wf2\") pod \"d0e41ca0-4806-49a1-bb76-55f140d4ad99\" (UID: \"d0e41ca0-4806-49a1-bb76-55f140d4ad99\") " Dec 01 15:53:51 crc kubenswrapper[4810]: I1201 15:53:51.039465 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0e41ca0-4806-49a1-bb76-55f140d4ad99-utilities\") pod \"d0e41ca0-4806-49a1-bb76-55f140d4ad99\" (UID: \"d0e41ca0-4806-49a1-bb76-55f140d4ad99\") " Dec 01 15:53:51 crc kubenswrapper[4810]: I1201 15:53:51.040246 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0e41ca0-4806-49a1-bb76-55f140d4ad99-utilities" (OuterVolumeSpecName: "utilities") pod "d0e41ca0-4806-49a1-bb76-55f140d4ad99" (UID: "d0e41ca0-4806-49a1-bb76-55f140d4ad99"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:53:51 crc kubenswrapper[4810]: I1201 15:53:51.044786 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0e41ca0-4806-49a1-bb76-55f140d4ad99-kube-api-access-b8wf2" (OuterVolumeSpecName: "kube-api-access-b8wf2") pod "d0e41ca0-4806-49a1-bb76-55f140d4ad99" (UID: "d0e41ca0-4806-49a1-bb76-55f140d4ad99"). InnerVolumeSpecName "kube-api-access-b8wf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:53:51 crc kubenswrapper[4810]: I1201 15:53:51.090860 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0e41ca0-4806-49a1-bb76-55f140d4ad99-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0e41ca0-4806-49a1-bb76-55f140d4ad99" (UID: "d0e41ca0-4806-49a1-bb76-55f140d4ad99"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:53:51 crc kubenswrapper[4810]: I1201 15:53:51.141515 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0e41ca0-4806-49a1-bb76-55f140d4ad99-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:51 crc kubenswrapper[4810]: I1201 15:53:51.141564 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8wf2\" (UniqueName: \"kubernetes.io/projected/d0e41ca0-4806-49a1-bb76-55f140d4ad99-kube-api-access-b8wf2\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:51 crc kubenswrapper[4810]: I1201 15:53:51.141580 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0e41ca0-4806-49a1-bb76-55f140d4ad99-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:53:51 crc kubenswrapper[4810]: I1201 15:53:51.515730 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zcdww" event={"ID":"d0e41ca0-4806-49a1-bb76-55f140d4ad99","Type":"ContainerDied","Data":"2ddcd44f3d83e5e0caefb41af9690a578544a689fd5f8b5124f3ac049d6cf510"} Dec 01 15:53:51 crc kubenswrapper[4810]: I1201 15:53:51.515779 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zcdww" Dec 01 15:53:51 crc kubenswrapper[4810]: I1201 15:53:51.515782 4810 scope.go:117] "RemoveContainer" containerID="351aa4d079c8d533d4855417d5a0ecdb3fa2637be834c9fdd03b0e6ab9646ec6" Dec 01 15:53:51 crc kubenswrapper[4810]: I1201 15:53:51.547139 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zcdww"] Dec 01 15:53:51 crc kubenswrapper[4810]: I1201 15:53:51.552774 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zcdww"] Dec 01 15:53:51 crc kubenswrapper[4810]: I1201 15:53:51.552931 4810 scope.go:117] "RemoveContainer" containerID="f071254a2c54da8de4c155fdd989f0555480d90cc553521063e705508da94da5" Dec 01 15:53:51 crc kubenswrapper[4810]: I1201 15:53:51.571011 4810 scope.go:117] "RemoveContainer" containerID="1713876d8eea0f3772be1a21fe71192f68eb5984ed846d1fc97a861c12fc7483" Dec 01 15:53:52 crc kubenswrapper[4810]: I1201 15:53:52.501543 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0e41ca0-4806-49a1-bb76-55f140d4ad99" path="/var/lib/kubelet/pods/d0e41ca0-4806-49a1-bb76-55f140d4ad99/volumes" Dec 01 15:54:04 crc kubenswrapper[4810]: I1201 15:54:04.501502 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:54:04 crc kubenswrapper[4810]: E1201 15:54:04.502191 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:54:16 crc kubenswrapper[4810]: I1201 15:54:16.491142 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:54:16 crc kubenswrapper[4810]: E1201 15:54:16.492045 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:54:30 crc kubenswrapper[4810]: I1201 15:54:30.491108 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:54:30 crc kubenswrapper[4810]: E1201 15:54:30.493094 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:54:45 crc kubenswrapper[4810]: I1201 15:54:45.491053 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:54:45 crc kubenswrapper[4810]: E1201 15:54:45.491821 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:54:56 crc kubenswrapper[4810]: I1201 15:54:56.490871 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:54:56 crc kubenswrapper[4810]: E1201 15:54:56.491608 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 15:55:10 crc kubenswrapper[4810]: I1201 15:55:10.491280 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:55:11 crc kubenswrapper[4810]: I1201 15:55:11.071358 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"27dffee837d69d3c7e4cdc8a30f709eb2082d61dd91e53fa5540a464840c57c8"} Dec 01 15:57:28 crc kubenswrapper[4810]: I1201 15:57:28.844162 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hfkkd"] Dec 01 15:57:28 crc kubenswrapper[4810]: E1201 15:57:28.846634 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0e41ca0-4806-49a1-bb76-55f140d4ad99" containerName="extract-utilities" Dec 01 15:57:28 crc kubenswrapper[4810]: I1201 15:57:28.846795 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0e41ca0-4806-49a1-bb76-55f140d4ad99" containerName="extract-utilities" Dec 01 15:57:28 crc kubenswrapper[4810]: E1201 15:57:28.847040 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0e41ca0-4806-49a1-bb76-55f140d4ad99" containerName="registry-server" Dec 01 15:57:28 crc kubenswrapper[4810]: I1201 15:57:28.847225 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0e41ca0-4806-49a1-bb76-55f140d4ad99" containerName="registry-server" Dec 01 15:57:28 crc kubenswrapper[4810]: E1201 15:57:28.847360 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0e41ca0-4806-49a1-bb76-55f140d4ad99" containerName="extract-content" Dec 01 15:57:28 crc kubenswrapper[4810]: I1201 15:57:28.847440 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0e41ca0-4806-49a1-bb76-55f140d4ad99" containerName="extract-content" Dec 01 15:57:28 crc kubenswrapper[4810]: I1201 15:57:28.847761 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0e41ca0-4806-49a1-bb76-55f140d4ad99" containerName="registry-server" Dec 01 15:57:28 crc kubenswrapper[4810]: I1201 15:57:28.849345 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hfkkd" Dec 01 15:57:28 crc kubenswrapper[4810]: I1201 15:57:28.857654 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hfkkd"] Dec 01 15:57:28 crc kubenswrapper[4810]: I1201 15:57:28.998048 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63ae6528-1e15-4444-b388-ec026ac078b0-catalog-content\") pod \"community-operators-hfkkd\" (UID: \"63ae6528-1e15-4444-b388-ec026ac078b0\") " pod="openshift-marketplace/community-operators-hfkkd" Dec 01 15:57:28 crc kubenswrapper[4810]: I1201 15:57:28.998338 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97g4w\" (UniqueName: \"kubernetes.io/projected/63ae6528-1e15-4444-b388-ec026ac078b0-kube-api-access-97g4w\") pod \"community-operators-hfkkd\" (UID: \"63ae6528-1e15-4444-b388-ec026ac078b0\") " pod="openshift-marketplace/community-operators-hfkkd" Dec 01 15:57:28 crc kubenswrapper[4810]: I1201 15:57:28.998527 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63ae6528-1e15-4444-b388-ec026ac078b0-utilities\") pod \"community-operators-hfkkd\" (UID: \"63ae6528-1e15-4444-b388-ec026ac078b0\") " pod="openshift-marketplace/community-operators-hfkkd" Dec 01 15:57:29 crc kubenswrapper[4810]: I1201 15:57:29.099564 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97g4w\" (UniqueName: \"kubernetes.io/projected/63ae6528-1e15-4444-b388-ec026ac078b0-kube-api-access-97g4w\") pod \"community-operators-hfkkd\" (UID: \"63ae6528-1e15-4444-b388-ec026ac078b0\") " pod="openshift-marketplace/community-operators-hfkkd" Dec 01 15:57:29 crc kubenswrapper[4810]: I1201 15:57:29.099656 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63ae6528-1e15-4444-b388-ec026ac078b0-utilities\") pod \"community-operators-hfkkd\" (UID: \"63ae6528-1e15-4444-b388-ec026ac078b0\") " pod="openshift-marketplace/community-operators-hfkkd" Dec 01 15:57:29 crc kubenswrapper[4810]: I1201 15:57:29.099738 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63ae6528-1e15-4444-b388-ec026ac078b0-catalog-content\") pod \"community-operators-hfkkd\" (UID: \"63ae6528-1e15-4444-b388-ec026ac078b0\") " pod="openshift-marketplace/community-operators-hfkkd" Dec 01 15:57:29 crc kubenswrapper[4810]: I1201 15:57:29.100302 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63ae6528-1e15-4444-b388-ec026ac078b0-catalog-content\") pod \"community-operators-hfkkd\" (UID: \"63ae6528-1e15-4444-b388-ec026ac078b0\") " pod="openshift-marketplace/community-operators-hfkkd" Dec 01 15:57:29 crc kubenswrapper[4810]: I1201 15:57:29.100303 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63ae6528-1e15-4444-b388-ec026ac078b0-utilities\") pod \"community-operators-hfkkd\" (UID: \"63ae6528-1e15-4444-b388-ec026ac078b0\") " pod="openshift-marketplace/community-operators-hfkkd" Dec 01 15:57:29 crc kubenswrapper[4810]: I1201 15:57:29.127659 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97g4w\" (UniqueName: \"kubernetes.io/projected/63ae6528-1e15-4444-b388-ec026ac078b0-kube-api-access-97g4w\") pod \"community-operators-hfkkd\" (UID: \"63ae6528-1e15-4444-b388-ec026ac078b0\") " pod="openshift-marketplace/community-operators-hfkkd" Dec 01 15:57:29 crc kubenswrapper[4810]: I1201 15:57:29.205350 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hfkkd" Dec 01 15:57:29 crc kubenswrapper[4810]: I1201 15:57:29.725715 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hfkkd"] Dec 01 15:57:29 crc kubenswrapper[4810]: W1201 15:57:29.728268 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63ae6528_1e15_4444_b388_ec026ac078b0.slice/crio-ff5adf9e1bbac23bd62db46e7c30530b2c95bf8c9e13b4391c8a4a1a1f0be531 WatchSource:0}: Error finding container ff5adf9e1bbac23bd62db46e7c30530b2c95bf8c9e13b4391c8a4a1a1f0be531: Status 404 returned error can't find the container with id ff5adf9e1bbac23bd62db46e7c30530b2c95bf8c9e13b4391c8a4a1a1f0be531 Dec 01 15:57:30 crc kubenswrapper[4810]: I1201 15:57:30.201717 4810 generic.go:334] "Generic (PLEG): container finished" podID="63ae6528-1e15-4444-b388-ec026ac078b0" containerID="95820f4043dcd3aaeea6d911995b88d7e17eb4039297d6182bfb2770199cec3d" exitCode=0 Dec 01 15:57:30 crc kubenswrapper[4810]: I1201 15:57:30.201833 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfkkd" event={"ID":"63ae6528-1e15-4444-b388-ec026ac078b0","Type":"ContainerDied","Data":"95820f4043dcd3aaeea6d911995b88d7e17eb4039297d6182bfb2770199cec3d"} Dec 01 15:57:30 crc kubenswrapper[4810]: I1201 15:57:30.202153 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfkkd" event={"ID":"63ae6528-1e15-4444-b388-ec026ac078b0","Type":"ContainerStarted","Data":"ff5adf9e1bbac23bd62db46e7c30530b2c95bf8c9e13b4391c8a4a1a1f0be531"} Dec 01 15:57:32 crc kubenswrapper[4810]: I1201 15:57:32.222009 4810 generic.go:334] "Generic (PLEG): container finished" podID="63ae6528-1e15-4444-b388-ec026ac078b0" containerID="79eba24adcc6bb9b767a85b1f15f328ce3d8444006234b8986efc8858bba697c" exitCode=0 Dec 01 15:57:32 crc kubenswrapper[4810]: I1201 15:57:32.222052 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfkkd" event={"ID":"63ae6528-1e15-4444-b388-ec026ac078b0","Type":"ContainerDied","Data":"79eba24adcc6bb9b767a85b1f15f328ce3d8444006234b8986efc8858bba697c"} Dec 01 15:57:32 crc kubenswrapper[4810]: I1201 15:57:32.972659 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:57:32 crc kubenswrapper[4810]: I1201 15:57:32.972711 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:57:33 crc kubenswrapper[4810]: I1201 15:57:33.232128 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfkkd" event={"ID":"63ae6528-1e15-4444-b388-ec026ac078b0","Type":"ContainerStarted","Data":"e089b1b7e4d036e607110a636a89ac4c4a8eb584744ba595901a0d1d50dddfce"} Dec 01 15:57:33 crc kubenswrapper[4810]: I1201 15:57:33.254128 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hfkkd" podStartSLOduration=2.493904894 podStartE2EDuration="5.254107362s" podCreationTimestamp="2025-12-01 15:57:28 +0000 UTC" firstStartedPulling="2025-12-01 15:57:30.204102379 +0000 UTC m=+5015.967611982" lastFinishedPulling="2025-12-01 15:57:32.964304847 +0000 UTC m=+5018.727814450" observedRunningTime="2025-12-01 15:57:33.252030125 +0000 UTC m=+5019.015539758" watchObservedRunningTime="2025-12-01 15:57:33.254107362 +0000 UTC m=+5019.017616965" Dec 01 15:57:36 crc kubenswrapper[4810]: I1201 15:57:36.609221 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-ss5nn"] Dec 01 15:57:36 crc kubenswrapper[4810]: I1201 15:57:36.614592 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-ss5nn"] Dec 01 15:57:36 crc kubenswrapper[4810]: I1201 15:57:36.753289 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-9tdhp"] Dec 01 15:57:36 crc kubenswrapper[4810]: I1201 15:57:36.754374 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-9tdhp" Dec 01 15:57:36 crc kubenswrapper[4810]: I1201 15:57:36.757392 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 01 15:57:36 crc kubenswrapper[4810]: I1201 15:57:36.757424 4810 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-zgjj8" Dec 01 15:57:36 crc kubenswrapper[4810]: I1201 15:57:36.757436 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 01 15:57:36 crc kubenswrapper[4810]: I1201 15:57:36.757544 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 01 15:57:36 crc kubenswrapper[4810]: I1201 15:57:36.766679 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-9tdhp"] Dec 01 15:57:36 crc kubenswrapper[4810]: I1201 15:57:36.803529 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdm7d\" (UniqueName: \"kubernetes.io/projected/2db2df1b-6628-4b6e-8b03-b51bba1e6825-kube-api-access-jdm7d\") pod \"crc-storage-crc-9tdhp\" (UID: \"2db2df1b-6628-4b6e-8b03-b51bba1e6825\") " pod="crc-storage/crc-storage-crc-9tdhp" Dec 01 15:57:36 crc kubenswrapper[4810]: I1201 15:57:36.803682 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/2db2df1b-6628-4b6e-8b03-b51bba1e6825-crc-storage\") pod \"crc-storage-crc-9tdhp\" (UID: \"2db2df1b-6628-4b6e-8b03-b51bba1e6825\") " pod="crc-storage/crc-storage-crc-9tdhp" Dec 01 15:57:36 crc kubenswrapper[4810]: I1201 15:57:36.803774 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/2db2df1b-6628-4b6e-8b03-b51bba1e6825-node-mnt\") pod \"crc-storage-crc-9tdhp\" (UID: \"2db2df1b-6628-4b6e-8b03-b51bba1e6825\") " pod="crc-storage/crc-storage-crc-9tdhp" Dec 01 15:57:36 crc kubenswrapper[4810]: I1201 15:57:36.904598 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/2db2df1b-6628-4b6e-8b03-b51bba1e6825-crc-storage\") pod \"crc-storage-crc-9tdhp\" (UID: \"2db2df1b-6628-4b6e-8b03-b51bba1e6825\") " pod="crc-storage/crc-storage-crc-9tdhp" Dec 01 15:57:36 crc kubenswrapper[4810]: I1201 15:57:36.904692 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/2db2df1b-6628-4b6e-8b03-b51bba1e6825-node-mnt\") pod \"crc-storage-crc-9tdhp\" (UID: \"2db2df1b-6628-4b6e-8b03-b51bba1e6825\") " pod="crc-storage/crc-storage-crc-9tdhp" Dec 01 15:57:36 crc kubenswrapper[4810]: I1201 15:57:36.904722 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdm7d\" (UniqueName: \"kubernetes.io/projected/2db2df1b-6628-4b6e-8b03-b51bba1e6825-kube-api-access-jdm7d\") pod \"crc-storage-crc-9tdhp\" (UID: \"2db2df1b-6628-4b6e-8b03-b51bba1e6825\") " pod="crc-storage/crc-storage-crc-9tdhp" Dec 01 15:57:36 crc kubenswrapper[4810]: I1201 15:57:36.905375 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/2db2df1b-6628-4b6e-8b03-b51bba1e6825-node-mnt\") pod \"crc-storage-crc-9tdhp\" (UID: \"2db2df1b-6628-4b6e-8b03-b51bba1e6825\") " pod="crc-storage/crc-storage-crc-9tdhp" Dec 01 15:57:36 crc kubenswrapper[4810]: I1201 15:57:36.905606 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/2db2df1b-6628-4b6e-8b03-b51bba1e6825-crc-storage\") pod \"crc-storage-crc-9tdhp\" (UID: \"2db2df1b-6628-4b6e-8b03-b51bba1e6825\") " pod="crc-storage/crc-storage-crc-9tdhp" Dec 01 15:57:36 crc kubenswrapper[4810]: I1201 15:57:36.923212 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdm7d\" (UniqueName: \"kubernetes.io/projected/2db2df1b-6628-4b6e-8b03-b51bba1e6825-kube-api-access-jdm7d\") pod \"crc-storage-crc-9tdhp\" (UID: \"2db2df1b-6628-4b6e-8b03-b51bba1e6825\") " pod="crc-storage/crc-storage-crc-9tdhp" Dec 01 15:57:37 crc kubenswrapper[4810]: I1201 15:57:37.077194 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-9tdhp" Dec 01 15:57:37 crc kubenswrapper[4810]: I1201 15:57:37.489897 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-9tdhp"] Dec 01 15:57:37 crc kubenswrapper[4810]: W1201 15:57:37.496050 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2db2df1b_6628_4b6e_8b03_b51bba1e6825.slice/crio-5a52c14e7cb800f3f88f3ecd2cfd5ad1c2f83ad556a84bc87cb3c5e64490ccdb WatchSource:0}: Error finding container 5a52c14e7cb800f3f88f3ecd2cfd5ad1c2f83ad556a84bc87cb3c5e64490ccdb: Status 404 returned error can't find the container with id 5a52c14e7cb800f3f88f3ecd2cfd5ad1c2f83ad556a84bc87cb3c5e64490ccdb Dec 01 15:57:38 crc kubenswrapper[4810]: I1201 15:57:38.279921 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-9tdhp" event={"ID":"2db2df1b-6628-4b6e-8b03-b51bba1e6825","Type":"ContainerStarted","Data":"5a52c14e7cb800f3f88f3ecd2cfd5ad1c2f83ad556a84bc87cb3c5e64490ccdb"} Dec 01 15:57:38 crc kubenswrapper[4810]: I1201 15:57:38.552374 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a67e6d2-3f65-4e27-aefe-27f04f7ca75e" path="/var/lib/kubelet/pods/0a67e6d2-3f65-4e27-aefe-27f04f7ca75e/volumes" Dec 01 15:57:39 crc kubenswrapper[4810]: I1201 15:57:39.206385 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hfkkd" Dec 01 15:57:39 crc kubenswrapper[4810]: I1201 15:57:39.206835 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hfkkd" Dec 01 15:57:39 crc kubenswrapper[4810]: I1201 15:57:39.257278 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hfkkd" Dec 01 15:57:39 crc kubenswrapper[4810]: I1201 15:57:39.288379 4810 generic.go:334] "Generic (PLEG): container finished" podID="2db2df1b-6628-4b6e-8b03-b51bba1e6825" containerID="79a31301bf0b55903f9ccb9d1d6ebee23f756fedc97f72fa4043821de323c517" exitCode=0 Dec 01 15:57:39 crc kubenswrapper[4810]: I1201 15:57:39.288446 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-9tdhp" event={"ID":"2db2df1b-6628-4b6e-8b03-b51bba1e6825","Type":"ContainerDied","Data":"79a31301bf0b55903f9ccb9d1d6ebee23f756fedc97f72fa4043821de323c517"} Dec 01 15:57:39 crc kubenswrapper[4810]: I1201 15:57:39.338849 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hfkkd" Dec 01 15:57:39 crc kubenswrapper[4810]: I1201 15:57:39.489434 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hfkkd"] Dec 01 15:57:40 crc kubenswrapper[4810]: I1201 15:57:40.580358 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-9tdhp" Dec 01 15:57:40 crc kubenswrapper[4810]: I1201 15:57:40.756065 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/2db2df1b-6628-4b6e-8b03-b51bba1e6825-node-mnt\") pod \"2db2df1b-6628-4b6e-8b03-b51bba1e6825\" (UID: \"2db2df1b-6628-4b6e-8b03-b51bba1e6825\") " Dec 01 15:57:40 crc kubenswrapper[4810]: I1201 15:57:40.756189 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2db2df1b-6628-4b6e-8b03-b51bba1e6825-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "2db2df1b-6628-4b6e-8b03-b51bba1e6825" (UID: "2db2df1b-6628-4b6e-8b03-b51bba1e6825"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:57:40 crc kubenswrapper[4810]: I1201 15:57:40.756381 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdm7d\" (UniqueName: \"kubernetes.io/projected/2db2df1b-6628-4b6e-8b03-b51bba1e6825-kube-api-access-jdm7d\") pod \"2db2df1b-6628-4b6e-8b03-b51bba1e6825\" (UID: \"2db2df1b-6628-4b6e-8b03-b51bba1e6825\") " Dec 01 15:57:40 crc kubenswrapper[4810]: I1201 15:57:40.756520 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/2db2df1b-6628-4b6e-8b03-b51bba1e6825-crc-storage\") pod \"2db2df1b-6628-4b6e-8b03-b51bba1e6825\" (UID: \"2db2df1b-6628-4b6e-8b03-b51bba1e6825\") " Dec 01 15:57:40 crc kubenswrapper[4810]: I1201 15:57:40.757109 4810 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/2db2df1b-6628-4b6e-8b03-b51bba1e6825-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 01 15:57:40 crc kubenswrapper[4810]: I1201 15:57:40.761311 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2db2df1b-6628-4b6e-8b03-b51bba1e6825-kube-api-access-jdm7d" (OuterVolumeSpecName: "kube-api-access-jdm7d") pod "2db2df1b-6628-4b6e-8b03-b51bba1e6825" (UID: "2db2df1b-6628-4b6e-8b03-b51bba1e6825"). InnerVolumeSpecName "kube-api-access-jdm7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:57:40 crc kubenswrapper[4810]: I1201 15:57:40.773880 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2db2df1b-6628-4b6e-8b03-b51bba1e6825-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "2db2df1b-6628-4b6e-8b03-b51bba1e6825" (UID: "2db2df1b-6628-4b6e-8b03-b51bba1e6825"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:57:40 crc kubenswrapper[4810]: I1201 15:57:40.858022 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdm7d\" (UniqueName: \"kubernetes.io/projected/2db2df1b-6628-4b6e-8b03-b51bba1e6825-kube-api-access-jdm7d\") on node \"crc\" DevicePath \"\"" Dec 01 15:57:40 crc kubenswrapper[4810]: I1201 15:57:40.858057 4810 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/2db2df1b-6628-4b6e-8b03-b51bba1e6825-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 01 15:57:41 crc kubenswrapper[4810]: I1201 15:57:41.305298 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-9tdhp" Dec 01 15:57:41 crc kubenswrapper[4810]: I1201 15:57:41.305319 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-9tdhp" event={"ID":"2db2df1b-6628-4b6e-8b03-b51bba1e6825","Type":"ContainerDied","Data":"5a52c14e7cb800f3f88f3ecd2cfd5ad1c2f83ad556a84bc87cb3c5e64490ccdb"} Dec 01 15:57:41 crc kubenswrapper[4810]: I1201 15:57:41.305403 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hfkkd" podUID="63ae6528-1e15-4444-b388-ec026ac078b0" containerName="registry-server" containerID="cri-o://e089b1b7e4d036e607110a636a89ac4c4a8eb584744ba595901a0d1d50dddfce" gracePeriod=2 Dec 01 15:57:41 crc kubenswrapper[4810]: I1201 15:57:41.305421 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a52c14e7cb800f3f88f3ecd2cfd5ad1c2f83ad556a84bc87cb3c5e64490ccdb" Dec 01 15:57:41 crc kubenswrapper[4810]: I1201 15:57:41.676569 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hfkkd" Dec 01 15:57:41 crc kubenswrapper[4810]: I1201 15:57:41.872951 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63ae6528-1e15-4444-b388-ec026ac078b0-utilities\") pod \"63ae6528-1e15-4444-b388-ec026ac078b0\" (UID: \"63ae6528-1e15-4444-b388-ec026ac078b0\") " Dec 01 15:57:41 crc kubenswrapper[4810]: I1201 15:57:41.873091 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97g4w\" (UniqueName: \"kubernetes.io/projected/63ae6528-1e15-4444-b388-ec026ac078b0-kube-api-access-97g4w\") pod \"63ae6528-1e15-4444-b388-ec026ac078b0\" (UID: \"63ae6528-1e15-4444-b388-ec026ac078b0\") " Dec 01 15:57:41 crc kubenswrapper[4810]: I1201 15:57:41.873174 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63ae6528-1e15-4444-b388-ec026ac078b0-catalog-content\") pod \"63ae6528-1e15-4444-b388-ec026ac078b0\" (UID: \"63ae6528-1e15-4444-b388-ec026ac078b0\") " Dec 01 15:57:41 crc kubenswrapper[4810]: I1201 15:57:41.874080 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63ae6528-1e15-4444-b388-ec026ac078b0-utilities" (OuterVolumeSpecName: "utilities") pod "63ae6528-1e15-4444-b388-ec026ac078b0" (UID: "63ae6528-1e15-4444-b388-ec026ac078b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:57:41 crc kubenswrapper[4810]: I1201 15:57:41.880206 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63ae6528-1e15-4444-b388-ec026ac078b0-kube-api-access-97g4w" (OuterVolumeSpecName: "kube-api-access-97g4w") pod "63ae6528-1e15-4444-b388-ec026ac078b0" (UID: "63ae6528-1e15-4444-b388-ec026ac078b0"). InnerVolumeSpecName "kube-api-access-97g4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:57:41 crc kubenswrapper[4810]: I1201 15:57:41.963537 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63ae6528-1e15-4444-b388-ec026ac078b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63ae6528-1e15-4444-b388-ec026ac078b0" (UID: "63ae6528-1e15-4444-b388-ec026ac078b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:57:41 crc kubenswrapper[4810]: I1201 15:57:41.975053 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63ae6528-1e15-4444-b388-ec026ac078b0-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:57:41 crc kubenswrapper[4810]: I1201 15:57:41.975098 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97g4w\" (UniqueName: \"kubernetes.io/projected/63ae6528-1e15-4444-b388-ec026ac078b0-kube-api-access-97g4w\") on node \"crc\" DevicePath \"\"" Dec 01 15:57:41 crc kubenswrapper[4810]: I1201 15:57:41.975109 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63ae6528-1e15-4444-b388-ec026ac078b0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.313861 4810 generic.go:334] "Generic (PLEG): container finished" podID="63ae6528-1e15-4444-b388-ec026ac078b0" containerID="e089b1b7e4d036e607110a636a89ac4c4a8eb584744ba595901a0d1d50dddfce" exitCode=0 Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.313905 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfkkd" event={"ID":"63ae6528-1e15-4444-b388-ec026ac078b0","Type":"ContainerDied","Data":"e089b1b7e4d036e607110a636a89ac4c4a8eb584744ba595901a0d1d50dddfce"} Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.313929 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hfkkd" event={"ID":"63ae6528-1e15-4444-b388-ec026ac078b0","Type":"ContainerDied","Data":"ff5adf9e1bbac23bd62db46e7c30530b2c95bf8c9e13b4391c8a4a1a1f0be531"} Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.313927 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hfkkd" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.314013 4810 scope.go:117] "RemoveContainer" containerID="e089b1b7e4d036e607110a636a89ac4c4a8eb584744ba595901a0d1d50dddfce" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.330785 4810 scope.go:117] "RemoveContainer" containerID="79eba24adcc6bb9b767a85b1f15f328ce3d8444006234b8986efc8858bba697c" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.353223 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hfkkd"] Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.365263 4810 scope.go:117] "RemoveContainer" containerID="95820f4043dcd3aaeea6d911995b88d7e17eb4039297d6182bfb2770199cec3d" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.368188 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hfkkd"] Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.384805 4810 scope.go:117] "RemoveContainer" containerID="e089b1b7e4d036e607110a636a89ac4c4a8eb584744ba595901a0d1d50dddfce" Dec 01 15:57:42 crc kubenswrapper[4810]: E1201 15:57:42.385414 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e089b1b7e4d036e607110a636a89ac4c4a8eb584744ba595901a0d1d50dddfce\": container with ID starting with e089b1b7e4d036e607110a636a89ac4c4a8eb584744ba595901a0d1d50dddfce not found: ID does not exist" containerID="e089b1b7e4d036e607110a636a89ac4c4a8eb584744ba595901a0d1d50dddfce" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.385463 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e089b1b7e4d036e607110a636a89ac4c4a8eb584744ba595901a0d1d50dddfce"} err="failed to get container status \"e089b1b7e4d036e607110a636a89ac4c4a8eb584744ba595901a0d1d50dddfce\": rpc error: code = NotFound desc = could not find container \"e089b1b7e4d036e607110a636a89ac4c4a8eb584744ba595901a0d1d50dddfce\": container with ID starting with e089b1b7e4d036e607110a636a89ac4c4a8eb584744ba595901a0d1d50dddfce not found: ID does not exist" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.385512 4810 scope.go:117] "RemoveContainer" containerID="79eba24adcc6bb9b767a85b1f15f328ce3d8444006234b8986efc8858bba697c" Dec 01 15:57:42 crc kubenswrapper[4810]: E1201 15:57:42.385924 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79eba24adcc6bb9b767a85b1f15f328ce3d8444006234b8986efc8858bba697c\": container with ID starting with 79eba24adcc6bb9b767a85b1f15f328ce3d8444006234b8986efc8858bba697c not found: ID does not exist" containerID="79eba24adcc6bb9b767a85b1f15f328ce3d8444006234b8986efc8858bba697c" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.386004 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79eba24adcc6bb9b767a85b1f15f328ce3d8444006234b8986efc8858bba697c"} err="failed to get container status \"79eba24adcc6bb9b767a85b1f15f328ce3d8444006234b8986efc8858bba697c\": rpc error: code = NotFound desc = could not find container \"79eba24adcc6bb9b767a85b1f15f328ce3d8444006234b8986efc8858bba697c\": container with ID starting with 79eba24adcc6bb9b767a85b1f15f328ce3d8444006234b8986efc8858bba697c not found: ID does not exist" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.386072 4810 scope.go:117] "RemoveContainer" containerID="95820f4043dcd3aaeea6d911995b88d7e17eb4039297d6182bfb2770199cec3d" Dec 01 15:57:42 crc kubenswrapper[4810]: E1201 15:57:42.386466 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95820f4043dcd3aaeea6d911995b88d7e17eb4039297d6182bfb2770199cec3d\": container with ID starting with 95820f4043dcd3aaeea6d911995b88d7e17eb4039297d6182bfb2770199cec3d not found: ID does not exist" containerID="95820f4043dcd3aaeea6d911995b88d7e17eb4039297d6182bfb2770199cec3d" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.386579 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95820f4043dcd3aaeea6d911995b88d7e17eb4039297d6182bfb2770199cec3d"} err="failed to get container status \"95820f4043dcd3aaeea6d911995b88d7e17eb4039297d6182bfb2770199cec3d\": rpc error: code = NotFound desc = could not find container \"95820f4043dcd3aaeea6d911995b88d7e17eb4039297d6182bfb2770199cec3d\": container with ID starting with 95820f4043dcd3aaeea6d911995b88d7e17eb4039297d6182bfb2770199cec3d not found: ID does not exist" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.499164 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63ae6528-1e15-4444-b388-ec026ac078b0" path="/var/lib/kubelet/pods/63ae6528-1e15-4444-b388-ec026ac078b0/volumes" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.815087 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-9tdhp"] Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.820161 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-9tdhp"] Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.948740 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-8gcfp"] Dec 01 15:57:42 crc kubenswrapper[4810]: E1201 15:57:42.949094 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2db2df1b-6628-4b6e-8b03-b51bba1e6825" containerName="storage" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.949111 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2db2df1b-6628-4b6e-8b03-b51bba1e6825" containerName="storage" Dec 01 15:57:42 crc kubenswrapper[4810]: E1201 15:57:42.949132 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ae6528-1e15-4444-b388-ec026ac078b0" containerName="registry-server" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.949140 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ae6528-1e15-4444-b388-ec026ac078b0" containerName="registry-server" Dec 01 15:57:42 crc kubenswrapper[4810]: E1201 15:57:42.949156 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ae6528-1e15-4444-b388-ec026ac078b0" containerName="extract-utilities" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.949165 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ae6528-1e15-4444-b388-ec026ac078b0" containerName="extract-utilities" Dec 01 15:57:42 crc kubenswrapper[4810]: E1201 15:57:42.949180 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ae6528-1e15-4444-b388-ec026ac078b0" containerName="extract-content" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.949189 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ae6528-1e15-4444-b388-ec026ac078b0" containerName="extract-content" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.949351 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="2db2df1b-6628-4b6e-8b03-b51bba1e6825" containerName="storage" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.949378 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ae6528-1e15-4444-b388-ec026ac078b0" containerName="registry-server" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.950134 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8gcfp" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.957668 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.957736 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.957858 4810 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-zgjj8" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.957939 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 01 15:57:42 crc kubenswrapper[4810]: I1201 15:57:42.964525 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-8gcfp"] Dec 01 15:57:43 crc kubenswrapper[4810]: I1201 15:57:43.090282 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5s9x\" (UniqueName: \"kubernetes.io/projected/f9a386e2-f7bb-4d6d-ad84-b344243deaee-kube-api-access-m5s9x\") pod \"crc-storage-crc-8gcfp\" (UID: \"f9a386e2-f7bb-4d6d-ad84-b344243deaee\") " pod="crc-storage/crc-storage-crc-8gcfp" Dec 01 15:57:43 crc kubenswrapper[4810]: I1201 15:57:43.090334 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f9a386e2-f7bb-4d6d-ad84-b344243deaee-node-mnt\") pod \"crc-storage-crc-8gcfp\" (UID: \"f9a386e2-f7bb-4d6d-ad84-b344243deaee\") " pod="crc-storage/crc-storage-crc-8gcfp" Dec 01 15:57:43 crc kubenswrapper[4810]: I1201 15:57:43.090397 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f9a386e2-f7bb-4d6d-ad84-b344243deaee-crc-storage\") pod \"crc-storage-crc-8gcfp\" (UID: \"f9a386e2-f7bb-4d6d-ad84-b344243deaee\") " pod="crc-storage/crc-storage-crc-8gcfp" Dec 01 15:57:43 crc kubenswrapper[4810]: I1201 15:57:43.192042 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5s9x\" (UniqueName: \"kubernetes.io/projected/f9a386e2-f7bb-4d6d-ad84-b344243deaee-kube-api-access-m5s9x\") pod \"crc-storage-crc-8gcfp\" (UID: \"f9a386e2-f7bb-4d6d-ad84-b344243deaee\") " pod="crc-storage/crc-storage-crc-8gcfp" Dec 01 15:57:43 crc kubenswrapper[4810]: I1201 15:57:43.192093 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f9a386e2-f7bb-4d6d-ad84-b344243deaee-node-mnt\") pod \"crc-storage-crc-8gcfp\" (UID: \"f9a386e2-f7bb-4d6d-ad84-b344243deaee\") " pod="crc-storage/crc-storage-crc-8gcfp" Dec 01 15:57:43 crc kubenswrapper[4810]: I1201 15:57:43.192153 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f9a386e2-f7bb-4d6d-ad84-b344243deaee-crc-storage\") pod \"crc-storage-crc-8gcfp\" (UID: \"f9a386e2-f7bb-4d6d-ad84-b344243deaee\") " pod="crc-storage/crc-storage-crc-8gcfp" Dec 01 15:57:43 crc kubenswrapper[4810]: I1201 15:57:43.192504 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f9a386e2-f7bb-4d6d-ad84-b344243deaee-node-mnt\") pod \"crc-storage-crc-8gcfp\" (UID: \"f9a386e2-f7bb-4d6d-ad84-b344243deaee\") " pod="crc-storage/crc-storage-crc-8gcfp" Dec 01 15:57:43 crc kubenswrapper[4810]: I1201 15:57:43.192893 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f9a386e2-f7bb-4d6d-ad84-b344243deaee-crc-storage\") pod \"crc-storage-crc-8gcfp\" (UID: \"f9a386e2-f7bb-4d6d-ad84-b344243deaee\") " pod="crc-storage/crc-storage-crc-8gcfp" Dec 01 15:57:43 crc kubenswrapper[4810]: I1201 15:57:43.209268 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5s9x\" (UniqueName: \"kubernetes.io/projected/f9a386e2-f7bb-4d6d-ad84-b344243deaee-kube-api-access-m5s9x\") pod \"crc-storage-crc-8gcfp\" (UID: \"f9a386e2-f7bb-4d6d-ad84-b344243deaee\") " pod="crc-storage/crc-storage-crc-8gcfp" Dec 01 15:57:43 crc kubenswrapper[4810]: I1201 15:57:43.275138 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8gcfp" Dec 01 15:57:43 crc kubenswrapper[4810]: I1201 15:57:43.676124 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-8gcfp"] Dec 01 15:57:44 crc kubenswrapper[4810]: I1201 15:57:44.327312 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-8gcfp" event={"ID":"f9a386e2-f7bb-4d6d-ad84-b344243deaee","Type":"ContainerStarted","Data":"849da8c223d9386f672f5ea34f66e76192ee1a62e0f2baf273caffb5864ebe44"} Dec 01 15:57:44 crc kubenswrapper[4810]: I1201 15:57:44.499051 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2db2df1b-6628-4b6e-8b03-b51bba1e6825" path="/var/lib/kubelet/pods/2db2df1b-6628-4b6e-8b03-b51bba1e6825/volumes" Dec 01 15:57:45 crc kubenswrapper[4810]: I1201 15:57:45.346457 4810 generic.go:334] "Generic (PLEG): container finished" podID="f9a386e2-f7bb-4d6d-ad84-b344243deaee" containerID="52b6f49c1c5d60235095e06de883707c8aeeb48efc427be522f661b96d06e988" exitCode=0 Dec 01 15:57:45 crc kubenswrapper[4810]: I1201 15:57:45.346929 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-8gcfp" event={"ID":"f9a386e2-f7bb-4d6d-ad84-b344243deaee","Type":"ContainerDied","Data":"52b6f49c1c5d60235095e06de883707c8aeeb48efc427be522f661b96d06e988"} Dec 01 15:57:46 crc kubenswrapper[4810]: I1201 15:57:46.665989 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8gcfp" Dec 01 15:57:46 crc kubenswrapper[4810]: I1201 15:57:46.843510 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f9a386e2-f7bb-4d6d-ad84-b344243deaee-crc-storage\") pod \"f9a386e2-f7bb-4d6d-ad84-b344243deaee\" (UID: \"f9a386e2-f7bb-4d6d-ad84-b344243deaee\") " Dec 01 15:57:46 crc kubenswrapper[4810]: I1201 15:57:46.843612 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5s9x\" (UniqueName: \"kubernetes.io/projected/f9a386e2-f7bb-4d6d-ad84-b344243deaee-kube-api-access-m5s9x\") pod \"f9a386e2-f7bb-4d6d-ad84-b344243deaee\" (UID: \"f9a386e2-f7bb-4d6d-ad84-b344243deaee\") " Dec 01 15:57:46 crc kubenswrapper[4810]: I1201 15:57:46.843705 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f9a386e2-f7bb-4d6d-ad84-b344243deaee-node-mnt\") pod \"f9a386e2-f7bb-4d6d-ad84-b344243deaee\" (UID: \"f9a386e2-f7bb-4d6d-ad84-b344243deaee\") " Dec 01 15:57:46 crc kubenswrapper[4810]: I1201 15:57:46.843919 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9a386e2-f7bb-4d6d-ad84-b344243deaee-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "f9a386e2-f7bb-4d6d-ad84-b344243deaee" (UID: "f9a386e2-f7bb-4d6d-ad84-b344243deaee"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 15:57:46 crc kubenswrapper[4810]: I1201 15:57:46.844406 4810 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f9a386e2-f7bb-4d6d-ad84-b344243deaee-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 01 15:57:46 crc kubenswrapper[4810]: I1201 15:57:46.848703 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9a386e2-f7bb-4d6d-ad84-b344243deaee-kube-api-access-m5s9x" (OuterVolumeSpecName: "kube-api-access-m5s9x") pod "f9a386e2-f7bb-4d6d-ad84-b344243deaee" (UID: "f9a386e2-f7bb-4d6d-ad84-b344243deaee"). InnerVolumeSpecName "kube-api-access-m5s9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:57:46 crc kubenswrapper[4810]: I1201 15:57:46.862231 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9a386e2-f7bb-4d6d-ad84-b344243deaee-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "f9a386e2-f7bb-4d6d-ad84-b344243deaee" (UID: "f9a386e2-f7bb-4d6d-ad84-b344243deaee"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:57:46 crc kubenswrapper[4810]: I1201 15:57:46.945851 4810 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f9a386e2-f7bb-4d6d-ad84-b344243deaee-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 01 15:57:46 crc kubenswrapper[4810]: I1201 15:57:46.945894 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5s9x\" (UniqueName: \"kubernetes.io/projected/f9a386e2-f7bb-4d6d-ad84-b344243deaee-kube-api-access-m5s9x\") on node \"crc\" DevicePath \"\"" Dec 01 15:57:47 crc kubenswrapper[4810]: I1201 15:57:47.365692 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-8gcfp" event={"ID":"f9a386e2-f7bb-4d6d-ad84-b344243deaee","Type":"ContainerDied","Data":"849da8c223d9386f672f5ea34f66e76192ee1a62e0f2baf273caffb5864ebe44"} Dec 01 15:57:47 crc kubenswrapper[4810]: I1201 15:57:47.365763 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-8gcfp" Dec 01 15:57:47 crc kubenswrapper[4810]: I1201 15:57:47.365789 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="849da8c223d9386f672f5ea34f66e76192ee1a62e0f2baf273caffb5864ebe44" Dec 01 15:58:02 crc kubenswrapper[4810]: I1201 15:58:02.971760 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:58:02 crc kubenswrapper[4810]: I1201 15:58:02.972257 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:58:08 crc kubenswrapper[4810]: I1201 15:58:08.655851 4810 scope.go:117] "RemoveContainer" containerID="66eed9ca83c9c0f52c6ec00f5d3a71b9b0613bb7ef079010557011bf014c6898" Dec 01 15:58:32 crc kubenswrapper[4810]: I1201 15:58:32.972610 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 15:58:32 crc kubenswrapper[4810]: I1201 15:58:32.973210 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 15:58:32 crc kubenswrapper[4810]: I1201 15:58:32.973266 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 15:58:32 crc kubenswrapper[4810]: I1201 15:58:32.973977 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"27dffee837d69d3c7e4cdc8a30f709eb2082d61dd91e53fa5540a464840c57c8"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 15:58:32 crc kubenswrapper[4810]: I1201 15:58:32.974046 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://27dffee837d69d3c7e4cdc8a30f709eb2082d61dd91e53fa5540a464840c57c8" gracePeriod=600 Dec 01 15:58:33 crc kubenswrapper[4810]: I1201 15:58:33.718597 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="27dffee837d69d3c7e4cdc8a30f709eb2082d61dd91e53fa5540a464840c57c8" exitCode=0 Dec 01 15:58:33 crc kubenswrapper[4810]: I1201 15:58:33.718699 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"27dffee837d69d3c7e4cdc8a30f709eb2082d61dd91e53fa5540a464840c57c8"} Dec 01 15:58:33 crc kubenswrapper[4810]: I1201 15:58:33.718953 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706"} Dec 01 15:58:33 crc kubenswrapper[4810]: I1201 15:58:33.718995 4810 scope.go:117] "RemoveContainer" containerID="bfb6114af71c9c3c9a3fea52c77b5258cdbbe9fa314bc43f50ef83f8942ba889" Dec 01 15:58:59 crc kubenswrapper[4810]: I1201 15:58:59.756254 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7h8nl"] Dec 01 15:58:59 crc kubenswrapper[4810]: E1201 15:58:59.757975 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9a386e2-f7bb-4d6d-ad84-b344243deaee" containerName="storage" Dec 01 15:58:59 crc kubenswrapper[4810]: I1201 15:58:59.757993 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9a386e2-f7bb-4d6d-ad84-b344243deaee" containerName="storage" Dec 01 15:58:59 crc kubenswrapper[4810]: I1201 15:58:59.758175 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9a386e2-f7bb-4d6d-ad84-b344243deaee" containerName="storage" Dec 01 15:58:59 crc kubenswrapper[4810]: I1201 15:58:59.759554 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7h8nl" Dec 01 15:58:59 crc kubenswrapper[4810]: I1201 15:58:59.767128 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7h8nl"] Dec 01 15:58:59 crc kubenswrapper[4810]: I1201 15:58:59.807302 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rxxg\" (UniqueName: \"kubernetes.io/projected/0341c875-f1fa-4a99-b850-202a0f3385db-kube-api-access-9rxxg\") pod \"redhat-operators-7h8nl\" (UID: \"0341c875-f1fa-4a99-b850-202a0f3385db\") " pod="openshift-marketplace/redhat-operators-7h8nl" Dec 01 15:58:59 crc kubenswrapper[4810]: I1201 15:58:59.807402 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0341c875-f1fa-4a99-b850-202a0f3385db-catalog-content\") pod \"redhat-operators-7h8nl\" (UID: \"0341c875-f1fa-4a99-b850-202a0f3385db\") " pod="openshift-marketplace/redhat-operators-7h8nl" Dec 01 15:58:59 crc kubenswrapper[4810]: I1201 15:58:59.807719 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0341c875-f1fa-4a99-b850-202a0f3385db-utilities\") pod \"redhat-operators-7h8nl\" (UID: \"0341c875-f1fa-4a99-b850-202a0f3385db\") " pod="openshift-marketplace/redhat-operators-7h8nl" Dec 01 15:58:59 crc kubenswrapper[4810]: I1201 15:58:59.908625 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0341c875-f1fa-4a99-b850-202a0f3385db-utilities\") pod \"redhat-operators-7h8nl\" (UID: \"0341c875-f1fa-4a99-b850-202a0f3385db\") " pod="openshift-marketplace/redhat-operators-7h8nl" Dec 01 15:58:59 crc kubenswrapper[4810]: I1201 15:58:59.908671 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rxxg\" (UniqueName: \"kubernetes.io/projected/0341c875-f1fa-4a99-b850-202a0f3385db-kube-api-access-9rxxg\") pod \"redhat-operators-7h8nl\" (UID: \"0341c875-f1fa-4a99-b850-202a0f3385db\") " pod="openshift-marketplace/redhat-operators-7h8nl" Dec 01 15:58:59 crc kubenswrapper[4810]: I1201 15:58:59.908712 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0341c875-f1fa-4a99-b850-202a0f3385db-catalog-content\") pod \"redhat-operators-7h8nl\" (UID: \"0341c875-f1fa-4a99-b850-202a0f3385db\") " pod="openshift-marketplace/redhat-operators-7h8nl" Dec 01 15:58:59 crc kubenswrapper[4810]: I1201 15:58:59.909104 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0341c875-f1fa-4a99-b850-202a0f3385db-utilities\") pod \"redhat-operators-7h8nl\" (UID: \"0341c875-f1fa-4a99-b850-202a0f3385db\") " pod="openshift-marketplace/redhat-operators-7h8nl" Dec 01 15:58:59 crc kubenswrapper[4810]: I1201 15:58:59.909144 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0341c875-f1fa-4a99-b850-202a0f3385db-catalog-content\") pod \"redhat-operators-7h8nl\" (UID: \"0341c875-f1fa-4a99-b850-202a0f3385db\") " pod="openshift-marketplace/redhat-operators-7h8nl" Dec 01 15:58:59 crc kubenswrapper[4810]: I1201 15:58:59.930438 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rxxg\" (UniqueName: \"kubernetes.io/projected/0341c875-f1fa-4a99-b850-202a0f3385db-kube-api-access-9rxxg\") pod \"redhat-operators-7h8nl\" (UID: \"0341c875-f1fa-4a99-b850-202a0f3385db\") " pod="openshift-marketplace/redhat-operators-7h8nl" Dec 01 15:59:00 crc kubenswrapper[4810]: I1201 15:59:00.120908 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7h8nl" Dec 01 15:59:00 crc kubenswrapper[4810]: I1201 15:59:00.381706 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7h8nl"] Dec 01 15:59:00 crc kubenswrapper[4810]: W1201 15:59:00.384069 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0341c875_f1fa_4a99_b850_202a0f3385db.slice/crio-c356d16feda731d4581c9a07a174774c0b026c7287d988f0f5581cd090f02ae7 WatchSource:0}: Error finding container c356d16feda731d4581c9a07a174774c0b026c7287d988f0f5581cd090f02ae7: Status 404 returned error can't find the container with id c356d16feda731d4581c9a07a174774c0b026c7287d988f0f5581cd090f02ae7 Dec 01 15:59:00 crc kubenswrapper[4810]: I1201 15:59:00.916523 4810 generic.go:334] "Generic (PLEG): container finished" podID="0341c875-f1fa-4a99-b850-202a0f3385db" containerID="4ed98088f5c2fb360a2deb9e3d90a041dbbffa43e3eefd755bfccba8c10b7cab" exitCode=0 Dec 01 15:59:00 crc kubenswrapper[4810]: I1201 15:59:00.916593 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7h8nl" event={"ID":"0341c875-f1fa-4a99-b850-202a0f3385db","Type":"ContainerDied","Data":"4ed98088f5c2fb360a2deb9e3d90a041dbbffa43e3eefd755bfccba8c10b7cab"} Dec 01 15:59:00 crc kubenswrapper[4810]: I1201 15:59:00.916656 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7h8nl" event={"ID":"0341c875-f1fa-4a99-b850-202a0f3385db","Type":"ContainerStarted","Data":"c356d16feda731d4581c9a07a174774c0b026c7287d988f0f5581cd090f02ae7"} Dec 01 15:59:00 crc kubenswrapper[4810]: I1201 15:59:00.918532 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 15:59:08 crc kubenswrapper[4810]: I1201 15:59:08.970695 4810 generic.go:334] "Generic (PLEG): container finished" podID="0341c875-f1fa-4a99-b850-202a0f3385db" containerID="c35ce6e543e3bf0aec1132243907d15329650dab8cdaace96164252996646c78" exitCode=0 Dec 01 15:59:08 crc kubenswrapper[4810]: I1201 15:59:08.970930 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7h8nl" event={"ID":"0341c875-f1fa-4a99-b850-202a0f3385db","Type":"ContainerDied","Data":"c35ce6e543e3bf0aec1132243907d15329650dab8cdaace96164252996646c78"} Dec 01 15:59:09 crc kubenswrapper[4810]: I1201 15:59:09.980365 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7h8nl" event={"ID":"0341c875-f1fa-4a99-b850-202a0f3385db","Type":"ContainerStarted","Data":"ac6c8ec5edc9df6f4a47cecdd9e9ed4715fefe5fc6e47e66e8415109b3d76c49"} Dec 01 15:59:09 crc kubenswrapper[4810]: I1201 15:59:09.998098 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7h8nl" podStartSLOduration=2.368417439 podStartE2EDuration="10.998073708s" podCreationTimestamp="2025-12-01 15:58:59 +0000 UTC" firstStartedPulling="2025-12-01 15:59:00.918101022 +0000 UTC m=+5106.681610665" lastFinishedPulling="2025-12-01 15:59:09.547757331 +0000 UTC m=+5115.311266934" observedRunningTime="2025-12-01 15:59:09.995796046 +0000 UTC m=+5115.759305669" watchObservedRunningTime="2025-12-01 15:59:09.998073708 +0000 UTC m=+5115.761583311" Dec 01 15:59:10 crc kubenswrapper[4810]: I1201 15:59:10.121109 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7h8nl" Dec 01 15:59:10 crc kubenswrapper[4810]: I1201 15:59:10.121647 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7h8nl" Dec 01 15:59:11 crc kubenswrapper[4810]: I1201 15:59:11.158349 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7h8nl" podUID="0341c875-f1fa-4a99-b850-202a0f3385db" containerName="registry-server" probeResult="failure" output=< Dec 01 15:59:11 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Dec 01 15:59:11 crc kubenswrapper[4810]: > Dec 01 15:59:20 crc kubenswrapper[4810]: I1201 15:59:20.181066 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7h8nl" Dec 01 15:59:20 crc kubenswrapper[4810]: I1201 15:59:20.228401 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7h8nl" Dec 01 15:59:20 crc kubenswrapper[4810]: I1201 15:59:20.299158 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7h8nl"] Dec 01 15:59:20 crc kubenswrapper[4810]: I1201 15:59:20.416999 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-djh66"] Dec 01 15:59:20 crc kubenswrapper[4810]: I1201 15:59:20.417565 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-djh66" podUID="b8cdfd34-81c3-49ff-bba6-8b0d4192520d" containerName="registry-server" containerID="cri-o://c190916a214ccbab94317963e484b4304815933e0346dcee4a94454baa18be9a" gracePeriod=2 Dec 01 15:59:23 crc kubenswrapper[4810]: I1201 15:59:23.065931 4810 generic.go:334] "Generic (PLEG): container finished" podID="b8cdfd34-81c3-49ff-bba6-8b0d4192520d" containerID="c190916a214ccbab94317963e484b4304815933e0346dcee4a94454baa18be9a" exitCode=0 Dec 01 15:59:23 crc kubenswrapper[4810]: I1201 15:59:23.065987 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djh66" event={"ID":"b8cdfd34-81c3-49ff-bba6-8b0d4192520d","Type":"ContainerDied","Data":"c190916a214ccbab94317963e484b4304815933e0346dcee4a94454baa18be9a"} Dec 01 15:59:23 crc kubenswrapper[4810]: I1201 15:59:23.482002 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djh66" Dec 01 15:59:23 crc kubenswrapper[4810]: I1201 15:59:23.546043 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gnm4\" (UniqueName: \"kubernetes.io/projected/b8cdfd34-81c3-49ff-bba6-8b0d4192520d-kube-api-access-7gnm4\") pod \"b8cdfd34-81c3-49ff-bba6-8b0d4192520d\" (UID: \"b8cdfd34-81c3-49ff-bba6-8b0d4192520d\") " Dec 01 15:59:23 crc kubenswrapper[4810]: I1201 15:59:23.546707 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8cdfd34-81c3-49ff-bba6-8b0d4192520d-utilities\") pod \"b8cdfd34-81c3-49ff-bba6-8b0d4192520d\" (UID: \"b8cdfd34-81c3-49ff-bba6-8b0d4192520d\") " Dec 01 15:59:23 crc kubenswrapper[4810]: I1201 15:59:23.546837 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8cdfd34-81c3-49ff-bba6-8b0d4192520d-catalog-content\") pod \"b8cdfd34-81c3-49ff-bba6-8b0d4192520d\" (UID: \"b8cdfd34-81c3-49ff-bba6-8b0d4192520d\") " Dec 01 15:59:23 crc kubenswrapper[4810]: I1201 15:59:23.547310 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8cdfd34-81c3-49ff-bba6-8b0d4192520d-utilities" (OuterVolumeSpecName: "utilities") pod "b8cdfd34-81c3-49ff-bba6-8b0d4192520d" (UID: "b8cdfd34-81c3-49ff-bba6-8b0d4192520d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:59:23 crc kubenswrapper[4810]: I1201 15:59:23.553715 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8cdfd34-81c3-49ff-bba6-8b0d4192520d-kube-api-access-7gnm4" (OuterVolumeSpecName: "kube-api-access-7gnm4") pod "b8cdfd34-81c3-49ff-bba6-8b0d4192520d" (UID: "b8cdfd34-81c3-49ff-bba6-8b0d4192520d"). InnerVolumeSpecName "kube-api-access-7gnm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:59:23 crc kubenswrapper[4810]: I1201 15:59:23.637530 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8cdfd34-81c3-49ff-bba6-8b0d4192520d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b8cdfd34-81c3-49ff-bba6-8b0d4192520d" (UID: "b8cdfd34-81c3-49ff-bba6-8b0d4192520d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 15:59:23 crc kubenswrapper[4810]: I1201 15:59:23.648404 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8cdfd34-81c3-49ff-bba6-8b0d4192520d-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 15:59:23 crc kubenswrapper[4810]: I1201 15:59:23.648640 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8cdfd34-81c3-49ff-bba6-8b0d4192520d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 15:59:23 crc kubenswrapper[4810]: I1201 15:59:23.648741 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gnm4\" (UniqueName: \"kubernetes.io/projected/b8cdfd34-81c3-49ff-bba6-8b0d4192520d-kube-api-access-7gnm4\") on node \"crc\" DevicePath \"\"" Dec 01 15:59:24 crc kubenswrapper[4810]: I1201 15:59:24.077381 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djh66" event={"ID":"b8cdfd34-81c3-49ff-bba6-8b0d4192520d","Type":"ContainerDied","Data":"431f800501d6e071a0021ac82554a8e9afe4194a99ab181d44c97e2180241618"} Dec 01 15:59:24 crc kubenswrapper[4810]: I1201 15:59:24.077442 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djh66" Dec 01 15:59:24 crc kubenswrapper[4810]: I1201 15:59:24.077529 4810 scope.go:117] "RemoveContainer" containerID="c190916a214ccbab94317963e484b4304815933e0346dcee4a94454baa18be9a" Dec 01 15:59:24 crc kubenswrapper[4810]: I1201 15:59:24.105838 4810 scope.go:117] "RemoveContainer" containerID="3fed3998eee8132d07f1ab86b0636824160c217f05d8a833e912e22fc3c49ae2" Dec 01 15:59:24 crc kubenswrapper[4810]: I1201 15:59:24.117147 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-djh66"] Dec 01 15:59:24 crc kubenswrapper[4810]: I1201 15:59:24.125072 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-djh66"] Dec 01 15:59:24 crc kubenswrapper[4810]: I1201 15:59:24.156876 4810 scope.go:117] "RemoveContainer" containerID="7b9a2f3f0d9962a491a066dc2a2be88fb77afe5efe4a5a93c71ae3880ea887ae" Dec 01 15:59:24 crc kubenswrapper[4810]: I1201 15:59:24.501488 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8cdfd34-81c3-49ff-bba6-8b0d4192520d" path="/var/lib/kubelet/pods/b8cdfd34-81c3-49ff-bba6-8b0d4192520d/volumes" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.612678 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79f96f64d9-4jcv8"] Dec 01 15:59:46 crc kubenswrapper[4810]: E1201 15:59:46.613562 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8cdfd34-81c3-49ff-bba6-8b0d4192520d" containerName="extract-utilities" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.613583 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8cdfd34-81c3-49ff-bba6-8b0d4192520d" containerName="extract-utilities" Dec 01 15:59:46 crc kubenswrapper[4810]: E1201 15:59:46.613612 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8cdfd34-81c3-49ff-bba6-8b0d4192520d" containerName="extract-content" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.613621 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8cdfd34-81c3-49ff-bba6-8b0d4192520d" containerName="extract-content" Dec 01 15:59:46 crc kubenswrapper[4810]: E1201 15:59:46.613648 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8cdfd34-81c3-49ff-bba6-8b0d4192520d" containerName="registry-server" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.613655 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8cdfd34-81c3-49ff-bba6-8b0d4192520d" containerName="registry-server" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.613851 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8cdfd34-81c3-49ff-bba6-8b0d4192520d" containerName="registry-server" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.614722 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79f96f64d9-4jcv8" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.618456 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-t7n9t" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.618937 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.619104 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.619231 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.621202 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-856cfd97cc-4jsph"] Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.622964 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-856cfd97cc-4jsph" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.625626 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.640315 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79f96f64d9-4jcv8"] Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.659428 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-856cfd97cc-4jsph"] Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.666229 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvh5g\" (UniqueName: \"kubernetes.io/projected/3e821830-241a-40ce-9374-acdc7af5a7df-kube-api-access-rvh5g\") pod \"dnsmasq-dns-856cfd97cc-4jsph\" (UID: \"3e821830-241a-40ce-9374-acdc7af5a7df\") " pod="openstack/dnsmasq-dns-856cfd97cc-4jsph" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.666318 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csznf\" (UniqueName: \"kubernetes.io/projected/d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc-kube-api-access-csznf\") pod \"dnsmasq-dns-79f96f64d9-4jcv8\" (UID: \"d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc\") " pod="openstack/dnsmasq-dns-79f96f64d9-4jcv8" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.666383 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc-config\") pod \"dnsmasq-dns-79f96f64d9-4jcv8\" (UID: \"d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc\") " pod="openstack/dnsmasq-dns-79f96f64d9-4jcv8" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.666413 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e821830-241a-40ce-9374-acdc7af5a7df-dns-svc\") pod \"dnsmasq-dns-856cfd97cc-4jsph\" (UID: \"3e821830-241a-40ce-9374-acdc7af5a7df\") " pod="openstack/dnsmasq-dns-856cfd97cc-4jsph" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.666451 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e821830-241a-40ce-9374-acdc7af5a7df-config\") pod \"dnsmasq-dns-856cfd97cc-4jsph\" (UID: \"3e821830-241a-40ce-9374-acdc7af5a7df\") " pod="openstack/dnsmasq-dns-856cfd97cc-4jsph" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.768167 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csznf\" (UniqueName: \"kubernetes.io/projected/d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc-kube-api-access-csznf\") pod \"dnsmasq-dns-79f96f64d9-4jcv8\" (UID: \"d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc\") " pod="openstack/dnsmasq-dns-79f96f64d9-4jcv8" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.768324 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc-config\") pod \"dnsmasq-dns-79f96f64d9-4jcv8\" (UID: \"d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc\") " pod="openstack/dnsmasq-dns-79f96f64d9-4jcv8" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.768362 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e821830-241a-40ce-9374-acdc7af5a7df-dns-svc\") pod \"dnsmasq-dns-856cfd97cc-4jsph\" (UID: \"3e821830-241a-40ce-9374-acdc7af5a7df\") " pod="openstack/dnsmasq-dns-856cfd97cc-4jsph" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.768389 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e821830-241a-40ce-9374-acdc7af5a7df-config\") pod \"dnsmasq-dns-856cfd97cc-4jsph\" (UID: \"3e821830-241a-40ce-9374-acdc7af5a7df\") " pod="openstack/dnsmasq-dns-856cfd97cc-4jsph" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.768461 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvh5g\" (UniqueName: \"kubernetes.io/projected/3e821830-241a-40ce-9374-acdc7af5a7df-kube-api-access-rvh5g\") pod \"dnsmasq-dns-856cfd97cc-4jsph\" (UID: \"3e821830-241a-40ce-9374-acdc7af5a7df\") " pod="openstack/dnsmasq-dns-856cfd97cc-4jsph" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.770446 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc-config\") pod \"dnsmasq-dns-79f96f64d9-4jcv8\" (UID: \"d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc\") " pod="openstack/dnsmasq-dns-79f96f64d9-4jcv8" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.771167 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e821830-241a-40ce-9374-acdc7af5a7df-dns-svc\") pod \"dnsmasq-dns-856cfd97cc-4jsph\" (UID: \"3e821830-241a-40ce-9374-acdc7af5a7df\") " pod="openstack/dnsmasq-dns-856cfd97cc-4jsph" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.777270 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e821830-241a-40ce-9374-acdc7af5a7df-config\") pod \"dnsmasq-dns-856cfd97cc-4jsph\" (UID: \"3e821830-241a-40ce-9374-acdc7af5a7df\") " pod="openstack/dnsmasq-dns-856cfd97cc-4jsph" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.800055 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csznf\" (UniqueName: \"kubernetes.io/projected/d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc-kube-api-access-csznf\") pod \"dnsmasq-dns-79f96f64d9-4jcv8\" (UID: \"d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc\") " pod="openstack/dnsmasq-dns-79f96f64d9-4jcv8" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.801022 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvh5g\" (UniqueName: \"kubernetes.io/projected/3e821830-241a-40ce-9374-acdc7af5a7df-kube-api-access-rvh5g\") pod \"dnsmasq-dns-856cfd97cc-4jsph\" (UID: \"3e821830-241a-40ce-9374-acdc7af5a7df\") " pod="openstack/dnsmasq-dns-856cfd97cc-4jsph" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.944944 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79f96f64d9-4jcv8" Dec 01 15:59:46 crc kubenswrapper[4810]: I1201 15:59:46.951749 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-856cfd97cc-4jsph" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.020026 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79f96f64d9-4jcv8"] Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.072355 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c7997469c-hghrb"] Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.075238 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c7997469c-hghrb" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.093764 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c7997469c-hghrb"] Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.179629 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a1de466-94d4-4b81-a784-f1a7c6b1921e-config\") pod \"dnsmasq-dns-c7997469c-hghrb\" (UID: \"4a1de466-94d4-4b81-a784-f1a7c6b1921e\") " pod="openstack/dnsmasq-dns-c7997469c-hghrb" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.180026 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxtj4\" (UniqueName: \"kubernetes.io/projected/4a1de466-94d4-4b81-a784-f1a7c6b1921e-kube-api-access-pxtj4\") pod \"dnsmasq-dns-c7997469c-hghrb\" (UID: \"4a1de466-94d4-4b81-a784-f1a7c6b1921e\") " pod="openstack/dnsmasq-dns-c7997469c-hghrb" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.180173 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a1de466-94d4-4b81-a784-f1a7c6b1921e-dns-svc\") pod \"dnsmasq-dns-c7997469c-hghrb\" (UID: \"4a1de466-94d4-4b81-a784-f1a7c6b1921e\") " pod="openstack/dnsmasq-dns-c7997469c-hghrb" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.282041 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a1de466-94d4-4b81-a784-f1a7c6b1921e-dns-svc\") pod \"dnsmasq-dns-c7997469c-hghrb\" (UID: \"4a1de466-94d4-4b81-a784-f1a7c6b1921e\") " pod="openstack/dnsmasq-dns-c7997469c-hghrb" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.282089 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a1de466-94d4-4b81-a784-f1a7c6b1921e-config\") pod \"dnsmasq-dns-c7997469c-hghrb\" (UID: \"4a1de466-94d4-4b81-a784-f1a7c6b1921e\") " pod="openstack/dnsmasq-dns-c7997469c-hghrb" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.282117 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxtj4\" (UniqueName: \"kubernetes.io/projected/4a1de466-94d4-4b81-a784-f1a7c6b1921e-kube-api-access-pxtj4\") pod \"dnsmasq-dns-c7997469c-hghrb\" (UID: \"4a1de466-94d4-4b81-a784-f1a7c6b1921e\") " pod="openstack/dnsmasq-dns-c7997469c-hghrb" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.283311 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a1de466-94d4-4b81-a784-f1a7c6b1921e-dns-svc\") pod \"dnsmasq-dns-c7997469c-hghrb\" (UID: \"4a1de466-94d4-4b81-a784-f1a7c6b1921e\") " pod="openstack/dnsmasq-dns-c7997469c-hghrb" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.283798 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a1de466-94d4-4b81-a784-f1a7c6b1921e-config\") pod \"dnsmasq-dns-c7997469c-hghrb\" (UID: \"4a1de466-94d4-4b81-a784-f1a7c6b1921e\") " pod="openstack/dnsmasq-dns-c7997469c-hghrb" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.308455 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxtj4\" (UniqueName: \"kubernetes.io/projected/4a1de466-94d4-4b81-a784-f1a7c6b1921e-kube-api-access-pxtj4\") pod \"dnsmasq-dns-c7997469c-hghrb\" (UID: \"4a1de466-94d4-4b81-a784-f1a7c6b1921e\") " pod="openstack/dnsmasq-dns-c7997469c-hghrb" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.435713 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c7997469c-hghrb" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.444494 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-856cfd97cc-4jsph"] Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.491617 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-657f7dd897-fsxbt"] Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.492989 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.513014 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-657f7dd897-fsxbt"] Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.586295 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw66f\" (UniqueName: \"kubernetes.io/projected/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021-kube-api-access-rw66f\") pod \"dnsmasq-dns-657f7dd897-fsxbt\" (UID: \"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021\") " pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.588048 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021-config\") pod \"dnsmasq-dns-657f7dd897-fsxbt\" (UID: \"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021\") " pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.588092 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021-dns-svc\") pod \"dnsmasq-dns-657f7dd897-fsxbt\" (UID: \"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021\") " pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.674146 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-856cfd97cc-4jsph"] Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.689391 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw66f\" (UniqueName: \"kubernetes.io/projected/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021-kube-api-access-rw66f\") pod \"dnsmasq-dns-657f7dd897-fsxbt\" (UID: \"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021\") " pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.689453 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021-config\") pod \"dnsmasq-dns-657f7dd897-fsxbt\" (UID: \"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021\") " pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.689756 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021-dns-svc\") pod \"dnsmasq-dns-657f7dd897-fsxbt\" (UID: \"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021\") " pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.690824 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021-dns-svc\") pod \"dnsmasq-dns-657f7dd897-fsxbt\" (UID: \"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021\") " pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.691859 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021-config\") pod \"dnsmasq-dns-657f7dd897-fsxbt\" (UID: \"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021\") " pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.712704 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw66f\" (UniqueName: \"kubernetes.io/projected/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021-kube-api-access-rw66f\") pod \"dnsmasq-dns-657f7dd897-fsxbt\" (UID: \"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021\") " pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.732753 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79f96f64d9-4jcv8"] Dec 01 15:59:47 crc kubenswrapper[4810]: W1201 15:59:47.741741 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9b2e3b4_1bb7_4b06_a291_b5af5fc695cc.slice/crio-a6a8dfc3aa0d6cbc80dabe52e93426244cc0aaa138c13cfa9df49774bcec67aa WatchSource:0}: Error finding container a6a8dfc3aa0d6cbc80dabe52e93426244cc0aaa138c13cfa9df49774bcec67aa: Status 404 returned error can't find the container with id a6a8dfc3aa0d6cbc80dabe52e93426244cc0aaa138c13cfa9df49774bcec67aa Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.789136 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c7997469c-hghrb"] Dec 01 15:59:47 crc kubenswrapper[4810]: I1201 15:59:47.836026 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.201732 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.203736 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.205928 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.205999 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.206307 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.206523 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.206731 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.206794 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.207277 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-mgw2x" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.219328 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.287421 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79f96f64d9-4jcv8" event={"ID":"d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc","Type":"ContainerStarted","Data":"a6a8dfc3aa0d6cbc80dabe52e93426244cc0aaa138c13cfa9df49774bcec67aa"} Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.288803 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-856cfd97cc-4jsph" event={"ID":"3e821830-241a-40ce-9374-acdc7af5a7df","Type":"ContainerStarted","Data":"25a5b43cba22962ef1f0f241da07fc2ba4bf4958ba4fa4b36f3e583325d85875"} Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.291022 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c7997469c-hghrb" event={"ID":"4a1de466-94d4-4b81-a784-f1a7c6b1921e","Type":"ContainerStarted","Data":"c333d680a3a1e2f7b938b7137403fa3211dc9a3550ff2443b7234bae0a987c67"} Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.300100 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0788ca8d-f584-47dd-8488-87d98c1e22f0-config-data\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.300154 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0788ca8d-f584-47dd-8488-87d98c1e22f0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.300178 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ltcv\" (UniqueName: \"kubernetes.io/projected/0788ca8d-f584-47dd-8488-87d98c1e22f0-kube-api-access-9ltcv\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.300196 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.300378 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.300584 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0788ca8d-f584-47dd-8488-87d98c1e22f0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.300615 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0788ca8d-f584-47dd-8488-87d98c1e22f0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.300719 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.300765 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.300792 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0788ca8d-f584-47dd-8488-87d98c1e22f0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.300830 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.303145 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-657f7dd897-fsxbt"] Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.402347 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.402406 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0788ca8d-f584-47dd-8488-87d98c1e22f0-config-data\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.402438 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0788ca8d-f584-47dd-8488-87d98c1e22f0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.402465 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ltcv\" (UniqueName: \"kubernetes.io/projected/0788ca8d-f584-47dd-8488-87d98c1e22f0-kube-api-access-9ltcv\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.402504 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.402542 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.402599 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0788ca8d-f584-47dd-8488-87d98c1e22f0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.402695 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0788ca8d-f584-47dd-8488-87d98c1e22f0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.402739 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.402768 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.402788 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0788ca8d-f584-47dd-8488-87d98c1e22f0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.403524 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.404683 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.408519 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0788ca8d-f584-47dd-8488-87d98c1e22f0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.410097 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.410118 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0788ca8d-f584-47dd-8488-87d98c1e22f0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.410537 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.410878 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.410904 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0788ca8d-f584-47dd-8488-87d98c1e22f0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.410907 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/61d3d42dde16e89ed7ad20c90a916e70e853d49d4b65a0dd42582d970d13e064/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.412300 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0788ca8d-f584-47dd-8488-87d98c1e22f0-config-data\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.412510 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0788ca8d-f584-47dd-8488-87d98c1e22f0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.423041 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ltcv\" (UniqueName: \"kubernetes.io/projected/0788ca8d-f584-47dd-8488-87d98c1e22f0-kube-api-access-9ltcv\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.455533 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\") pod \"rabbitmq-server-0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.531987 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.589197 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.590696 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.596790 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.596828 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.597771 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-dkzvm" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.606926 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.607036 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.607165 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.606964 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.622278 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.724450 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/090f0e84-a3f5-4706-872f-b62086e22577-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.725084 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.725131 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.725216 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.725244 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2vrk\" (UniqueName: \"kubernetes.io/projected/090f0e84-a3f5-4706-872f-b62086e22577-kube-api-access-s2vrk\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.725276 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/090f0e84-a3f5-4706-872f-b62086e22577-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.725319 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.725350 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/090f0e84-a3f5-4706-872f-b62086e22577-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.725379 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/090f0e84-a3f5-4706-872f-b62086e22577-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.725395 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/090f0e84-a3f5-4706-872f-b62086e22577-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.725412 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.827268 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.827311 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2vrk\" (UniqueName: \"kubernetes.io/projected/090f0e84-a3f5-4706-872f-b62086e22577-kube-api-access-s2vrk\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.827348 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/090f0e84-a3f5-4706-872f-b62086e22577-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.827382 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.827411 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/090f0e84-a3f5-4706-872f-b62086e22577-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.827437 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/090f0e84-a3f5-4706-872f-b62086e22577-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.827466 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/090f0e84-a3f5-4706-872f-b62086e22577-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.827518 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.827558 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/090f0e84-a3f5-4706-872f-b62086e22577-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.827597 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.827620 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.827980 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.828254 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.829185 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/090f0e84-a3f5-4706-872f-b62086e22577-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.829207 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/090f0e84-a3f5-4706-872f-b62086e22577-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.830062 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/090f0e84-a3f5-4706-872f-b62086e22577-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.832413 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/090f0e84-a3f5-4706-872f-b62086e22577-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.832588 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/090f0e84-a3f5-4706-872f-b62086e22577-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.832630 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.832861 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.832892 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c8c54e25171899303fbcbbb2f3f9c41dea8d63d5254c746c192fb7c677ca914d/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.832860 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.845431 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2vrk\" (UniqueName: \"kubernetes.io/projected/090f0e84-a3f5-4706-872f-b62086e22577-kube-api-access-s2vrk\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.867163 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\") pod \"rabbitmq-cell1-server-0\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:48 crc kubenswrapper[4810]: I1201 15:59:48.919383 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.025214 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 15:59:49 crc kubenswrapper[4810]: W1201 15:59:49.086703 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0788ca8d_f584_47dd_8488_87d98c1e22f0.slice/crio-6858861ea0b81d0f555cafb73ab5a963e6881e7d320dd72dc480720d1813936b WatchSource:0}: Error finding container 6858861ea0b81d0f555cafb73ab5a963e6881e7d320dd72dc480720d1813936b: Status 404 returned error can't find the container with id 6858861ea0b81d0f555cafb73ab5a963e6881e7d320dd72dc480720d1813936b Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.298453 4810 generic.go:334] "Generic (PLEG): container finished" podID="3e821830-241a-40ce-9374-acdc7af5a7df" containerID="d6cadf5cf7ccb19a835e1919d0dbbbb63f5691392ce9c9f3f75c71eb5a12c163" exitCode=0 Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.298538 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-856cfd97cc-4jsph" event={"ID":"3e821830-241a-40ce-9374-acdc7af5a7df","Type":"ContainerDied","Data":"d6cadf5cf7ccb19a835e1919d0dbbbb63f5691392ce9c9f3f75c71eb5a12c163"} Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.300572 4810 generic.go:334] "Generic (PLEG): container finished" podID="9710aadb-4ea9-4bd5-bbaa-ce3d474c9021" containerID="6b2cff3a2019f8bbaa63fd334b5c25c02460651640b06f2870e44937a5f6f626" exitCode=0 Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.300700 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" event={"ID":"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021","Type":"ContainerDied","Data":"6b2cff3a2019f8bbaa63fd334b5c25c02460651640b06f2870e44937a5f6f626"} Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.300730 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" event={"ID":"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021","Type":"ContainerStarted","Data":"8d4781b23752081ec0585802f60cf8a76e96093d278de0423b7016e4da00d107"} Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.302667 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0788ca8d-f584-47dd-8488-87d98c1e22f0","Type":"ContainerStarted","Data":"6858861ea0b81d0f555cafb73ab5a963e6881e7d320dd72dc480720d1813936b"} Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.304589 4810 generic.go:334] "Generic (PLEG): container finished" podID="4a1de466-94d4-4b81-a784-f1a7c6b1921e" containerID="f33c140ebef426c851a66a3c2da1b8d0add9d137609020a76ca395f26b6d771f" exitCode=0 Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.304677 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c7997469c-hghrb" event={"ID":"4a1de466-94d4-4b81-a784-f1a7c6b1921e","Type":"ContainerDied","Data":"f33c140ebef426c851a66a3c2da1b8d0add9d137609020a76ca395f26b6d771f"} Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.307177 4810 generic.go:334] "Generic (PLEG): container finished" podID="d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc" containerID="888826fb139c07d33477d5f4ccb16a5324e2a3f4e6dcbd61f83471a544a0866a" exitCode=0 Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.307215 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79f96f64d9-4jcv8" event={"ID":"d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc","Type":"ContainerDied","Data":"888826fb139c07d33477d5f4ccb16a5324e2a3f4e6dcbd61f83471a544a0866a"} Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.393920 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.409426 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.410704 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.415316 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.416133 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.416499 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.424027 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-dblcj" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.424439 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 01 15:59:49 crc kubenswrapper[4810]: W1201 15:59:49.431528 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod090f0e84_a3f5_4706_872f_b62086e22577.slice/crio-ee01afe9d7b912f957c8355a5950bab0e03a2ef4cd20508c5c59a1d151d2ac24 WatchSource:0}: Error finding container ee01afe9d7b912f957c8355a5950bab0e03a2ef4cd20508c5c59a1d151d2ac24: Status 404 returned error can't find the container with id ee01afe9d7b912f957c8355a5950bab0e03a2ef4cd20508c5c59a1d151d2ac24 Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.436087 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.540521 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9a264de6-cd64-441c-b935-d7c30e7ec733-kolla-config\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.540571 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9a264de6-cd64-441c-b935-d7c30e7ec733-config-data-default\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.540605 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9a264de6-cd64-441c-b935-d7c30e7ec733-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.540621 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a264de6-cd64-441c-b935-d7c30e7ec733-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.540660 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-dbec6bf0-3e2a-495e-ab9f-ab90dd29ceff\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dbec6bf0-3e2a-495e-ab9f-ab90dd29ceff\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.540686 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk6pm\" (UniqueName: \"kubernetes.io/projected/9a264de6-cd64-441c-b935-d7c30e7ec733-kube-api-access-tk6pm\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.540708 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a264de6-cd64-441c-b935-d7c30e7ec733-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.540732 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a264de6-cd64-441c-b935-d7c30e7ec733-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.627799 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79f96f64d9-4jcv8" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.646425 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-dbec6bf0-3e2a-495e-ab9f-ab90dd29ceff\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dbec6bf0-3e2a-495e-ab9f-ab90dd29ceff\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.646487 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk6pm\" (UniqueName: \"kubernetes.io/projected/9a264de6-cd64-441c-b935-d7c30e7ec733-kube-api-access-tk6pm\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.646513 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a264de6-cd64-441c-b935-d7c30e7ec733-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.646538 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a264de6-cd64-441c-b935-d7c30e7ec733-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.646576 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9a264de6-cd64-441c-b935-d7c30e7ec733-kolla-config\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.646610 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9a264de6-cd64-441c-b935-d7c30e7ec733-config-data-default\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.646663 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9a264de6-cd64-441c-b935-d7c30e7ec733-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.646680 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a264de6-cd64-441c-b935-d7c30e7ec733-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.650309 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9a264de6-cd64-441c-b935-d7c30e7ec733-kolla-config\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.650460 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a264de6-cd64-441c-b935-d7c30e7ec733-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.650733 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9a264de6-cd64-441c-b935-d7c30e7ec733-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.650845 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9a264de6-cd64-441c-b935-d7c30e7ec733-config-data-default\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.653816 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a264de6-cd64-441c-b935-d7c30e7ec733-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: E1201 15:59:49.655766 4810 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 01 15:59:49 crc kubenswrapper[4810]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/4a1de466-94d4-4b81-a784-f1a7c6b1921e/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 01 15:59:49 crc kubenswrapper[4810]: > podSandboxID="c333d680a3a1e2f7b938b7137403fa3211dc9a3550ff2443b7234bae0a987c67" Dec 01 15:59:49 crc kubenswrapper[4810]: E1201 15:59:49.655906 4810 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 01 15:59:49 crc kubenswrapper[4810]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:4218330ae90f65f4a2c1d93334812c4d04a4ed1d46013269252aba16e1138627,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8chc6h5bh56fh546hb7hc8h67h5bchffh577h697h5b5h5bdh59bhf6hf4h558hb5h578h595h5cchfbh644h59ch7fh654h547h587h5cbh5d5h8fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pxtj4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-c7997469c-hghrb_openstack(4a1de466-94d4-4b81-a784-f1a7c6b1921e): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/4a1de466-94d4-4b81-a784-f1a7c6b1921e/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 01 15:59:49 crc kubenswrapper[4810]: > logger="UnhandledError" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.656043 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.656065 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-dbec6bf0-3e2a-495e-ab9f-ab90dd29ceff\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dbec6bf0-3e2a-495e-ab9f-ab90dd29ceff\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/04ed5b2f96c93c3f1e66d789594333e39b37a0c9226a690ee6feb5bdb59e394e/globalmount\"" pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: E1201 15:59:49.657010 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/4a1de466-94d4-4b81-a784-f1a7c6b1921e/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-c7997469c-hghrb" podUID="4a1de466-94d4-4b81-a784-f1a7c6b1921e" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.658841 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a264de6-cd64-441c-b935-d7c30e7ec733-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.667120 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk6pm\" (UniqueName: \"kubernetes.io/projected/9a264de6-cd64-441c-b935-d7c30e7ec733-kube-api-access-tk6pm\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.692917 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-dbec6bf0-3e2a-495e-ab9f-ab90dd29ceff\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dbec6bf0-3e2a-495e-ab9f-ab90dd29ceff\") pod \"openstack-galera-0\" (UID: \"9a264de6-cd64-441c-b935-d7c30e7ec733\") " pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.718243 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-856cfd97cc-4jsph" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.748050 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csznf\" (UniqueName: \"kubernetes.io/projected/d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc-kube-api-access-csznf\") pod \"d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc\" (UID: \"d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc\") " Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.748103 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc-config\") pod \"d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc\" (UID: \"d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc\") " Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.752823 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc-kube-api-access-csznf" (OuterVolumeSpecName: "kube-api-access-csznf") pod "d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc" (UID: "d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc"). InnerVolumeSpecName "kube-api-access-csznf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.768898 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.849807 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e821830-241a-40ce-9374-acdc7af5a7df-dns-svc\") pod \"3e821830-241a-40ce-9374-acdc7af5a7df\" (UID: \"3e821830-241a-40ce-9374-acdc7af5a7df\") " Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.850002 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvh5g\" (UniqueName: \"kubernetes.io/projected/3e821830-241a-40ce-9374-acdc7af5a7df-kube-api-access-rvh5g\") pod \"3e821830-241a-40ce-9374-acdc7af5a7df\" (UID: \"3e821830-241a-40ce-9374-acdc7af5a7df\") " Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.850068 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e821830-241a-40ce-9374-acdc7af5a7df-config\") pod \"3e821830-241a-40ce-9374-acdc7af5a7df\" (UID: \"3e821830-241a-40ce-9374-acdc7af5a7df\") " Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.850395 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csznf\" (UniqueName: \"kubernetes.io/projected/d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc-kube-api-access-csznf\") on node \"crc\" DevicePath \"\"" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.954062 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e821830-241a-40ce-9374-acdc7af5a7df-kube-api-access-rvh5g" (OuterVolumeSpecName: "kube-api-access-rvh5g") pod "3e821830-241a-40ce-9374-acdc7af5a7df" (UID: "3e821830-241a-40ce-9374-acdc7af5a7df"). InnerVolumeSpecName "kube-api-access-rvh5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:59:49 crc kubenswrapper[4810]: I1201 15:59:49.975335 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e821830-241a-40ce-9374-acdc7af5a7df-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3e821830-241a-40ce-9374-acdc7af5a7df" (UID: "3e821830-241a-40ce-9374-acdc7af5a7df"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.029712 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.052889 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e821830-241a-40ce-9374-acdc7af5a7df-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.052915 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvh5g\" (UniqueName: \"kubernetes.io/projected/3e821830-241a-40ce-9374-acdc7af5a7df-kube-api-access-rvh5g\") on node \"crc\" DevicePath \"\"" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.060911 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc-config" (OuterVolumeSpecName: "config") pod "d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc" (UID: "d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.066351 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e821830-241a-40ce-9374-acdc7af5a7df-config" (OuterVolumeSpecName: "config") pod "3e821830-241a-40ce-9374-acdc7af5a7df" (UID: "3e821830-241a-40ce-9374-acdc7af5a7df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.155345 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e821830-241a-40ce-9374-acdc7af5a7df-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.155384 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.316008 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" event={"ID":"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021","Type":"ContainerStarted","Data":"03779612bae7e481a9889c92835631a5c5c49129f5c4cb323d7f4fe9d9963b4a"} Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.316169 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.317930 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0788ca8d-f584-47dd-8488-87d98c1e22f0","Type":"ContainerStarted","Data":"e2454a338a9d1bb4e69ba8caeb7fb292da7aa5cf3876c0272796812e2600841d"} Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.319635 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"090f0e84-a3f5-4706-872f-b62086e22577","Type":"ContainerStarted","Data":"ee01afe9d7b912f957c8355a5950bab0e03a2ef4cd20508c5c59a1d151d2ac24"} Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.321291 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9a264de6-cd64-441c-b935-d7c30e7ec733","Type":"ContainerStarted","Data":"b1ff7f8dc6a1845bf9e223078b09e65a1405773d660ba4042c0be00ce6528664"} Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.321321 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9a264de6-cd64-441c-b935-d7c30e7ec733","Type":"ContainerStarted","Data":"822be5f6eb3ad03971da6dfcf1c361a1cd556d55b85c083fc2b3c7154b46e348"} Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.322828 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79f96f64d9-4jcv8" event={"ID":"d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc","Type":"ContainerDied","Data":"a6a8dfc3aa0d6cbc80dabe52e93426244cc0aaa138c13cfa9df49774bcec67aa"} Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.322870 4810 scope.go:117] "RemoveContainer" containerID="888826fb139c07d33477d5f4ccb16a5324e2a3f4e6dcbd61f83471a544a0866a" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.323015 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79f96f64d9-4jcv8" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.325079 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-856cfd97cc-4jsph" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.326570 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-856cfd97cc-4jsph" event={"ID":"3e821830-241a-40ce-9374-acdc7af5a7df","Type":"ContainerDied","Data":"25a5b43cba22962ef1f0f241da07fc2ba4bf4958ba4fa4b36f3e583325d85875"} Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.346725 4810 scope.go:117] "RemoveContainer" containerID="d6cadf5cf7ccb19a835e1919d0dbbbb63f5691392ce9c9f3f75c71eb5a12c163" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.381310 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" podStartSLOduration=3.381288433 podStartE2EDuration="3.381288433s" podCreationTimestamp="2025-12-01 15:59:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:59:50.346007085 +0000 UTC m=+5156.109516708" watchObservedRunningTime="2025-12-01 15:59:50.381288433 +0000 UTC m=+5156.144798026" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.571000 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-856cfd97cc-4jsph"] Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.583638 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-856cfd97cc-4jsph"] Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.616761 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79f96f64d9-4jcv8"] Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.631370 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79f96f64d9-4jcv8"] Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.929999 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 15:59:50 crc kubenswrapper[4810]: E1201 15:59:50.930870 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e821830-241a-40ce-9374-acdc7af5a7df" containerName="init" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.930885 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e821830-241a-40ce-9374-acdc7af5a7df" containerName="init" Dec 01 15:59:50 crc kubenswrapper[4810]: E1201 15:59:50.930899 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc" containerName="init" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.930906 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc" containerName="init" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.931058 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc" containerName="init" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.931070 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e821830-241a-40ce-9374-acdc7af5a7df" containerName="init" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.931826 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.933889 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.933995 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.935810 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-xt8pp" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.935937 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 01 15:59:50 crc kubenswrapper[4810]: I1201 15:59:50.947681 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.067062 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxnht\" (UniqueName: \"kubernetes.io/projected/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-kube-api-access-jxnht\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.067158 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.067191 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-cc9197d1-73f7-489f-b056-230b3df1716a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cc9197d1-73f7-489f-b056-230b3df1716a\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.067208 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.067233 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.067303 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.067368 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.067424 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.168461 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.168560 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.168606 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.168676 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxnht\" (UniqueName: \"kubernetes.io/projected/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-kube-api-access-jxnht\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.168713 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.168753 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-cc9197d1-73f7-489f-b056-230b3df1716a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cc9197d1-73f7-489f-b056-230b3df1716a\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.168793 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.168830 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.169445 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.169767 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.169913 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.170709 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.171305 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.171352 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-cc9197d1-73f7-489f-b056-230b3df1716a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cc9197d1-73f7-489f-b056-230b3df1716a\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5393b86b46fcc13a80a64c2fe9f291cf122f80bb42ba87030e6fb9be75b753c0/globalmount\"" pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.173929 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.175564 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.204046 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxnht\" (UniqueName: \"kubernetes.io/projected/bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877-kube-api-access-jxnht\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.204594 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-cc9197d1-73f7-489f-b056-230b3df1716a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cc9197d1-73f7-489f-b056-230b3df1716a\") pod \"openstack-cell1-galera-0\" (UID: \"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877\") " pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.248363 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.314407 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.315588 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.317779 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.318090 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-shltn" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.318266 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.339828 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"090f0e84-a3f5-4706-872f-b62086e22577","Type":"ContainerStarted","Data":"e724435386ae94fcebeff126f3b3387876d53bb30199d1134178eb3f74e49bf2"} Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.340010 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.341912 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c7997469c-hghrb" event={"ID":"4a1de466-94d4-4b81-a784-f1a7c6b1921e","Type":"ContainerStarted","Data":"5c3b32e6fa15609985d407ca049ed28c2cbfe8c0360c8b5bb01f28355e99caba"} Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.342345 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-c7997469c-hghrb" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.397362 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-c7997469c-hghrb" podStartSLOduration=4.397334013 podStartE2EDuration="4.397334013s" podCreationTimestamp="2025-12-01 15:59:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:59:51.390104147 +0000 UTC m=+5157.153613750" watchObservedRunningTime="2025-12-01 15:59:51.397334013 +0000 UTC m=+5157.160843616" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.472866 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0b48f20e-a414-4e62-a9a7-8ecfaef12dc7-kolla-config\") pod \"memcached-0\" (UID: \"0b48f20e-a414-4e62-a9a7-8ecfaef12dc7\") " pod="openstack/memcached-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.473315 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b48f20e-a414-4e62-a9a7-8ecfaef12dc7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0b48f20e-a414-4e62-a9a7-8ecfaef12dc7\") " pod="openstack/memcached-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.473520 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrfj7\" (UniqueName: \"kubernetes.io/projected/0b48f20e-a414-4e62-a9a7-8ecfaef12dc7-kube-api-access-mrfj7\") pod \"memcached-0\" (UID: \"0b48f20e-a414-4e62-a9a7-8ecfaef12dc7\") " pod="openstack/memcached-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.473565 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b48f20e-a414-4e62-a9a7-8ecfaef12dc7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0b48f20e-a414-4e62-a9a7-8ecfaef12dc7\") " pod="openstack/memcached-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.473624 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b48f20e-a414-4e62-a9a7-8ecfaef12dc7-config-data\") pod \"memcached-0\" (UID: \"0b48f20e-a414-4e62-a9a7-8ecfaef12dc7\") " pod="openstack/memcached-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.575405 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrfj7\" (UniqueName: \"kubernetes.io/projected/0b48f20e-a414-4e62-a9a7-8ecfaef12dc7-kube-api-access-mrfj7\") pod \"memcached-0\" (UID: \"0b48f20e-a414-4e62-a9a7-8ecfaef12dc7\") " pod="openstack/memcached-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.575507 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b48f20e-a414-4e62-a9a7-8ecfaef12dc7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0b48f20e-a414-4e62-a9a7-8ecfaef12dc7\") " pod="openstack/memcached-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.575539 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b48f20e-a414-4e62-a9a7-8ecfaef12dc7-config-data\") pod \"memcached-0\" (UID: \"0b48f20e-a414-4e62-a9a7-8ecfaef12dc7\") " pod="openstack/memcached-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.575612 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0b48f20e-a414-4e62-a9a7-8ecfaef12dc7-kolla-config\") pod \"memcached-0\" (UID: \"0b48f20e-a414-4e62-a9a7-8ecfaef12dc7\") " pod="openstack/memcached-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.575657 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b48f20e-a414-4e62-a9a7-8ecfaef12dc7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0b48f20e-a414-4e62-a9a7-8ecfaef12dc7\") " pod="openstack/memcached-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.576759 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b48f20e-a414-4e62-a9a7-8ecfaef12dc7-config-data\") pod \"memcached-0\" (UID: \"0b48f20e-a414-4e62-a9a7-8ecfaef12dc7\") " pod="openstack/memcached-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.577116 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0b48f20e-a414-4e62-a9a7-8ecfaef12dc7-kolla-config\") pod \"memcached-0\" (UID: \"0b48f20e-a414-4e62-a9a7-8ecfaef12dc7\") " pod="openstack/memcached-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.579586 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b48f20e-a414-4e62-a9a7-8ecfaef12dc7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0b48f20e-a414-4e62-a9a7-8ecfaef12dc7\") " pod="openstack/memcached-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.579656 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b48f20e-a414-4e62-a9a7-8ecfaef12dc7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0b48f20e-a414-4e62-a9a7-8ecfaef12dc7\") " pod="openstack/memcached-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.594575 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrfj7\" (UniqueName: \"kubernetes.io/projected/0b48f20e-a414-4e62-a9a7-8ecfaef12dc7-kube-api-access-mrfj7\") pod \"memcached-0\" (UID: \"0b48f20e-a414-4e62-a9a7-8ecfaef12dc7\") " pod="openstack/memcached-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.677847 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 01 15:59:51 crc kubenswrapper[4810]: I1201 15:59:51.726934 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 15:59:51 crc kubenswrapper[4810]: W1201 15:59:51.727281 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbcb3289_cf5d_4f8d_bf1f_fe30d45ef877.slice/crio-dda0889555ec133450879ea8d85e5290218f51c78434ce6bd32c4d0cc266ca31 WatchSource:0}: Error finding container dda0889555ec133450879ea8d85e5290218f51c78434ce6bd32c4d0cc266ca31: Status 404 returned error can't find the container with id dda0889555ec133450879ea8d85e5290218f51c78434ce6bd32c4d0cc266ca31 Dec 01 15:59:52 crc kubenswrapper[4810]: I1201 15:59:52.078534 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 01 15:59:52 crc kubenswrapper[4810]: W1201 15:59:52.085726 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b48f20e_a414_4e62_a9a7_8ecfaef12dc7.slice/crio-0804f53b2b69616840e64e5b02eee7304c34b67cc2b183df7c0c803e883dbaa4 WatchSource:0}: Error finding container 0804f53b2b69616840e64e5b02eee7304c34b67cc2b183df7c0c803e883dbaa4: Status 404 returned error can't find the container with id 0804f53b2b69616840e64e5b02eee7304c34b67cc2b183df7c0c803e883dbaa4 Dec 01 15:59:52 crc kubenswrapper[4810]: I1201 15:59:52.374284 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877","Type":"ContainerStarted","Data":"6e1aeacb9217308dbb1579a5122699480985a2b137584c982783cfdc38bf2096"} Dec 01 15:59:52 crc kubenswrapper[4810]: I1201 15:59:52.374805 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877","Type":"ContainerStarted","Data":"dda0889555ec133450879ea8d85e5290218f51c78434ce6bd32c4d0cc266ca31"} Dec 01 15:59:52 crc kubenswrapper[4810]: I1201 15:59:52.376325 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0b48f20e-a414-4e62-a9a7-8ecfaef12dc7","Type":"ContainerStarted","Data":"03625f980e46b21fdb0c10552fc6052e1f373e3ae8cad08dcd6b48dc51855fb8"} Dec 01 15:59:52 crc kubenswrapper[4810]: I1201 15:59:52.376358 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0b48f20e-a414-4e62-a9a7-8ecfaef12dc7","Type":"ContainerStarted","Data":"0804f53b2b69616840e64e5b02eee7304c34b67cc2b183df7c0c803e883dbaa4"} Dec 01 15:59:52 crc kubenswrapper[4810]: I1201 15:59:52.376966 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 01 15:59:52 crc kubenswrapper[4810]: I1201 15:59:52.418255 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=1.418236067 podStartE2EDuration="1.418236067s" podCreationTimestamp="2025-12-01 15:59:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:59:52.409737526 +0000 UTC m=+5158.173247129" watchObservedRunningTime="2025-12-01 15:59:52.418236067 +0000 UTC m=+5158.181745670" Dec 01 15:59:52 crc kubenswrapper[4810]: I1201 15:59:52.500526 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e821830-241a-40ce-9374-acdc7af5a7df" path="/var/lib/kubelet/pods/3e821830-241a-40ce-9374-acdc7af5a7df/volumes" Dec 01 15:59:52 crc kubenswrapper[4810]: I1201 15:59:52.501003 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc" path="/var/lib/kubelet/pods/d9b2e3b4-1bb7-4b06-a291-b5af5fc695cc/volumes" Dec 01 15:59:54 crc kubenswrapper[4810]: I1201 15:59:54.390999 4810 generic.go:334] "Generic (PLEG): container finished" podID="9a264de6-cd64-441c-b935-d7c30e7ec733" containerID="b1ff7f8dc6a1845bf9e223078b09e65a1405773d660ba4042c0be00ce6528664" exitCode=0 Dec 01 15:59:54 crc kubenswrapper[4810]: I1201 15:59:54.391277 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9a264de6-cd64-441c-b935-d7c30e7ec733","Type":"ContainerDied","Data":"b1ff7f8dc6a1845bf9e223078b09e65a1405773d660ba4042c0be00ce6528664"} Dec 01 15:59:55 crc kubenswrapper[4810]: I1201 15:59:55.398857 4810 generic.go:334] "Generic (PLEG): container finished" podID="bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877" containerID="6e1aeacb9217308dbb1579a5122699480985a2b137584c982783cfdc38bf2096" exitCode=0 Dec 01 15:59:55 crc kubenswrapper[4810]: I1201 15:59:55.398935 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877","Type":"ContainerDied","Data":"6e1aeacb9217308dbb1579a5122699480985a2b137584c982783cfdc38bf2096"} Dec 01 15:59:55 crc kubenswrapper[4810]: I1201 15:59:55.400985 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9a264de6-cd64-441c-b935-d7c30e7ec733","Type":"ContainerStarted","Data":"1b86a9e1fcf43f90fac4ec7d7a0b560928bce212082b023f72ad4d767e51234b"} Dec 01 15:59:55 crc kubenswrapper[4810]: I1201 15:59:55.449664 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=7.449641855 podStartE2EDuration="7.449641855s" podCreationTimestamp="2025-12-01 15:59:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:59:55.442790789 +0000 UTC m=+5161.206300392" watchObservedRunningTime="2025-12-01 15:59:55.449641855 +0000 UTC m=+5161.213151458" Dec 01 15:59:56 crc kubenswrapper[4810]: I1201 15:59:56.409403 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877","Type":"ContainerStarted","Data":"1946931005a40a0692795ec5e1b0072043cc5e9547d3515a9b71322c6b8f7750"} Dec 01 15:59:56 crc kubenswrapper[4810]: I1201 15:59:56.433833 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.43381407 podStartE2EDuration="7.43381407s" podCreationTimestamp="2025-12-01 15:59:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 15:59:56.432044171 +0000 UTC m=+5162.195553784" watchObservedRunningTime="2025-12-01 15:59:56.43381407 +0000 UTC m=+5162.197323673" Dec 01 15:59:57 crc kubenswrapper[4810]: I1201 15:59:57.437616 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-c7997469c-hghrb" Dec 01 15:59:57 crc kubenswrapper[4810]: I1201 15:59:57.837656 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" Dec 01 15:59:57 crc kubenswrapper[4810]: I1201 15:59:57.903708 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c7997469c-hghrb"] Dec 01 15:59:58 crc kubenswrapper[4810]: I1201 15:59:58.421770 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-c7997469c-hghrb" podUID="4a1de466-94d4-4b81-a784-f1a7c6b1921e" containerName="dnsmasq-dns" containerID="cri-o://5c3b32e6fa15609985d407ca049ed28c2cbfe8c0360c8b5bb01f28355e99caba" gracePeriod=10 Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.411625 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c7997469c-hghrb" Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.433609 4810 generic.go:334] "Generic (PLEG): container finished" podID="4a1de466-94d4-4b81-a784-f1a7c6b1921e" containerID="5c3b32e6fa15609985d407ca049ed28c2cbfe8c0360c8b5bb01f28355e99caba" exitCode=0 Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.433676 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c7997469c-hghrb" event={"ID":"4a1de466-94d4-4b81-a784-f1a7c6b1921e","Type":"ContainerDied","Data":"5c3b32e6fa15609985d407ca049ed28c2cbfe8c0360c8b5bb01f28355e99caba"} Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.433716 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c7997469c-hghrb" Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.433735 4810 scope.go:117] "RemoveContainer" containerID="5c3b32e6fa15609985d407ca049ed28c2cbfe8c0360c8b5bb01f28355e99caba" Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.433720 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c7997469c-hghrb" event={"ID":"4a1de466-94d4-4b81-a784-f1a7c6b1921e","Type":"ContainerDied","Data":"c333d680a3a1e2f7b938b7137403fa3211dc9a3550ff2443b7234bae0a987c67"} Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.457672 4810 scope.go:117] "RemoveContainer" containerID="f33c140ebef426c851a66a3c2da1b8d0add9d137609020a76ca395f26b6d771f" Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.530370 4810 scope.go:117] "RemoveContainer" containerID="5c3b32e6fa15609985d407ca049ed28c2cbfe8c0360c8b5bb01f28355e99caba" Dec 01 15:59:59 crc kubenswrapper[4810]: E1201 15:59:59.531393 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c3b32e6fa15609985d407ca049ed28c2cbfe8c0360c8b5bb01f28355e99caba\": container with ID starting with 5c3b32e6fa15609985d407ca049ed28c2cbfe8c0360c8b5bb01f28355e99caba not found: ID does not exist" containerID="5c3b32e6fa15609985d407ca049ed28c2cbfe8c0360c8b5bb01f28355e99caba" Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.531436 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c3b32e6fa15609985d407ca049ed28c2cbfe8c0360c8b5bb01f28355e99caba"} err="failed to get container status \"5c3b32e6fa15609985d407ca049ed28c2cbfe8c0360c8b5bb01f28355e99caba\": rpc error: code = NotFound desc = could not find container \"5c3b32e6fa15609985d407ca049ed28c2cbfe8c0360c8b5bb01f28355e99caba\": container with ID starting with 5c3b32e6fa15609985d407ca049ed28c2cbfe8c0360c8b5bb01f28355e99caba not found: ID does not exist" Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.531461 4810 scope.go:117] "RemoveContainer" containerID="f33c140ebef426c851a66a3c2da1b8d0add9d137609020a76ca395f26b6d771f" Dec 01 15:59:59 crc kubenswrapper[4810]: E1201 15:59:59.531919 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f33c140ebef426c851a66a3c2da1b8d0add9d137609020a76ca395f26b6d771f\": container with ID starting with f33c140ebef426c851a66a3c2da1b8d0add9d137609020a76ca395f26b6d771f not found: ID does not exist" containerID="f33c140ebef426c851a66a3c2da1b8d0add9d137609020a76ca395f26b6d771f" Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.531956 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f33c140ebef426c851a66a3c2da1b8d0add9d137609020a76ca395f26b6d771f"} err="failed to get container status \"f33c140ebef426c851a66a3c2da1b8d0add9d137609020a76ca395f26b6d771f\": rpc error: code = NotFound desc = could not find container \"f33c140ebef426c851a66a3c2da1b8d0add9d137609020a76ca395f26b6d771f\": container with ID starting with f33c140ebef426c851a66a3c2da1b8d0add9d137609020a76ca395f26b6d771f not found: ID does not exist" Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.603378 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxtj4\" (UniqueName: \"kubernetes.io/projected/4a1de466-94d4-4b81-a784-f1a7c6b1921e-kube-api-access-pxtj4\") pod \"4a1de466-94d4-4b81-a784-f1a7c6b1921e\" (UID: \"4a1de466-94d4-4b81-a784-f1a7c6b1921e\") " Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.603556 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a1de466-94d4-4b81-a784-f1a7c6b1921e-config\") pod \"4a1de466-94d4-4b81-a784-f1a7c6b1921e\" (UID: \"4a1de466-94d4-4b81-a784-f1a7c6b1921e\") " Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.603609 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a1de466-94d4-4b81-a784-f1a7c6b1921e-dns-svc\") pod \"4a1de466-94d4-4b81-a784-f1a7c6b1921e\" (UID: \"4a1de466-94d4-4b81-a784-f1a7c6b1921e\") " Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.608345 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a1de466-94d4-4b81-a784-f1a7c6b1921e-kube-api-access-pxtj4" (OuterVolumeSpecName: "kube-api-access-pxtj4") pod "4a1de466-94d4-4b81-a784-f1a7c6b1921e" (UID: "4a1de466-94d4-4b81-a784-f1a7c6b1921e"). InnerVolumeSpecName "kube-api-access-pxtj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.638133 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a1de466-94d4-4b81-a784-f1a7c6b1921e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4a1de466-94d4-4b81-a784-f1a7c6b1921e" (UID: "4a1de466-94d4-4b81-a784-f1a7c6b1921e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.640777 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a1de466-94d4-4b81-a784-f1a7c6b1921e-config" (OuterVolumeSpecName: "config") pod "4a1de466-94d4-4b81-a784-f1a7c6b1921e" (UID: "4a1de466-94d4-4b81-a784-f1a7c6b1921e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.705793 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a1de466-94d4-4b81-a784-f1a7c6b1921e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.705835 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxtj4\" (UniqueName: \"kubernetes.io/projected/4a1de466-94d4-4b81-a784-f1a7c6b1921e-kube-api-access-pxtj4\") on node \"crc\" DevicePath \"\"" Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.705853 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a1de466-94d4-4b81-a784-f1a7c6b1921e-config\") on node \"crc\" DevicePath \"\"" Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.768055 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c7997469c-hghrb"] Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.769185 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.769418 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 01 15:59:59 crc kubenswrapper[4810]: I1201 15:59:59.775112 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-c7997469c-hghrb"] Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.154898 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz"] Dec 01 16:00:00 crc kubenswrapper[4810]: E1201 16:00:00.155316 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a1de466-94d4-4b81-a784-f1a7c6b1921e" containerName="dnsmasq-dns" Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.155331 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a1de466-94d4-4b81-a784-f1a7c6b1921e" containerName="dnsmasq-dns" Dec 01 16:00:00 crc kubenswrapper[4810]: E1201 16:00:00.155353 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a1de466-94d4-4b81-a784-f1a7c6b1921e" containerName="init" Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.155364 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a1de466-94d4-4b81-a784-f1a7c6b1921e" containerName="init" Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.155580 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a1de466-94d4-4b81-a784-f1a7c6b1921e" containerName="dnsmasq-dns" Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.156253 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz" Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.159638 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.159866 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.182280 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz"] Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.317091 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/767af1af-2227-4c98-9b5d-24f8e64a7d6d-config-volume\") pod \"collect-profiles-29410080-4nllz\" (UID: \"767af1af-2227-4c98-9b5d-24f8e64a7d6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz" Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.317191 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx4qb\" (UniqueName: \"kubernetes.io/projected/767af1af-2227-4c98-9b5d-24f8e64a7d6d-kube-api-access-nx4qb\") pod \"collect-profiles-29410080-4nllz\" (UID: \"767af1af-2227-4c98-9b5d-24f8e64a7d6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz" Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.317241 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/767af1af-2227-4c98-9b5d-24f8e64a7d6d-secret-volume\") pod \"collect-profiles-29410080-4nllz\" (UID: \"767af1af-2227-4c98-9b5d-24f8e64a7d6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz" Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.419011 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx4qb\" (UniqueName: \"kubernetes.io/projected/767af1af-2227-4c98-9b5d-24f8e64a7d6d-kube-api-access-nx4qb\") pod \"collect-profiles-29410080-4nllz\" (UID: \"767af1af-2227-4c98-9b5d-24f8e64a7d6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz" Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.419092 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/767af1af-2227-4c98-9b5d-24f8e64a7d6d-secret-volume\") pod \"collect-profiles-29410080-4nllz\" (UID: \"767af1af-2227-4c98-9b5d-24f8e64a7d6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz" Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.419156 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/767af1af-2227-4c98-9b5d-24f8e64a7d6d-config-volume\") pod \"collect-profiles-29410080-4nllz\" (UID: \"767af1af-2227-4c98-9b5d-24f8e64a7d6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz" Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.420085 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/767af1af-2227-4c98-9b5d-24f8e64a7d6d-config-volume\") pod \"collect-profiles-29410080-4nllz\" (UID: \"767af1af-2227-4c98-9b5d-24f8e64a7d6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz" Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.424127 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/767af1af-2227-4c98-9b5d-24f8e64a7d6d-secret-volume\") pod \"collect-profiles-29410080-4nllz\" (UID: \"767af1af-2227-4c98-9b5d-24f8e64a7d6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz" Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.439408 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx4qb\" (UniqueName: \"kubernetes.io/projected/767af1af-2227-4c98-9b5d-24f8e64a7d6d-kube-api-access-nx4qb\") pod \"collect-profiles-29410080-4nllz\" (UID: \"767af1af-2227-4c98-9b5d-24f8e64a7d6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz" Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.482327 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz" Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.500490 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a1de466-94d4-4b81-a784-f1a7c6b1921e" path="/var/lib/kubelet/pods/4a1de466-94d4-4b81-a784-f1a7c6b1921e/volumes" Dec 01 16:00:00 crc kubenswrapper[4810]: I1201 16:00:00.951960 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz"] Dec 01 16:00:00 crc kubenswrapper[4810]: W1201 16:00:00.953831 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod767af1af_2227_4c98_9b5d_24f8e64a7d6d.slice/crio-c8b434ca08e31a94d9027c1f5701ac76d1bd591bd550667f6055e6c2b56b8e79 WatchSource:0}: Error finding container c8b434ca08e31a94d9027c1f5701ac76d1bd591bd550667f6055e6c2b56b8e79: Status 404 returned error can't find the container with id c8b434ca08e31a94d9027c1f5701ac76d1bd591bd550667f6055e6c2b56b8e79 Dec 01 16:00:01 crc kubenswrapper[4810]: I1201 16:00:01.249417 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 01 16:00:01 crc kubenswrapper[4810]: I1201 16:00:01.249748 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 01 16:00:01 crc kubenswrapper[4810]: I1201 16:00:01.450410 4810 generic.go:334] "Generic (PLEG): container finished" podID="767af1af-2227-4c98-9b5d-24f8e64a7d6d" containerID="fabd7de3f73e504c542541093725e23523892d07d1879c6b7a4f6160f29bcc59" exitCode=0 Dec 01 16:00:01 crc kubenswrapper[4810]: I1201 16:00:01.450486 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz" event={"ID":"767af1af-2227-4c98-9b5d-24f8e64a7d6d","Type":"ContainerDied","Data":"fabd7de3f73e504c542541093725e23523892d07d1879c6b7a4f6160f29bcc59"} Dec 01 16:00:01 crc kubenswrapper[4810]: I1201 16:00:01.451172 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz" event={"ID":"767af1af-2227-4c98-9b5d-24f8e64a7d6d","Type":"ContainerStarted","Data":"c8b434ca08e31a94d9027c1f5701ac76d1bd591bd550667f6055e6c2b56b8e79"} Dec 01 16:00:01 crc kubenswrapper[4810]: I1201 16:00:01.679217 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 01 16:00:01 crc kubenswrapper[4810]: I1201 16:00:01.943004 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 01 16:00:02 crc kubenswrapper[4810]: I1201 16:00:02.016116 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 01 16:00:02 crc kubenswrapper[4810]: I1201 16:00:02.753200 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz" Dec 01 16:00:02 crc kubenswrapper[4810]: I1201 16:00:02.855512 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nx4qb\" (UniqueName: \"kubernetes.io/projected/767af1af-2227-4c98-9b5d-24f8e64a7d6d-kube-api-access-nx4qb\") pod \"767af1af-2227-4c98-9b5d-24f8e64a7d6d\" (UID: \"767af1af-2227-4c98-9b5d-24f8e64a7d6d\") " Dec 01 16:00:02 crc kubenswrapper[4810]: I1201 16:00:02.855618 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/767af1af-2227-4c98-9b5d-24f8e64a7d6d-config-volume\") pod \"767af1af-2227-4c98-9b5d-24f8e64a7d6d\" (UID: \"767af1af-2227-4c98-9b5d-24f8e64a7d6d\") " Dec 01 16:00:02 crc kubenswrapper[4810]: I1201 16:00:02.855676 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/767af1af-2227-4c98-9b5d-24f8e64a7d6d-secret-volume\") pod \"767af1af-2227-4c98-9b5d-24f8e64a7d6d\" (UID: \"767af1af-2227-4c98-9b5d-24f8e64a7d6d\") " Dec 01 16:00:02 crc kubenswrapper[4810]: I1201 16:00:02.856356 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/767af1af-2227-4c98-9b5d-24f8e64a7d6d-config-volume" (OuterVolumeSpecName: "config-volume") pod "767af1af-2227-4c98-9b5d-24f8e64a7d6d" (UID: "767af1af-2227-4c98-9b5d-24f8e64a7d6d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:00:02 crc kubenswrapper[4810]: I1201 16:00:02.861054 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/767af1af-2227-4c98-9b5d-24f8e64a7d6d-kube-api-access-nx4qb" (OuterVolumeSpecName: "kube-api-access-nx4qb") pod "767af1af-2227-4c98-9b5d-24f8e64a7d6d" (UID: "767af1af-2227-4c98-9b5d-24f8e64a7d6d"). InnerVolumeSpecName "kube-api-access-nx4qb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:00:02 crc kubenswrapper[4810]: I1201 16:00:02.861220 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/767af1af-2227-4c98-9b5d-24f8e64a7d6d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "767af1af-2227-4c98-9b5d-24f8e64a7d6d" (UID: "767af1af-2227-4c98-9b5d-24f8e64a7d6d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:00:02 crc kubenswrapper[4810]: I1201 16:00:02.960457 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nx4qb\" (UniqueName: \"kubernetes.io/projected/767af1af-2227-4c98-9b5d-24f8e64a7d6d-kube-api-access-nx4qb\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:02 crc kubenswrapper[4810]: I1201 16:00:02.960798 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/767af1af-2227-4c98-9b5d-24f8e64a7d6d-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:02 crc kubenswrapper[4810]: I1201 16:00:02.960811 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/767af1af-2227-4c98-9b5d-24f8e64a7d6d-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:03 crc kubenswrapper[4810]: I1201 16:00:03.425229 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 01 16:00:03 crc kubenswrapper[4810]: I1201 16:00:03.466576 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz" event={"ID":"767af1af-2227-4c98-9b5d-24f8e64a7d6d","Type":"ContainerDied","Data":"c8b434ca08e31a94d9027c1f5701ac76d1bd591bd550667f6055e6c2b56b8e79"} Dec 01 16:00:03 crc kubenswrapper[4810]: I1201 16:00:03.466612 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8b434ca08e31a94d9027c1f5701ac76d1bd591bd550667f6055e6c2b56b8e79" Dec 01 16:00:03 crc kubenswrapper[4810]: I1201 16:00:03.466615 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz" Dec 01 16:00:03 crc kubenswrapper[4810]: I1201 16:00:03.496979 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 01 16:00:03 crc kubenswrapper[4810]: I1201 16:00:03.831443 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9"] Dec 01 16:00:03 crc kubenswrapper[4810]: I1201 16:00:03.837596 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410035-sbvh9"] Dec 01 16:00:04 crc kubenswrapper[4810]: I1201 16:00:04.501727 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2c3a243-4fa8-4c09-90ba-fb901628b7da" path="/var/lib/kubelet/pods/c2c3a243-4fa8-4c09-90ba-fb901628b7da/volumes" Dec 01 16:00:08 crc kubenswrapper[4810]: I1201 16:00:08.761583 4810 scope.go:117] "RemoveContainer" containerID="59c98a47ec8e218fb94ced748c247d8b6232e1e05323abf0ee48987c383faaf3" Dec 01 16:00:22 crc kubenswrapper[4810]: I1201 16:00:22.611392 4810 generic.go:334] "Generic (PLEG): container finished" podID="0788ca8d-f584-47dd-8488-87d98c1e22f0" containerID="e2454a338a9d1bb4e69ba8caeb7fb292da7aa5cf3876c0272796812e2600841d" exitCode=0 Dec 01 16:00:22 crc kubenswrapper[4810]: I1201 16:00:22.611556 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0788ca8d-f584-47dd-8488-87d98c1e22f0","Type":"ContainerDied","Data":"e2454a338a9d1bb4e69ba8caeb7fb292da7aa5cf3876c0272796812e2600841d"} Dec 01 16:00:22 crc kubenswrapper[4810]: I1201 16:00:22.616839 4810 generic.go:334] "Generic (PLEG): container finished" podID="090f0e84-a3f5-4706-872f-b62086e22577" containerID="e724435386ae94fcebeff126f3b3387876d53bb30199d1134178eb3f74e49bf2" exitCode=0 Dec 01 16:00:22 crc kubenswrapper[4810]: I1201 16:00:22.616897 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"090f0e84-a3f5-4706-872f-b62086e22577","Type":"ContainerDied","Data":"e724435386ae94fcebeff126f3b3387876d53bb30199d1134178eb3f74e49bf2"} Dec 01 16:00:23 crc kubenswrapper[4810]: I1201 16:00:23.627027 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0788ca8d-f584-47dd-8488-87d98c1e22f0","Type":"ContainerStarted","Data":"4794e0892c96e9d985051845f733fda1c06719545108d6ebcafb6662f600f4a8"} Dec 01 16:00:23 crc kubenswrapper[4810]: I1201 16:00:23.627557 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 01 16:00:23 crc kubenswrapper[4810]: I1201 16:00:23.629324 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"090f0e84-a3f5-4706-872f-b62086e22577","Type":"ContainerStarted","Data":"830d2f4682ce22e2fcbbe9ba3ba054d64a9347e2e7694441c75add8d7c9c3f30"} Dec 01 16:00:23 crc kubenswrapper[4810]: I1201 16:00:23.629550 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:23 crc kubenswrapper[4810]: I1201 16:00:23.648706 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.648687877 podStartE2EDuration="36.648687877s" podCreationTimestamp="2025-12-01 15:59:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:00:23.646381814 +0000 UTC m=+5189.409891427" watchObservedRunningTime="2025-12-01 16:00:23.648687877 +0000 UTC m=+5189.412197480" Dec 01 16:00:38 crc kubenswrapper[4810]: I1201 16:00:38.535708 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 01 16:00:38 crc kubenswrapper[4810]: I1201 16:00:38.571836 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=51.57180575 podStartE2EDuration="51.57180575s" podCreationTimestamp="2025-12-01 15:59:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:00:23.671393344 +0000 UTC m=+5189.434902947" watchObservedRunningTime="2025-12-01 16:00:38.57180575 +0000 UTC m=+5204.335315363" Dec 01 16:00:38 crc kubenswrapper[4810]: I1201 16:00:38.922744 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:43 crc kubenswrapper[4810]: I1201 16:00:43.375929 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-964767c4f-mxwfj"] Dec 01 16:00:43 crc kubenswrapper[4810]: E1201 16:00:43.376567 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="767af1af-2227-4c98-9b5d-24f8e64a7d6d" containerName="collect-profiles" Dec 01 16:00:43 crc kubenswrapper[4810]: I1201 16:00:43.376580 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="767af1af-2227-4c98-9b5d-24f8e64a7d6d" containerName="collect-profiles" Dec 01 16:00:43 crc kubenswrapper[4810]: I1201 16:00:43.376712 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="767af1af-2227-4c98-9b5d-24f8e64a7d6d" containerName="collect-profiles" Dec 01 16:00:43 crc kubenswrapper[4810]: I1201 16:00:43.377633 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-964767c4f-mxwfj" Dec 01 16:00:43 crc kubenswrapper[4810]: I1201 16:00:43.388959 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-964767c4f-mxwfj"] Dec 01 16:00:43 crc kubenswrapper[4810]: I1201 16:00:43.493629 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjpqd\" (UniqueName: \"kubernetes.io/projected/47f4dca5-a8bd-4017-948f-8add5492e78f-kube-api-access-wjpqd\") pod \"dnsmasq-dns-964767c4f-mxwfj\" (UID: \"47f4dca5-a8bd-4017-948f-8add5492e78f\") " pod="openstack/dnsmasq-dns-964767c4f-mxwfj" Dec 01 16:00:43 crc kubenswrapper[4810]: I1201 16:00:43.494042 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47f4dca5-a8bd-4017-948f-8add5492e78f-config\") pod \"dnsmasq-dns-964767c4f-mxwfj\" (UID: \"47f4dca5-a8bd-4017-948f-8add5492e78f\") " pod="openstack/dnsmasq-dns-964767c4f-mxwfj" Dec 01 16:00:43 crc kubenswrapper[4810]: I1201 16:00:43.494079 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47f4dca5-a8bd-4017-948f-8add5492e78f-dns-svc\") pod \"dnsmasq-dns-964767c4f-mxwfj\" (UID: \"47f4dca5-a8bd-4017-948f-8add5492e78f\") " pod="openstack/dnsmasq-dns-964767c4f-mxwfj" Dec 01 16:00:43 crc kubenswrapper[4810]: I1201 16:00:43.595206 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjpqd\" (UniqueName: \"kubernetes.io/projected/47f4dca5-a8bd-4017-948f-8add5492e78f-kube-api-access-wjpqd\") pod \"dnsmasq-dns-964767c4f-mxwfj\" (UID: \"47f4dca5-a8bd-4017-948f-8add5492e78f\") " pod="openstack/dnsmasq-dns-964767c4f-mxwfj" Dec 01 16:00:43 crc kubenswrapper[4810]: I1201 16:00:43.595312 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47f4dca5-a8bd-4017-948f-8add5492e78f-config\") pod \"dnsmasq-dns-964767c4f-mxwfj\" (UID: \"47f4dca5-a8bd-4017-948f-8add5492e78f\") " pod="openstack/dnsmasq-dns-964767c4f-mxwfj" Dec 01 16:00:43 crc kubenswrapper[4810]: I1201 16:00:43.595337 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47f4dca5-a8bd-4017-948f-8add5492e78f-dns-svc\") pod \"dnsmasq-dns-964767c4f-mxwfj\" (UID: \"47f4dca5-a8bd-4017-948f-8add5492e78f\") " pod="openstack/dnsmasq-dns-964767c4f-mxwfj" Dec 01 16:00:43 crc kubenswrapper[4810]: I1201 16:00:43.596676 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47f4dca5-a8bd-4017-948f-8add5492e78f-config\") pod \"dnsmasq-dns-964767c4f-mxwfj\" (UID: \"47f4dca5-a8bd-4017-948f-8add5492e78f\") " pod="openstack/dnsmasq-dns-964767c4f-mxwfj" Dec 01 16:00:43 crc kubenswrapper[4810]: I1201 16:00:43.596915 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47f4dca5-a8bd-4017-948f-8add5492e78f-dns-svc\") pod \"dnsmasq-dns-964767c4f-mxwfj\" (UID: \"47f4dca5-a8bd-4017-948f-8add5492e78f\") " pod="openstack/dnsmasq-dns-964767c4f-mxwfj" Dec 01 16:00:43 crc kubenswrapper[4810]: I1201 16:00:43.615200 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjpqd\" (UniqueName: \"kubernetes.io/projected/47f4dca5-a8bd-4017-948f-8add5492e78f-kube-api-access-wjpqd\") pod \"dnsmasq-dns-964767c4f-mxwfj\" (UID: \"47f4dca5-a8bd-4017-948f-8add5492e78f\") " pod="openstack/dnsmasq-dns-964767c4f-mxwfj" Dec 01 16:00:43 crc kubenswrapper[4810]: I1201 16:00:43.693835 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-964767c4f-mxwfj" Dec 01 16:00:44 crc kubenswrapper[4810]: I1201 16:00:44.078043 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 16:00:44 crc kubenswrapper[4810]: I1201 16:00:44.125036 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-964767c4f-mxwfj"] Dec 01 16:00:44 crc kubenswrapper[4810]: W1201 16:00:44.152107 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47f4dca5_a8bd_4017_948f_8add5492e78f.slice/crio-9df1fd9da6a928e0dbe4defb38a9d56a565638f7e47cf0abe2a06563fe82d97e WatchSource:0}: Error finding container 9df1fd9da6a928e0dbe4defb38a9d56a565638f7e47cf0abe2a06563fe82d97e: Status 404 returned error can't find the container with id 9df1fd9da6a928e0dbe4defb38a9d56a565638f7e47cf0abe2a06563fe82d97e Dec 01 16:00:44 crc kubenswrapper[4810]: I1201 16:00:44.723122 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xnpkl"] Dec 01 16:00:44 crc kubenswrapper[4810]: I1201 16:00:44.724751 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xnpkl" Dec 01 16:00:44 crc kubenswrapper[4810]: I1201 16:00:44.746822 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xnpkl"] Dec 01 16:00:44 crc kubenswrapper[4810]: I1201 16:00:44.796534 4810 generic.go:334] "Generic (PLEG): container finished" podID="47f4dca5-a8bd-4017-948f-8add5492e78f" containerID="121e95f4c3189ba3c64f8dd383a65c3c03a106899a7d7c0e58558a88344f67d9" exitCode=0 Dec 01 16:00:44 crc kubenswrapper[4810]: I1201 16:00:44.796583 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-964767c4f-mxwfj" event={"ID":"47f4dca5-a8bd-4017-948f-8add5492e78f","Type":"ContainerDied","Data":"121e95f4c3189ba3c64f8dd383a65c3c03a106899a7d7c0e58558a88344f67d9"} Dec 01 16:00:44 crc kubenswrapper[4810]: I1201 16:00:44.796609 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-964767c4f-mxwfj" event={"ID":"47f4dca5-a8bd-4017-948f-8add5492e78f","Type":"ContainerStarted","Data":"9df1fd9da6a928e0dbe4defb38a9d56a565638f7e47cf0abe2a06563fe82d97e"} Dec 01 16:00:44 crc kubenswrapper[4810]: I1201 16:00:44.813287 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc86968-4a25-41c5-9612-d0f5570bd122-utilities\") pod \"redhat-marketplace-xnpkl\" (UID: \"bfc86968-4a25-41c5-9612-d0f5570bd122\") " pod="openshift-marketplace/redhat-marketplace-xnpkl" Dec 01 16:00:44 crc kubenswrapper[4810]: I1201 16:00:44.813348 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwvlv\" (UniqueName: \"kubernetes.io/projected/bfc86968-4a25-41c5-9612-d0f5570bd122-kube-api-access-pwvlv\") pod \"redhat-marketplace-xnpkl\" (UID: \"bfc86968-4a25-41c5-9612-d0f5570bd122\") " pod="openshift-marketplace/redhat-marketplace-xnpkl" Dec 01 16:00:44 crc kubenswrapper[4810]: I1201 16:00:44.813378 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc86968-4a25-41c5-9612-d0f5570bd122-catalog-content\") pod \"redhat-marketplace-xnpkl\" (UID: \"bfc86968-4a25-41c5-9612-d0f5570bd122\") " pod="openshift-marketplace/redhat-marketplace-xnpkl" Dec 01 16:00:44 crc kubenswrapper[4810]: I1201 16:00:44.914212 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc86968-4a25-41c5-9612-d0f5570bd122-utilities\") pod \"redhat-marketplace-xnpkl\" (UID: \"bfc86968-4a25-41c5-9612-d0f5570bd122\") " pod="openshift-marketplace/redhat-marketplace-xnpkl" Dec 01 16:00:44 crc kubenswrapper[4810]: I1201 16:00:44.914280 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwvlv\" (UniqueName: \"kubernetes.io/projected/bfc86968-4a25-41c5-9612-d0f5570bd122-kube-api-access-pwvlv\") pod \"redhat-marketplace-xnpkl\" (UID: \"bfc86968-4a25-41c5-9612-d0f5570bd122\") " pod="openshift-marketplace/redhat-marketplace-xnpkl" Dec 01 16:00:44 crc kubenswrapper[4810]: I1201 16:00:44.914314 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc86968-4a25-41c5-9612-d0f5570bd122-catalog-content\") pod \"redhat-marketplace-xnpkl\" (UID: \"bfc86968-4a25-41c5-9612-d0f5570bd122\") " pod="openshift-marketplace/redhat-marketplace-xnpkl" Dec 01 16:00:44 crc kubenswrapper[4810]: I1201 16:00:44.914847 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc86968-4a25-41c5-9612-d0f5570bd122-catalog-content\") pod \"redhat-marketplace-xnpkl\" (UID: \"bfc86968-4a25-41c5-9612-d0f5570bd122\") " pod="openshift-marketplace/redhat-marketplace-xnpkl" Dec 01 16:00:44 crc kubenswrapper[4810]: I1201 16:00:44.915094 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc86968-4a25-41c5-9612-d0f5570bd122-utilities\") pod \"redhat-marketplace-xnpkl\" (UID: \"bfc86968-4a25-41c5-9612-d0f5570bd122\") " pod="openshift-marketplace/redhat-marketplace-xnpkl" Dec 01 16:00:44 crc kubenswrapper[4810]: I1201 16:00:44.946541 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwvlv\" (UniqueName: \"kubernetes.io/projected/bfc86968-4a25-41c5-9612-d0f5570bd122-kube-api-access-pwvlv\") pod \"redhat-marketplace-xnpkl\" (UID: \"bfc86968-4a25-41c5-9612-d0f5570bd122\") " pod="openshift-marketplace/redhat-marketplace-xnpkl" Dec 01 16:00:45 crc kubenswrapper[4810]: I1201 16:00:45.004510 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 16:00:45 crc kubenswrapper[4810]: I1201 16:00:45.044265 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xnpkl" Dec 01 16:00:45 crc kubenswrapper[4810]: I1201 16:00:45.504233 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xnpkl"] Dec 01 16:00:45 crc kubenswrapper[4810]: W1201 16:00:45.509963 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbfc86968_4a25_41c5_9612_d0f5570bd122.slice/crio-bc4acc9bd94ec87e2bf135cf85cd2077b569d24ea356f9d1d21a5316392a0d13 WatchSource:0}: Error finding container bc4acc9bd94ec87e2bf135cf85cd2077b569d24ea356f9d1d21a5316392a0d13: Status 404 returned error can't find the container with id bc4acc9bd94ec87e2bf135cf85cd2077b569d24ea356f9d1d21a5316392a0d13 Dec 01 16:00:45 crc kubenswrapper[4810]: I1201 16:00:45.805652 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-964767c4f-mxwfj" event={"ID":"47f4dca5-a8bd-4017-948f-8add5492e78f","Type":"ContainerStarted","Data":"a91cbfa70cc72acc400e8cb6eb1913ac06cbf54150c2af7da400cd56e29191c6"} Dec 01 16:00:45 crc kubenswrapper[4810]: I1201 16:00:45.805945 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-964767c4f-mxwfj" Dec 01 16:00:45 crc kubenswrapper[4810]: I1201 16:00:45.808012 4810 generic.go:334] "Generic (PLEG): container finished" podID="bfc86968-4a25-41c5-9612-d0f5570bd122" containerID="3e36f9a660690c5334a6d7b01cc1f21ff9c4d9a4472307d4eb45c5c05e47229d" exitCode=0 Dec 01 16:00:45 crc kubenswrapper[4810]: I1201 16:00:45.808088 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xnpkl" event={"ID":"bfc86968-4a25-41c5-9612-d0f5570bd122","Type":"ContainerDied","Data":"3e36f9a660690c5334a6d7b01cc1f21ff9c4d9a4472307d4eb45c5c05e47229d"} Dec 01 16:00:45 crc kubenswrapper[4810]: I1201 16:00:45.808147 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xnpkl" event={"ID":"bfc86968-4a25-41c5-9612-d0f5570bd122","Type":"ContainerStarted","Data":"bc4acc9bd94ec87e2bf135cf85cd2077b569d24ea356f9d1d21a5316392a0d13"} Dec 01 16:00:45 crc kubenswrapper[4810]: I1201 16:00:45.827496 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-964767c4f-mxwfj" podStartSLOduration=2.82744639 podStartE2EDuration="2.82744639s" podCreationTimestamp="2025-12-01 16:00:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:00:45.824602493 +0000 UTC m=+5211.588112096" watchObservedRunningTime="2025-12-01 16:00:45.82744639 +0000 UTC m=+5211.590955993" Dec 01 16:00:47 crc kubenswrapper[4810]: I1201 16:00:47.831153 4810 generic.go:334] "Generic (PLEG): container finished" podID="bfc86968-4a25-41c5-9612-d0f5570bd122" containerID="9340620f3eea5359ba4cbcc22b12e3fdf0487a0b761821583e642a0e98ee933d" exitCode=0 Dec 01 16:00:47 crc kubenswrapper[4810]: I1201 16:00:47.831963 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xnpkl" event={"ID":"bfc86968-4a25-41c5-9612-d0f5570bd122","Type":"ContainerDied","Data":"9340620f3eea5359ba4cbcc22b12e3fdf0487a0b761821583e642a0e98ee933d"} Dec 01 16:00:48 crc kubenswrapper[4810]: I1201 16:00:48.704954 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="0788ca8d-f584-47dd-8488-87d98c1e22f0" containerName="rabbitmq" containerID="cri-o://4794e0892c96e9d985051845f733fda1c06719545108d6ebcafb6662f600f4a8" gracePeriod=604796 Dec 01 16:00:48 crc kubenswrapper[4810]: I1201 16:00:48.841696 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xnpkl" event={"ID":"bfc86968-4a25-41c5-9612-d0f5570bd122","Type":"ContainerStarted","Data":"2b827f8ab11bbe2f0d2e72802614d10898e91971a7a1ef660bf07f94eb0114db"} Dec 01 16:00:48 crc kubenswrapper[4810]: I1201 16:00:48.870064 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xnpkl" podStartSLOduration=2.340035829 podStartE2EDuration="4.870040702s" podCreationTimestamp="2025-12-01 16:00:44 +0000 UTC" firstStartedPulling="2025-12-01 16:00:45.809091061 +0000 UTC m=+5211.572600654" lastFinishedPulling="2025-12-01 16:00:48.339095924 +0000 UTC m=+5214.102605527" observedRunningTime="2025-12-01 16:00:48.864767038 +0000 UTC m=+5214.628276641" watchObservedRunningTime="2025-12-01 16:00:48.870040702 +0000 UTC m=+5214.633550305" Dec 01 16:00:49 crc kubenswrapper[4810]: I1201 16:00:49.611434 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="090f0e84-a3f5-4706-872f-b62086e22577" containerName="rabbitmq" containerID="cri-o://830d2f4682ce22e2fcbbe9ba3ba054d64a9347e2e7694441c75add8d7c9c3f30" gracePeriod=604796 Dec 01 16:00:53 crc kubenswrapper[4810]: I1201 16:00:53.695656 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-964767c4f-mxwfj" Dec 01 16:00:53 crc kubenswrapper[4810]: I1201 16:00:53.753011 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-657f7dd897-fsxbt"] Dec 01 16:00:53 crc kubenswrapper[4810]: I1201 16:00:53.753292 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" podUID="9710aadb-4ea9-4bd5-bbaa-ce3d474c9021" containerName="dnsmasq-dns" containerID="cri-o://03779612bae7e481a9889c92835631a5c5c49129f5c4cb323d7f4fe9d9963b4a" gracePeriod=10 Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.654371 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.785809 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021-dns-svc\") pod \"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021\" (UID: \"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021\") " Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.785900 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw66f\" (UniqueName: \"kubernetes.io/projected/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021-kube-api-access-rw66f\") pod \"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021\" (UID: \"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021\") " Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.785991 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021-config\") pod \"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021\" (UID: \"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021\") " Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.803723 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021-kube-api-access-rw66f" (OuterVolumeSpecName: "kube-api-access-rw66f") pod "9710aadb-4ea9-4bd5-bbaa-ce3d474c9021" (UID: "9710aadb-4ea9-4bd5-bbaa-ce3d474c9021"). InnerVolumeSpecName "kube-api-access-rw66f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.828731 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021-config" (OuterVolumeSpecName: "config") pod "9710aadb-4ea9-4bd5-bbaa-ce3d474c9021" (UID: "9710aadb-4ea9-4bd5-bbaa-ce3d474c9021"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.843045 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9710aadb-4ea9-4bd5-bbaa-ce3d474c9021" (UID: "9710aadb-4ea9-4bd5-bbaa-ce3d474c9021"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.885912 4810 generic.go:334] "Generic (PLEG): container finished" podID="9710aadb-4ea9-4bd5-bbaa-ce3d474c9021" containerID="03779612bae7e481a9889c92835631a5c5c49129f5c4cb323d7f4fe9d9963b4a" exitCode=0 Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.885997 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" event={"ID":"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021","Type":"ContainerDied","Data":"03779612bae7e481a9889c92835631a5c5c49129f5c4cb323d7f4fe9d9963b4a"} Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.886044 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" event={"ID":"9710aadb-4ea9-4bd5-bbaa-ce3d474c9021","Type":"ContainerDied","Data":"8d4781b23752081ec0585802f60cf8a76e96093d278de0423b7016e4da00d107"} Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.886076 4810 scope.go:117] "RemoveContainer" containerID="03779612bae7e481a9889c92835631a5c5c49129f5c4cb323d7f4fe9d9963b4a" Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.886276 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-657f7dd897-fsxbt" Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.890024 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.890053 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.890069 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw66f\" (UniqueName: \"kubernetes.io/projected/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021-kube-api-access-rw66f\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.907794 4810 scope.go:117] "RemoveContainer" containerID="6b2cff3a2019f8bbaa63fd334b5c25c02460651640b06f2870e44937a5f6f626" Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.934910 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-657f7dd897-fsxbt"] Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.941377 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-657f7dd897-fsxbt"] Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.949159 4810 scope.go:117] "RemoveContainer" containerID="03779612bae7e481a9889c92835631a5c5c49129f5c4cb323d7f4fe9d9963b4a" Dec 01 16:00:54 crc kubenswrapper[4810]: E1201 16:00:54.950019 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03779612bae7e481a9889c92835631a5c5c49129f5c4cb323d7f4fe9d9963b4a\": container with ID starting with 03779612bae7e481a9889c92835631a5c5c49129f5c4cb323d7f4fe9d9963b4a not found: ID does not exist" containerID="03779612bae7e481a9889c92835631a5c5c49129f5c4cb323d7f4fe9d9963b4a" Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.950057 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03779612bae7e481a9889c92835631a5c5c49129f5c4cb323d7f4fe9d9963b4a"} err="failed to get container status \"03779612bae7e481a9889c92835631a5c5c49129f5c4cb323d7f4fe9d9963b4a\": rpc error: code = NotFound desc = could not find container \"03779612bae7e481a9889c92835631a5c5c49129f5c4cb323d7f4fe9d9963b4a\": container with ID starting with 03779612bae7e481a9889c92835631a5c5c49129f5c4cb323d7f4fe9d9963b4a not found: ID does not exist" Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.950084 4810 scope.go:117] "RemoveContainer" containerID="6b2cff3a2019f8bbaa63fd334b5c25c02460651640b06f2870e44937a5f6f626" Dec 01 16:00:54 crc kubenswrapper[4810]: E1201 16:00:54.950711 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b2cff3a2019f8bbaa63fd334b5c25c02460651640b06f2870e44937a5f6f626\": container with ID starting with 6b2cff3a2019f8bbaa63fd334b5c25c02460651640b06f2870e44937a5f6f626 not found: ID does not exist" containerID="6b2cff3a2019f8bbaa63fd334b5c25c02460651640b06f2870e44937a5f6f626" Dec 01 16:00:54 crc kubenswrapper[4810]: I1201 16:00:54.950737 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b2cff3a2019f8bbaa63fd334b5c25c02460651640b06f2870e44937a5f6f626"} err="failed to get container status \"6b2cff3a2019f8bbaa63fd334b5c25c02460651640b06f2870e44937a5f6f626\": rpc error: code = NotFound desc = could not find container \"6b2cff3a2019f8bbaa63fd334b5c25c02460651640b06f2870e44937a5f6f626\": container with ID starting with 6b2cff3a2019f8bbaa63fd334b5c25c02460651640b06f2870e44937a5f6f626 not found: ID does not exist" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.045892 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xnpkl" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.045952 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xnpkl" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.088596 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xnpkl" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.370331 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.509692 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0788ca8d-f584-47dd-8488-87d98c1e22f0-config-data\") pod \"0788ca8d-f584-47dd-8488-87d98c1e22f0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.509774 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-erlang-cookie\") pod \"0788ca8d-f584-47dd-8488-87d98c1e22f0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.509840 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0788ca8d-f584-47dd-8488-87d98c1e22f0-erlang-cookie-secret\") pod \"0788ca8d-f584-47dd-8488-87d98c1e22f0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.509940 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-plugins\") pod \"0788ca8d-f584-47dd-8488-87d98c1e22f0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.509969 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ltcv\" (UniqueName: \"kubernetes.io/projected/0788ca8d-f584-47dd-8488-87d98c1e22f0-kube-api-access-9ltcv\") pod \"0788ca8d-f584-47dd-8488-87d98c1e22f0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.510001 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-tls\") pod \"0788ca8d-f584-47dd-8488-87d98c1e22f0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.510032 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-confd\") pod \"0788ca8d-f584-47dd-8488-87d98c1e22f0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.510053 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0788ca8d-f584-47dd-8488-87d98c1e22f0-server-conf\") pod \"0788ca8d-f584-47dd-8488-87d98c1e22f0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.510095 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0788ca8d-f584-47dd-8488-87d98c1e22f0-pod-info\") pod \"0788ca8d-f584-47dd-8488-87d98c1e22f0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.510122 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0788ca8d-f584-47dd-8488-87d98c1e22f0-plugins-conf\") pod \"0788ca8d-f584-47dd-8488-87d98c1e22f0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.510223 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\") pod \"0788ca8d-f584-47dd-8488-87d98c1e22f0\" (UID: \"0788ca8d-f584-47dd-8488-87d98c1e22f0\") " Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.511169 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0788ca8d-f584-47dd-8488-87d98c1e22f0-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0788ca8d-f584-47dd-8488-87d98c1e22f0" (UID: "0788ca8d-f584-47dd-8488-87d98c1e22f0"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.513595 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0788ca8d-f584-47dd-8488-87d98c1e22f0" (UID: "0788ca8d-f584-47dd-8488-87d98c1e22f0"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.513921 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0788ca8d-f584-47dd-8488-87d98c1e22f0" (UID: "0788ca8d-f584-47dd-8488-87d98c1e22f0"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.514379 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0788ca8d-f584-47dd-8488-87d98c1e22f0" (UID: "0788ca8d-f584-47dd-8488-87d98c1e22f0"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.517015 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0788ca8d-f584-47dd-8488-87d98c1e22f0-pod-info" (OuterVolumeSpecName: "pod-info") pod "0788ca8d-f584-47dd-8488-87d98c1e22f0" (UID: "0788ca8d-f584-47dd-8488-87d98c1e22f0"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.517117 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0788ca8d-f584-47dd-8488-87d98c1e22f0-kube-api-access-9ltcv" (OuterVolumeSpecName: "kube-api-access-9ltcv") pod "0788ca8d-f584-47dd-8488-87d98c1e22f0" (UID: "0788ca8d-f584-47dd-8488-87d98c1e22f0"). InnerVolumeSpecName "kube-api-access-9ltcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.520664 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0788ca8d-f584-47dd-8488-87d98c1e22f0-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0788ca8d-f584-47dd-8488-87d98c1e22f0" (UID: "0788ca8d-f584-47dd-8488-87d98c1e22f0"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.554110 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0788ca8d-f584-47dd-8488-87d98c1e22f0-config-data" (OuterVolumeSpecName: "config-data") pod "0788ca8d-f584-47dd-8488-87d98c1e22f0" (UID: "0788ca8d-f584-47dd-8488-87d98c1e22f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.588122 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0788ca8d-f584-47dd-8488-87d98c1e22f0-server-conf" (OuterVolumeSpecName: "server-conf") pod "0788ca8d-f584-47dd-8488-87d98c1e22f0" (UID: "0788ca8d-f584-47dd-8488-87d98c1e22f0"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.614415 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0788ca8d-f584-47dd-8488-87d98c1e22f0-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.614447 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.614462 4810 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0788ca8d-f584-47dd-8488-87d98c1e22f0-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.614494 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.614511 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ltcv\" (UniqueName: \"kubernetes.io/projected/0788ca8d-f584-47dd-8488-87d98c1e22f0-kube-api-access-9ltcv\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.614522 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.614532 4810 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0788ca8d-f584-47dd-8488-87d98c1e22f0-server-conf\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.614542 4810 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0788ca8d-f584-47dd-8488-87d98c1e22f0-pod-info\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.614554 4810 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0788ca8d-f584-47dd-8488-87d98c1e22f0-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.630722 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0788ca8d-f584-47dd-8488-87d98c1e22f0" (UID: "0788ca8d-f584-47dd-8488-87d98c1e22f0"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.702327 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5" (OuterVolumeSpecName: "persistence") pod "0788ca8d-f584-47dd-8488-87d98c1e22f0" (UID: "0788ca8d-f584-47dd-8488-87d98c1e22f0"). InnerVolumeSpecName "pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.715940 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0788ca8d-f584-47dd-8488-87d98c1e22f0-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.716004 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\") on node \"crc\" " Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.735810 4810 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.736015 4810 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5") on node "crc" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.817428 4810 reconciler_common.go:293] "Volume detached for volume \"pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.896370 4810 generic.go:334] "Generic (PLEG): container finished" podID="090f0e84-a3f5-4706-872f-b62086e22577" containerID="830d2f4682ce22e2fcbbe9ba3ba054d64a9347e2e7694441c75add8d7c9c3f30" exitCode=0 Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.896460 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"090f0e84-a3f5-4706-872f-b62086e22577","Type":"ContainerDied","Data":"830d2f4682ce22e2fcbbe9ba3ba054d64a9347e2e7694441c75add8d7c9c3f30"} Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.900338 4810 generic.go:334] "Generic (PLEG): container finished" podID="0788ca8d-f584-47dd-8488-87d98c1e22f0" containerID="4794e0892c96e9d985051845f733fda1c06719545108d6ebcafb6662f600f4a8" exitCode=0 Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.900482 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0788ca8d-f584-47dd-8488-87d98c1e22f0","Type":"ContainerDied","Data":"4794e0892c96e9d985051845f733fda1c06719545108d6ebcafb6662f600f4a8"} Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.900561 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.900634 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0788ca8d-f584-47dd-8488-87d98c1e22f0","Type":"ContainerDied","Data":"6858861ea0b81d0f555cafb73ab5a963e6881e7d320dd72dc480720d1813936b"} Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.900665 4810 scope.go:117] "RemoveContainer" containerID="4794e0892c96e9d985051845f733fda1c06719545108d6ebcafb6662f600f4a8" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.920876 4810 scope.go:117] "RemoveContainer" containerID="e2454a338a9d1bb4e69ba8caeb7fb292da7aa5cf3876c0272796812e2600841d" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.949157 4810 scope.go:117] "RemoveContainer" containerID="4794e0892c96e9d985051845f733fda1c06719545108d6ebcafb6662f600f4a8" Dec 01 16:00:55 crc kubenswrapper[4810]: E1201 16:00:55.952662 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4794e0892c96e9d985051845f733fda1c06719545108d6ebcafb6662f600f4a8\": container with ID starting with 4794e0892c96e9d985051845f733fda1c06719545108d6ebcafb6662f600f4a8 not found: ID does not exist" containerID="4794e0892c96e9d985051845f733fda1c06719545108d6ebcafb6662f600f4a8" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.952702 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4794e0892c96e9d985051845f733fda1c06719545108d6ebcafb6662f600f4a8"} err="failed to get container status \"4794e0892c96e9d985051845f733fda1c06719545108d6ebcafb6662f600f4a8\": rpc error: code = NotFound desc = could not find container \"4794e0892c96e9d985051845f733fda1c06719545108d6ebcafb6662f600f4a8\": container with ID starting with 4794e0892c96e9d985051845f733fda1c06719545108d6ebcafb6662f600f4a8 not found: ID does not exist" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.952730 4810 scope.go:117] "RemoveContainer" containerID="e2454a338a9d1bb4e69ba8caeb7fb292da7aa5cf3876c0272796812e2600841d" Dec 01 16:00:55 crc kubenswrapper[4810]: E1201 16:00:55.953137 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2454a338a9d1bb4e69ba8caeb7fb292da7aa5cf3876c0272796812e2600841d\": container with ID starting with e2454a338a9d1bb4e69ba8caeb7fb292da7aa5cf3876c0272796812e2600841d not found: ID does not exist" containerID="e2454a338a9d1bb4e69ba8caeb7fb292da7aa5cf3876c0272796812e2600841d" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.953176 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2454a338a9d1bb4e69ba8caeb7fb292da7aa5cf3876c0272796812e2600841d"} err="failed to get container status \"e2454a338a9d1bb4e69ba8caeb7fb292da7aa5cf3876c0272796812e2600841d\": rpc error: code = NotFound desc = could not find container \"e2454a338a9d1bb4e69ba8caeb7fb292da7aa5cf3876c0272796812e2600841d\": container with ID starting with e2454a338a9d1bb4e69ba8caeb7fb292da7aa5cf3876c0272796812e2600841d not found: ID does not exist" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.953288 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xnpkl" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.959986 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.968344 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.984579 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 16:00:55 crc kubenswrapper[4810]: E1201 16:00:55.985017 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9710aadb-4ea9-4bd5-bbaa-ce3d474c9021" containerName="init" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.985030 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9710aadb-4ea9-4bd5-bbaa-ce3d474c9021" containerName="init" Dec 01 16:00:55 crc kubenswrapper[4810]: E1201 16:00:55.985046 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9710aadb-4ea9-4bd5-bbaa-ce3d474c9021" containerName="dnsmasq-dns" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.985053 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9710aadb-4ea9-4bd5-bbaa-ce3d474c9021" containerName="dnsmasq-dns" Dec 01 16:00:55 crc kubenswrapper[4810]: E1201 16:00:55.985071 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0788ca8d-f584-47dd-8488-87d98c1e22f0" containerName="setup-container" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.985078 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0788ca8d-f584-47dd-8488-87d98c1e22f0" containerName="setup-container" Dec 01 16:00:55 crc kubenswrapper[4810]: E1201 16:00:55.985096 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0788ca8d-f584-47dd-8488-87d98c1e22f0" containerName="rabbitmq" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.985102 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0788ca8d-f584-47dd-8488-87d98c1e22f0" containerName="rabbitmq" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.985248 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9710aadb-4ea9-4bd5-bbaa-ce3d474c9021" containerName="dnsmasq-dns" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.985264 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0788ca8d-f584-47dd-8488-87d98c1e22f0" containerName="rabbitmq" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.986031 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.991941 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.992121 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.992226 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.992417 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.992600 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-mgw2x" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.992699 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 01 16:00:55 crc kubenswrapper[4810]: I1201 16:00:55.992803 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.003251 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.027238 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xnpkl"] Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.127393 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.127460 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fc454c84-0d74-446b-a396-3b0185e716a2-config-data\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.127514 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fc454c84-0d74-446b-a396-3b0185e716a2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.127546 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fc454c84-0d74-446b-a396-3b0185e716a2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.127566 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fc454c84-0d74-446b-a396-3b0185e716a2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.127663 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bltdj\" (UniqueName: \"kubernetes.io/projected/fc454c84-0d74-446b-a396-3b0185e716a2-kube-api-access-bltdj\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.127800 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fc454c84-0d74-446b-a396-3b0185e716a2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.127909 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fc454c84-0d74-446b-a396-3b0185e716a2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.127934 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fc454c84-0d74-446b-a396-3b0185e716a2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.127953 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fc454c84-0d74-446b-a396-3b0185e716a2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.127975 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fc454c84-0d74-446b-a396-3b0185e716a2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.182553 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.229750 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fc454c84-0d74-446b-a396-3b0185e716a2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.229800 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bltdj\" (UniqueName: \"kubernetes.io/projected/fc454c84-0d74-446b-a396-3b0185e716a2-kube-api-access-bltdj\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.229852 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fc454c84-0d74-446b-a396-3b0185e716a2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.229882 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fc454c84-0d74-446b-a396-3b0185e716a2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.229901 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fc454c84-0d74-446b-a396-3b0185e716a2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.229922 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fc454c84-0d74-446b-a396-3b0185e716a2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.229947 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fc454c84-0d74-446b-a396-3b0185e716a2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.229970 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.230004 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fc454c84-0d74-446b-a396-3b0185e716a2-config-data\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.230030 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fc454c84-0d74-446b-a396-3b0185e716a2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.230059 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fc454c84-0d74-446b-a396-3b0185e716a2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.230785 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fc454c84-0d74-446b-a396-3b0185e716a2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.230876 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fc454c84-0d74-446b-a396-3b0185e716a2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.231146 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fc454c84-0d74-446b-a396-3b0185e716a2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.232975 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fc454c84-0d74-446b-a396-3b0185e716a2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.233485 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fc454c84-0d74-446b-a396-3b0185e716a2-config-data\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.235572 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.235622 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/61d3d42dde16e89ed7ad20c90a916e70e853d49d4b65a0dd42582d970d13e064/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.235869 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fc454c84-0d74-446b-a396-3b0185e716a2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.235926 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fc454c84-0d74-446b-a396-3b0185e716a2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.240784 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fc454c84-0d74-446b-a396-3b0185e716a2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.242794 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fc454c84-0d74-446b-a396-3b0185e716a2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.262957 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bltdj\" (UniqueName: \"kubernetes.io/projected/fc454c84-0d74-446b-a396-3b0185e716a2-kube-api-access-bltdj\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.280506 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e75b38b7-f141-4654-855c-82d2f78a9ab5\") pod \"rabbitmq-server-0\" (UID: \"fc454c84-0d74-446b-a396-3b0185e716a2\") " pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.307650 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.330918 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2vrk\" (UniqueName: \"kubernetes.io/projected/090f0e84-a3f5-4706-872f-b62086e22577-kube-api-access-s2vrk\") pod \"090f0e84-a3f5-4706-872f-b62086e22577\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.331137 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\") pod \"090f0e84-a3f5-4706-872f-b62086e22577\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.331175 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/090f0e84-a3f5-4706-872f-b62086e22577-config-data\") pod \"090f0e84-a3f5-4706-872f-b62086e22577\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.331270 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/090f0e84-a3f5-4706-872f-b62086e22577-plugins-conf\") pod \"090f0e84-a3f5-4706-872f-b62086e22577\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.331326 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-plugins\") pod \"090f0e84-a3f5-4706-872f-b62086e22577\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.331342 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/090f0e84-a3f5-4706-872f-b62086e22577-server-conf\") pod \"090f0e84-a3f5-4706-872f-b62086e22577\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.331388 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-erlang-cookie\") pod \"090f0e84-a3f5-4706-872f-b62086e22577\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.331680 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-tls\") pod \"090f0e84-a3f5-4706-872f-b62086e22577\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.331708 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-confd\") pod \"090f0e84-a3f5-4706-872f-b62086e22577\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.331745 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/090f0e84-a3f5-4706-872f-b62086e22577-pod-info\") pod \"090f0e84-a3f5-4706-872f-b62086e22577\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.331777 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/090f0e84-a3f5-4706-872f-b62086e22577-erlang-cookie-secret\") pod \"090f0e84-a3f5-4706-872f-b62086e22577\" (UID: \"090f0e84-a3f5-4706-872f-b62086e22577\") " Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.332058 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/090f0e84-a3f5-4706-872f-b62086e22577-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "090f0e84-a3f5-4706-872f-b62086e22577" (UID: "090f0e84-a3f5-4706-872f-b62086e22577"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.332701 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "090f0e84-a3f5-4706-872f-b62086e22577" (UID: "090f0e84-a3f5-4706-872f-b62086e22577"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.332717 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "090f0e84-a3f5-4706-872f-b62086e22577" (UID: "090f0e84-a3f5-4706-872f-b62086e22577"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.337164 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/090f0e84-a3f5-4706-872f-b62086e22577-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "090f0e84-a3f5-4706-872f-b62086e22577" (UID: "090f0e84-a3f5-4706-872f-b62086e22577"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.338387 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "090f0e84-a3f5-4706-872f-b62086e22577" (UID: "090f0e84-a3f5-4706-872f-b62086e22577"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.338647 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/090f0e84-a3f5-4706-872f-b62086e22577-pod-info" (OuterVolumeSpecName: "pod-info") pod "090f0e84-a3f5-4706-872f-b62086e22577" (UID: "090f0e84-a3f5-4706-872f-b62086e22577"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.346545 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/090f0e84-a3f5-4706-872f-b62086e22577-kube-api-access-s2vrk" (OuterVolumeSpecName: "kube-api-access-s2vrk") pod "090f0e84-a3f5-4706-872f-b62086e22577" (UID: "090f0e84-a3f5-4706-872f-b62086e22577"). InnerVolumeSpecName "kube-api-access-s2vrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.351726 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/090f0e84-a3f5-4706-872f-b62086e22577-config-data" (OuterVolumeSpecName: "config-data") pod "090f0e84-a3f5-4706-872f-b62086e22577" (UID: "090f0e84-a3f5-4706-872f-b62086e22577"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.359336 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8" (OuterVolumeSpecName: "persistence") pod "090f0e84-a3f5-4706-872f-b62086e22577" (UID: "090f0e84-a3f5-4706-872f-b62086e22577"). InnerVolumeSpecName "pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.372816 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/090f0e84-a3f5-4706-872f-b62086e22577-server-conf" (OuterVolumeSpecName: "server-conf") pod "090f0e84-a3f5-4706-872f-b62086e22577" (UID: "090f0e84-a3f5-4706-872f-b62086e22577"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.427633 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "090f0e84-a3f5-4706-872f-b62086e22577" (UID: "090f0e84-a3f5-4706-872f-b62086e22577"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.434678 4810 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/090f0e84-a3f5-4706-872f-b62086e22577-pod-info\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.434729 4810 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/090f0e84-a3f5-4706-872f-b62086e22577-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.434747 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2vrk\" (UniqueName: \"kubernetes.io/projected/090f0e84-a3f5-4706-872f-b62086e22577-kube-api-access-s2vrk\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.434798 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\") on node \"crc\" " Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.434920 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/090f0e84-a3f5-4706-872f-b62086e22577-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.434934 4810 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/090f0e84-a3f5-4706-872f-b62086e22577-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.434946 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.434964 4810 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/090f0e84-a3f5-4706-872f-b62086e22577-server-conf\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.434977 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.434988 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.434999 4810 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/090f0e84-a3f5-4706-872f-b62086e22577-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.454627 4810 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.454827 4810 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8") on node "crc" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.502600 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0788ca8d-f584-47dd-8488-87d98c1e22f0" path="/var/lib/kubelet/pods/0788ca8d-f584-47dd-8488-87d98c1e22f0/volumes" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.503174 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9710aadb-4ea9-4bd5-bbaa-ce3d474c9021" path="/var/lib/kubelet/pods/9710aadb-4ea9-4bd5-bbaa-ce3d474c9021/volumes" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.535871 4810 reconciler_common.go:293] "Volume detached for volume \"pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.747439 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 16:00:56 crc kubenswrapper[4810]: W1201 16:00:56.752162 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc454c84_0d74_446b_a396_3b0185e716a2.slice/crio-577ce3e48fd1522240053e4c0393961941b4871d8f3d9e61c90d6213410eb7bd WatchSource:0}: Error finding container 577ce3e48fd1522240053e4c0393961941b4871d8f3d9e61c90d6213410eb7bd: Status 404 returned error can't find the container with id 577ce3e48fd1522240053e4c0393961941b4871d8f3d9e61c90d6213410eb7bd Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.910450 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"fc454c84-0d74-446b-a396-3b0185e716a2","Type":"ContainerStarted","Data":"577ce3e48fd1522240053e4c0393961941b4871d8f3d9e61c90d6213410eb7bd"} Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.915084 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.915594 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"090f0e84-a3f5-4706-872f-b62086e22577","Type":"ContainerDied","Data":"ee01afe9d7b912f957c8355a5950bab0e03a2ef4cd20508c5c59a1d151d2ac24"} Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.915628 4810 scope.go:117] "RemoveContainer" containerID="830d2f4682ce22e2fcbbe9ba3ba054d64a9347e2e7694441c75add8d7c9c3f30" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.952505 4810 scope.go:117] "RemoveContainer" containerID="e724435386ae94fcebeff126f3b3387876d53bb30199d1134178eb3f74e49bf2" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.966782 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.978911 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.986510 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 16:00:56 crc kubenswrapper[4810]: E1201 16:00:56.987135 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="090f0e84-a3f5-4706-872f-b62086e22577" containerName="setup-container" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.987232 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="090f0e84-a3f5-4706-872f-b62086e22577" containerName="setup-container" Dec 01 16:00:56 crc kubenswrapper[4810]: E1201 16:00:56.987335 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="090f0e84-a3f5-4706-872f-b62086e22577" containerName="rabbitmq" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.987424 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="090f0e84-a3f5-4706-872f-b62086e22577" containerName="rabbitmq" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.987715 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="090f0e84-a3f5-4706-872f-b62086e22577" containerName="rabbitmq" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.989356 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.992925 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.993292 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.994030 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.994282 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.994677 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.994836 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-dkzvm" Dec 01 16:00:56 crc kubenswrapper[4810]: I1201 16:00:56.995089 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.003404 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.051099 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8d4f1ab5-87de-41d0-877c-8ff39bdff385-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.051163 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8d4f1ab5-87de-41d0-877c-8ff39bdff385-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.051193 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8d4f1ab5-87de-41d0-877c-8ff39bdff385-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.051220 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8d4f1ab5-87de-41d0-877c-8ff39bdff385-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.051244 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.051279 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8d4f1ab5-87de-41d0-877c-8ff39bdff385-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.051302 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8d4f1ab5-87de-41d0-877c-8ff39bdff385-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.051326 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8d4f1ab5-87de-41d0-877c-8ff39bdff385-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.051355 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b9kw\" (UniqueName: \"kubernetes.io/projected/8d4f1ab5-87de-41d0-877c-8ff39bdff385-kube-api-access-6b9kw\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.051411 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8d4f1ab5-87de-41d0-877c-8ff39bdff385-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.051430 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8d4f1ab5-87de-41d0-877c-8ff39bdff385-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.152839 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.152914 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8d4f1ab5-87de-41d0-877c-8ff39bdff385-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.152977 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8d4f1ab5-87de-41d0-877c-8ff39bdff385-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.153683 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8d4f1ab5-87de-41d0-877c-8ff39bdff385-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.153780 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b9kw\" (UniqueName: \"kubernetes.io/projected/8d4f1ab5-87de-41d0-877c-8ff39bdff385-kube-api-access-6b9kw\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.153857 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8d4f1ab5-87de-41d0-877c-8ff39bdff385-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.153889 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8d4f1ab5-87de-41d0-877c-8ff39bdff385-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.153941 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8d4f1ab5-87de-41d0-877c-8ff39bdff385-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.154008 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8d4f1ab5-87de-41d0-877c-8ff39bdff385-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.154036 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8d4f1ab5-87de-41d0-877c-8ff39bdff385-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.154070 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8d4f1ab5-87de-41d0-877c-8ff39bdff385-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.154113 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8d4f1ab5-87de-41d0-877c-8ff39bdff385-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.154377 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8d4f1ab5-87de-41d0-877c-8ff39bdff385-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.154995 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8d4f1ab5-87de-41d0-877c-8ff39bdff385-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.155194 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8d4f1ab5-87de-41d0-877c-8ff39bdff385-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.155305 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8d4f1ab5-87de-41d0-877c-8ff39bdff385-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.156505 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.156545 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c8c54e25171899303fbcbbb2f3f9c41dea8d63d5254c746c192fb7c677ca914d/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.157680 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8d4f1ab5-87de-41d0-877c-8ff39bdff385-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.159209 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8d4f1ab5-87de-41d0-877c-8ff39bdff385-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.159572 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8d4f1ab5-87de-41d0-877c-8ff39bdff385-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.164384 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8d4f1ab5-87de-41d0-877c-8ff39bdff385-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.180631 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b9kw\" (UniqueName: \"kubernetes.io/projected/8d4f1ab5-87de-41d0-877c-8ff39bdff385-kube-api-access-6b9kw\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.184117 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f49fa175-3816-4ea3-b149-d0ac466bc4b8\") pod \"rabbitmq-cell1-server-0\" (UID: \"8d4f1ab5-87de-41d0-877c-8ff39bdff385\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.308174 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.533881 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 16:00:57 crc kubenswrapper[4810]: W1201 16:00:57.556975 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d4f1ab5_87de_41d0_877c_8ff39bdff385.slice/crio-bb2de6bbee0af72cc9d045019c4e1e6106192077d7c604e7138d46556abf7e57 WatchSource:0}: Error finding container bb2de6bbee0af72cc9d045019c4e1e6106192077d7c604e7138d46556abf7e57: Status 404 returned error can't find the container with id bb2de6bbee0af72cc9d045019c4e1e6106192077d7c604e7138d46556abf7e57 Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.922309 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8d4f1ab5-87de-41d0-877c-8ff39bdff385","Type":"ContainerStarted","Data":"bb2de6bbee0af72cc9d045019c4e1e6106192077d7c604e7138d46556abf7e57"} Dec 01 16:00:57 crc kubenswrapper[4810]: I1201 16:00:57.922499 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xnpkl" podUID="bfc86968-4a25-41c5-9612-d0f5570bd122" containerName="registry-server" containerID="cri-o://2b827f8ab11bbe2f0d2e72802614d10898e91971a7a1ef660bf07f94eb0114db" gracePeriod=2 Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.355068 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xnpkl" Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.480964 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc86968-4a25-41c5-9612-d0f5570bd122-utilities\") pod \"bfc86968-4a25-41c5-9612-d0f5570bd122\" (UID: \"bfc86968-4a25-41c5-9612-d0f5570bd122\") " Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.481012 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc86968-4a25-41c5-9612-d0f5570bd122-catalog-content\") pod \"bfc86968-4a25-41c5-9612-d0f5570bd122\" (UID: \"bfc86968-4a25-41c5-9612-d0f5570bd122\") " Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.481112 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwvlv\" (UniqueName: \"kubernetes.io/projected/bfc86968-4a25-41c5-9612-d0f5570bd122-kube-api-access-pwvlv\") pod \"bfc86968-4a25-41c5-9612-d0f5570bd122\" (UID: \"bfc86968-4a25-41c5-9612-d0f5570bd122\") " Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.482327 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfc86968-4a25-41c5-9612-d0f5570bd122-utilities" (OuterVolumeSpecName: "utilities") pod "bfc86968-4a25-41c5-9612-d0f5570bd122" (UID: "bfc86968-4a25-41c5-9612-d0f5570bd122"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.486129 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfc86968-4a25-41c5-9612-d0f5570bd122-kube-api-access-pwvlv" (OuterVolumeSpecName: "kube-api-access-pwvlv") pod "bfc86968-4a25-41c5-9612-d0f5570bd122" (UID: "bfc86968-4a25-41c5-9612-d0f5570bd122"). InnerVolumeSpecName "kube-api-access-pwvlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.501444 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="090f0e84-a3f5-4706-872f-b62086e22577" path="/var/lib/kubelet/pods/090f0e84-a3f5-4706-872f-b62086e22577/volumes" Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.513009 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfc86968-4a25-41c5-9612-d0f5570bd122-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bfc86968-4a25-41c5-9612-d0f5570bd122" (UID: "bfc86968-4a25-41c5-9612-d0f5570bd122"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.582568 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfc86968-4a25-41c5-9612-d0f5570bd122-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.582894 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfc86968-4a25-41c5-9612-d0f5570bd122-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.582910 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwvlv\" (UniqueName: \"kubernetes.io/projected/bfc86968-4a25-41c5-9612-d0f5570bd122-kube-api-access-pwvlv\") on node \"crc\" DevicePath \"\"" Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.931765 4810 generic.go:334] "Generic (PLEG): container finished" podID="bfc86968-4a25-41c5-9612-d0f5570bd122" containerID="2b827f8ab11bbe2f0d2e72802614d10898e91971a7a1ef660bf07f94eb0114db" exitCode=0 Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.931827 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xnpkl" event={"ID":"bfc86968-4a25-41c5-9612-d0f5570bd122","Type":"ContainerDied","Data":"2b827f8ab11bbe2f0d2e72802614d10898e91971a7a1ef660bf07f94eb0114db"} Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.931859 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xnpkl" event={"ID":"bfc86968-4a25-41c5-9612-d0f5570bd122","Type":"ContainerDied","Data":"bc4acc9bd94ec87e2bf135cf85cd2077b569d24ea356f9d1d21a5316392a0d13"} Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.931881 4810 scope.go:117] "RemoveContainer" containerID="2b827f8ab11bbe2f0d2e72802614d10898e91971a7a1ef660bf07f94eb0114db" Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.931982 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xnpkl" Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.949838 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"fc454c84-0d74-446b-a396-3b0185e716a2","Type":"ContainerStarted","Data":"63cf8983550ff06197e54a87d22101bbf8f050f91103b6a3138561c5df0b0c4c"} Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.972814 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xnpkl"] Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.975526 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xnpkl"] Dec 01 16:00:58 crc kubenswrapper[4810]: I1201 16:00:58.975724 4810 scope.go:117] "RemoveContainer" containerID="9340620f3eea5359ba4cbcc22b12e3fdf0487a0b761821583e642a0e98ee933d" Dec 01 16:00:59 crc kubenswrapper[4810]: I1201 16:00:59.131985 4810 scope.go:117] "RemoveContainer" containerID="3e36f9a660690c5334a6d7b01cc1f21ff9c4d9a4472307d4eb45c5c05e47229d" Dec 01 16:00:59 crc kubenswrapper[4810]: I1201 16:00:59.162184 4810 scope.go:117] "RemoveContainer" containerID="2b827f8ab11bbe2f0d2e72802614d10898e91971a7a1ef660bf07f94eb0114db" Dec 01 16:00:59 crc kubenswrapper[4810]: E1201 16:00:59.162540 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b827f8ab11bbe2f0d2e72802614d10898e91971a7a1ef660bf07f94eb0114db\": container with ID starting with 2b827f8ab11bbe2f0d2e72802614d10898e91971a7a1ef660bf07f94eb0114db not found: ID does not exist" containerID="2b827f8ab11bbe2f0d2e72802614d10898e91971a7a1ef660bf07f94eb0114db" Dec 01 16:00:59 crc kubenswrapper[4810]: I1201 16:00:59.162567 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b827f8ab11bbe2f0d2e72802614d10898e91971a7a1ef660bf07f94eb0114db"} err="failed to get container status \"2b827f8ab11bbe2f0d2e72802614d10898e91971a7a1ef660bf07f94eb0114db\": rpc error: code = NotFound desc = could not find container \"2b827f8ab11bbe2f0d2e72802614d10898e91971a7a1ef660bf07f94eb0114db\": container with ID starting with 2b827f8ab11bbe2f0d2e72802614d10898e91971a7a1ef660bf07f94eb0114db not found: ID does not exist" Dec 01 16:00:59 crc kubenswrapper[4810]: I1201 16:00:59.162587 4810 scope.go:117] "RemoveContainer" containerID="9340620f3eea5359ba4cbcc22b12e3fdf0487a0b761821583e642a0e98ee933d" Dec 01 16:00:59 crc kubenswrapper[4810]: E1201 16:00:59.162835 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9340620f3eea5359ba4cbcc22b12e3fdf0487a0b761821583e642a0e98ee933d\": container with ID starting with 9340620f3eea5359ba4cbcc22b12e3fdf0487a0b761821583e642a0e98ee933d not found: ID does not exist" containerID="9340620f3eea5359ba4cbcc22b12e3fdf0487a0b761821583e642a0e98ee933d" Dec 01 16:00:59 crc kubenswrapper[4810]: I1201 16:00:59.162858 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9340620f3eea5359ba4cbcc22b12e3fdf0487a0b761821583e642a0e98ee933d"} err="failed to get container status \"9340620f3eea5359ba4cbcc22b12e3fdf0487a0b761821583e642a0e98ee933d\": rpc error: code = NotFound desc = could not find container \"9340620f3eea5359ba4cbcc22b12e3fdf0487a0b761821583e642a0e98ee933d\": container with ID starting with 9340620f3eea5359ba4cbcc22b12e3fdf0487a0b761821583e642a0e98ee933d not found: ID does not exist" Dec 01 16:00:59 crc kubenswrapper[4810]: I1201 16:00:59.162871 4810 scope.go:117] "RemoveContainer" containerID="3e36f9a660690c5334a6d7b01cc1f21ff9c4d9a4472307d4eb45c5c05e47229d" Dec 01 16:00:59 crc kubenswrapper[4810]: E1201 16:00:59.163185 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e36f9a660690c5334a6d7b01cc1f21ff9c4d9a4472307d4eb45c5c05e47229d\": container with ID starting with 3e36f9a660690c5334a6d7b01cc1f21ff9c4d9a4472307d4eb45c5c05e47229d not found: ID does not exist" containerID="3e36f9a660690c5334a6d7b01cc1f21ff9c4d9a4472307d4eb45c5c05e47229d" Dec 01 16:00:59 crc kubenswrapper[4810]: I1201 16:00:59.163204 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e36f9a660690c5334a6d7b01cc1f21ff9c4d9a4472307d4eb45c5c05e47229d"} err="failed to get container status \"3e36f9a660690c5334a6d7b01cc1f21ff9c4d9a4472307d4eb45c5c05e47229d\": rpc error: code = NotFound desc = could not find container \"3e36f9a660690c5334a6d7b01cc1f21ff9c4d9a4472307d4eb45c5c05e47229d\": container with ID starting with 3e36f9a660690c5334a6d7b01cc1f21ff9c4d9a4472307d4eb45c5c05e47229d not found: ID does not exist" Dec 01 16:00:59 crc kubenswrapper[4810]: I1201 16:00:59.958724 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8d4f1ab5-87de-41d0-877c-8ff39bdff385","Type":"ContainerStarted","Data":"a7fa70b65d1e760ba5f6e27d5707666b8e548108fdf3c7f2051738ae9bb37296"} Dec 01 16:01:00 crc kubenswrapper[4810]: I1201 16:01:00.503304 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfc86968-4a25-41c5-9612-d0f5570bd122" path="/var/lib/kubelet/pods/bfc86968-4a25-41c5-9612-d0f5570bd122/volumes" Dec 01 16:01:02 crc kubenswrapper[4810]: I1201 16:01:02.972677 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:01:02 crc kubenswrapper[4810]: I1201 16:01:02.973023 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:01:30 crc kubenswrapper[4810]: I1201 16:01:30.193368 4810 generic.go:334] "Generic (PLEG): container finished" podID="fc454c84-0d74-446b-a396-3b0185e716a2" containerID="63cf8983550ff06197e54a87d22101bbf8f050f91103b6a3138561c5df0b0c4c" exitCode=0 Dec 01 16:01:30 crc kubenswrapper[4810]: I1201 16:01:30.193517 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"fc454c84-0d74-446b-a396-3b0185e716a2","Type":"ContainerDied","Data":"63cf8983550ff06197e54a87d22101bbf8f050f91103b6a3138561c5df0b0c4c"} Dec 01 16:01:31 crc kubenswrapper[4810]: I1201 16:01:31.202886 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"fc454c84-0d74-446b-a396-3b0185e716a2","Type":"ContainerStarted","Data":"4c7a1b343d9c20f20e4bac4160287f614ff0410d57f5ddc1014d90a4c876367b"} Dec 01 16:01:31 crc kubenswrapper[4810]: I1201 16:01:31.203449 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 01 16:01:31 crc kubenswrapper[4810]: I1201 16:01:31.204667 4810 generic.go:334] "Generic (PLEG): container finished" podID="8d4f1ab5-87de-41d0-877c-8ff39bdff385" containerID="a7fa70b65d1e760ba5f6e27d5707666b8e548108fdf3c7f2051738ae9bb37296" exitCode=0 Dec 01 16:01:31 crc kubenswrapper[4810]: I1201 16:01:31.204706 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8d4f1ab5-87de-41d0-877c-8ff39bdff385","Type":"ContainerDied","Data":"a7fa70b65d1e760ba5f6e27d5707666b8e548108fdf3c7f2051738ae9bb37296"} Dec 01 16:01:31 crc kubenswrapper[4810]: I1201 16:01:31.224767 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.224750152 podStartE2EDuration="36.224750152s" podCreationTimestamp="2025-12-01 16:00:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:01:31.222893252 +0000 UTC m=+5256.986402875" watchObservedRunningTime="2025-12-01 16:01:31.224750152 +0000 UTC m=+5256.988259755" Dec 01 16:01:32 crc kubenswrapper[4810]: I1201 16:01:32.214629 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"8d4f1ab5-87de-41d0-877c-8ff39bdff385","Type":"ContainerStarted","Data":"8d0c24d354cc35aee962415b13ef691dcf16f06d6a47fe026e07e16cc2667440"} Dec 01 16:01:32 crc kubenswrapper[4810]: I1201 16:01:32.215449 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:01:32 crc kubenswrapper[4810]: I1201 16:01:32.245025 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.245004578 podStartE2EDuration="36.245004578s" podCreationTimestamp="2025-12-01 16:00:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:01:32.235706915 +0000 UTC m=+5257.999216528" watchObservedRunningTime="2025-12-01 16:01:32.245004578 +0000 UTC m=+5258.008514171" Dec 01 16:01:32 crc kubenswrapper[4810]: I1201 16:01:32.972070 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:01:32 crc kubenswrapper[4810]: I1201 16:01:32.972448 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:01:46 crc kubenswrapper[4810]: I1201 16:01:46.311651 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 01 16:01:47 crc kubenswrapper[4810]: I1201 16:01:47.312008 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 01 16:01:51 crc kubenswrapper[4810]: I1201 16:01:51.124964 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Dec 01 16:01:51 crc kubenswrapper[4810]: E1201 16:01:51.126199 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfc86968-4a25-41c5-9612-d0f5570bd122" containerName="extract-content" Dec 01 16:01:51 crc kubenswrapper[4810]: I1201 16:01:51.126218 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfc86968-4a25-41c5-9612-d0f5570bd122" containerName="extract-content" Dec 01 16:01:51 crc kubenswrapper[4810]: E1201 16:01:51.126247 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfc86968-4a25-41c5-9612-d0f5570bd122" containerName="registry-server" Dec 01 16:01:51 crc kubenswrapper[4810]: I1201 16:01:51.126256 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfc86968-4a25-41c5-9612-d0f5570bd122" containerName="registry-server" Dec 01 16:01:51 crc kubenswrapper[4810]: E1201 16:01:51.126268 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfc86968-4a25-41c5-9612-d0f5570bd122" containerName="extract-utilities" Dec 01 16:01:51 crc kubenswrapper[4810]: I1201 16:01:51.126276 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfc86968-4a25-41c5-9612-d0f5570bd122" containerName="extract-utilities" Dec 01 16:01:51 crc kubenswrapper[4810]: I1201 16:01:51.126656 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfc86968-4a25-41c5-9612-d0f5570bd122" containerName="registry-server" Dec 01 16:01:51 crc kubenswrapper[4810]: I1201 16:01:51.127320 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 01 16:01:51 crc kubenswrapper[4810]: I1201 16:01:51.129873 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-xs2jm" Dec 01 16:01:51 crc kubenswrapper[4810]: I1201 16:01:51.132580 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 01 16:01:51 crc kubenswrapper[4810]: I1201 16:01:51.147946 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m255c\" (UniqueName: \"kubernetes.io/projected/cc9ca194-7250-44f4-840a-155240e1bd07-kube-api-access-m255c\") pod \"mariadb-client-1-default\" (UID: \"cc9ca194-7250-44f4-840a-155240e1bd07\") " pod="openstack/mariadb-client-1-default" Dec 01 16:01:51 crc kubenswrapper[4810]: I1201 16:01:51.250577 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m255c\" (UniqueName: \"kubernetes.io/projected/cc9ca194-7250-44f4-840a-155240e1bd07-kube-api-access-m255c\") pod \"mariadb-client-1-default\" (UID: \"cc9ca194-7250-44f4-840a-155240e1bd07\") " pod="openstack/mariadb-client-1-default" Dec 01 16:01:51 crc kubenswrapper[4810]: I1201 16:01:51.273700 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m255c\" (UniqueName: \"kubernetes.io/projected/cc9ca194-7250-44f4-840a-155240e1bd07-kube-api-access-m255c\") pod \"mariadb-client-1-default\" (UID: \"cc9ca194-7250-44f4-840a-155240e1bd07\") " pod="openstack/mariadb-client-1-default" Dec 01 16:01:51 crc kubenswrapper[4810]: I1201 16:01:51.448281 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 01 16:01:51 crc kubenswrapper[4810]: I1201 16:01:51.948864 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 01 16:01:52 crc kubenswrapper[4810]: I1201 16:01:52.367615 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"cc9ca194-7250-44f4-840a-155240e1bd07","Type":"ContainerStarted","Data":"d42da7ac954aedd77c03fac47eebc7e9a14ba5a6619af8cb467f7905d8020c6e"} Dec 01 16:01:56 crc kubenswrapper[4810]: I1201 16:01:56.398830 4810 generic.go:334] "Generic (PLEG): container finished" podID="cc9ca194-7250-44f4-840a-155240e1bd07" containerID="49ec02d5346e11e8f636f3169f4c6d7f580b7ea8185198a5cfc6b935e42c9ca3" exitCode=0 Dec 01 16:01:56 crc kubenswrapper[4810]: I1201 16:01:56.398914 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"cc9ca194-7250-44f4-840a-155240e1bd07","Type":"ContainerDied","Data":"49ec02d5346e11e8f636f3169f4c6d7f580b7ea8185198a5cfc6b935e42c9ca3"} Dec 01 16:01:57 crc kubenswrapper[4810]: I1201 16:01:57.745766 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 01 16:01:57 crc kubenswrapper[4810]: I1201 16:01:57.771428 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_cc9ca194-7250-44f4-840a-155240e1bd07/mariadb-client-1-default/0.log" Dec 01 16:01:57 crc kubenswrapper[4810]: I1201 16:01:57.803552 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 01 16:01:57 crc kubenswrapper[4810]: I1201 16:01:57.813619 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 01 16:01:57 crc kubenswrapper[4810]: I1201 16:01:57.946811 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m255c\" (UniqueName: \"kubernetes.io/projected/cc9ca194-7250-44f4-840a-155240e1bd07-kube-api-access-m255c\") pod \"cc9ca194-7250-44f4-840a-155240e1bd07\" (UID: \"cc9ca194-7250-44f4-840a-155240e1bd07\") " Dec 01 16:01:57 crc kubenswrapper[4810]: I1201 16:01:57.957815 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc9ca194-7250-44f4-840a-155240e1bd07-kube-api-access-m255c" (OuterVolumeSpecName: "kube-api-access-m255c") pod "cc9ca194-7250-44f4-840a-155240e1bd07" (UID: "cc9ca194-7250-44f4-840a-155240e1bd07"). InnerVolumeSpecName "kube-api-access-m255c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:01:58 crc kubenswrapper[4810]: I1201 16:01:58.050651 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m255c\" (UniqueName: \"kubernetes.io/projected/cc9ca194-7250-44f4-840a-155240e1bd07-kube-api-access-m255c\") on node \"crc\" DevicePath \"\"" Dec 01 16:01:58 crc kubenswrapper[4810]: I1201 16:01:58.333785 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Dec 01 16:01:58 crc kubenswrapper[4810]: E1201 16:01:58.334136 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc9ca194-7250-44f4-840a-155240e1bd07" containerName="mariadb-client-1-default" Dec 01 16:01:58 crc kubenswrapper[4810]: I1201 16:01:58.334157 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc9ca194-7250-44f4-840a-155240e1bd07" containerName="mariadb-client-1-default" Dec 01 16:01:58 crc kubenswrapper[4810]: I1201 16:01:58.334443 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc9ca194-7250-44f4-840a-155240e1bd07" containerName="mariadb-client-1-default" Dec 01 16:01:58 crc kubenswrapper[4810]: I1201 16:01:58.335177 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 01 16:01:58 crc kubenswrapper[4810]: I1201 16:01:58.345067 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 01 16:01:58 crc kubenswrapper[4810]: I1201 16:01:58.417503 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d42da7ac954aedd77c03fac47eebc7e9a14ba5a6619af8cb467f7905d8020c6e" Dec 01 16:01:58 crc kubenswrapper[4810]: I1201 16:01:58.417602 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 01 16:01:58 crc kubenswrapper[4810]: I1201 16:01:58.456691 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm79z\" (UniqueName: \"kubernetes.io/projected/28a8853e-5752-4e38-9f26-f0cd023a6d28-kube-api-access-qm79z\") pod \"mariadb-client-2-default\" (UID: \"28a8853e-5752-4e38-9f26-f0cd023a6d28\") " pod="openstack/mariadb-client-2-default" Dec 01 16:01:58 crc kubenswrapper[4810]: I1201 16:01:58.508624 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc9ca194-7250-44f4-840a-155240e1bd07" path="/var/lib/kubelet/pods/cc9ca194-7250-44f4-840a-155240e1bd07/volumes" Dec 01 16:01:58 crc kubenswrapper[4810]: I1201 16:01:58.559064 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm79z\" (UniqueName: \"kubernetes.io/projected/28a8853e-5752-4e38-9f26-f0cd023a6d28-kube-api-access-qm79z\") pod \"mariadb-client-2-default\" (UID: \"28a8853e-5752-4e38-9f26-f0cd023a6d28\") " pod="openstack/mariadb-client-2-default" Dec 01 16:01:58 crc kubenswrapper[4810]: I1201 16:01:58.575874 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm79z\" (UniqueName: \"kubernetes.io/projected/28a8853e-5752-4e38-9f26-f0cd023a6d28-kube-api-access-qm79z\") pod \"mariadb-client-2-default\" (UID: \"28a8853e-5752-4e38-9f26-f0cd023a6d28\") " pod="openstack/mariadb-client-2-default" Dec 01 16:01:58 crc kubenswrapper[4810]: I1201 16:01:58.656648 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 01 16:01:59 crc kubenswrapper[4810]: I1201 16:01:59.162018 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 01 16:01:59 crc kubenswrapper[4810]: I1201 16:01:59.428863 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"28a8853e-5752-4e38-9f26-f0cd023a6d28","Type":"ContainerStarted","Data":"d451256e7412bbd5f2eef5a3d23a3015d190d220b52d9b5f40ca6bd12e47d0cd"} Dec 01 16:01:59 crc kubenswrapper[4810]: I1201 16:01:59.429236 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"28a8853e-5752-4e38-9f26-f0cd023a6d28","Type":"ContainerStarted","Data":"6c08f688eefc4d84e10732797d5478c78e64627341fa6d32d95ab54704c272a6"} Dec 01 16:01:59 crc kubenswrapper[4810]: I1201 16:01:59.454136 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-2-default" podStartSLOduration=1.45410128 podStartE2EDuration="1.45410128s" podCreationTimestamp="2025-12-01 16:01:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:01:59.442785643 +0000 UTC m=+5285.206295276" watchObservedRunningTime="2025-12-01 16:01:59.45410128 +0000 UTC m=+5285.217610923" Dec 01 16:02:00 crc kubenswrapper[4810]: I1201 16:02:00.452608 4810 generic.go:334] "Generic (PLEG): container finished" podID="28a8853e-5752-4e38-9f26-f0cd023a6d28" containerID="d451256e7412bbd5f2eef5a3d23a3015d190d220b52d9b5f40ca6bd12e47d0cd" exitCode=1 Dec 01 16:02:00 crc kubenswrapper[4810]: I1201 16:02:00.452698 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"28a8853e-5752-4e38-9f26-f0cd023a6d28","Type":"ContainerDied","Data":"d451256e7412bbd5f2eef5a3d23a3015d190d220b52d9b5f40ca6bd12e47d0cd"} Dec 01 16:02:01 crc kubenswrapper[4810]: I1201 16:02:01.801682 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 01 16:02:01 crc kubenswrapper[4810]: I1201 16:02:01.851001 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 01 16:02:01 crc kubenswrapper[4810]: I1201 16:02:01.859139 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 01 16:02:01 crc kubenswrapper[4810]: I1201 16:02:01.910451 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qm79z\" (UniqueName: \"kubernetes.io/projected/28a8853e-5752-4e38-9f26-f0cd023a6d28-kube-api-access-qm79z\") pod \"28a8853e-5752-4e38-9f26-f0cd023a6d28\" (UID: \"28a8853e-5752-4e38-9f26-f0cd023a6d28\") " Dec 01 16:02:01 crc kubenswrapper[4810]: I1201 16:02:01.915755 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28a8853e-5752-4e38-9f26-f0cd023a6d28-kube-api-access-qm79z" (OuterVolumeSpecName: "kube-api-access-qm79z") pod "28a8853e-5752-4e38-9f26-f0cd023a6d28" (UID: "28a8853e-5752-4e38-9f26-f0cd023a6d28"). InnerVolumeSpecName "kube-api-access-qm79z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:02:02 crc kubenswrapper[4810]: I1201 16:02:02.011965 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qm79z\" (UniqueName: \"kubernetes.io/projected/28a8853e-5752-4e38-9f26-f0cd023a6d28-kube-api-access-qm79z\") on node \"crc\" DevicePath \"\"" Dec 01 16:02:02 crc kubenswrapper[4810]: I1201 16:02:02.339803 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Dec 01 16:02:02 crc kubenswrapper[4810]: E1201 16:02:02.340501 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28a8853e-5752-4e38-9f26-f0cd023a6d28" containerName="mariadb-client-2-default" Dec 01 16:02:02 crc kubenswrapper[4810]: I1201 16:02:02.340524 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="28a8853e-5752-4e38-9f26-f0cd023a6d28" containerName="mariadb-client-2-default" Dec 01 16:02:02 crc kubenswrapper[4810]: I1201 16:02:02.340778 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="28a8853e-5752-4e38-9f26-f0cd023a6d28" containerName="mariadb-client-2-default" Dec 01 16:02:02 crc kubenswrapper[4810]: I1201 16:02:02.341995 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 01 16:02:02 crc kubenswrapper[4810]: I1201 16:02:02.353798 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Dec 01 16:02:02 crc kubenswrapper[4810]: I1201 16:02:02.418006 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44497\" (UniqueName: \"kubernetes.io/projected/3ff88757-ea48-4e78-9452-d9f89ac909df-kube-api-access-44497\") pod \"mariadb-client-1\" (UID: \"3ff88757-ea48-4e78-9452-d9f89ac909df\") " pod="openstack/mariadb-client-1" Dec 01 16:02:02 crc kubenswrapper[4810]: I1201 16:02:02.468274 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c08f688eefc4d84e10732797d5478c78e64627341fa6d32d95ab54704c272a6" Dec 01 16:02:02 crc kubenswrapper[4810]: I1201 16:02:02.468346 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 01 16:02:02 crc kubenswrapper[4810]: I1201 16:02:02.500393 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28a8853e-5752-4e38-9f26-f0cd023a6d28" path="/var/lib/kubelet/pods/28a8853e-5752-4e38-9f26-f0cd023a6d28/volumes" Dec 01 16:02:02 crc kubenswrapper[4810]: I1201 16:02:02.519328 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44497\" (UniqueName: \"kubernetes.io/projected/3ff88757-ea48-4e78-9452-d9f89ac909df-kube-api-access-44497\") pod \"mariadb-client-1\" (UID: \"3ff88757-ea48-4e78-9452-d9f89ac909df\") " pod="openstack/mariadb-client-1" Dec 01 16:02:02 crc kubenswrapper[4810]: I1201 16:02:02.535463 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44497\" (UniqueName: \"kubernetes.io/projected/3ff88757-ea48-4e78-9452-d9f89ac909df-kube-api-access-44497\") pod \"mariadb-client-1\" (UID: \"3ff88757-ea48-4e78-9452-d9f89ac909df\") " pod="openstack/mariadb-client-1" Dec 01 16:02:02 crc kubenswrapper[4810]: I1201 16:02:02.659810 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 01 16:02:02 crc kubenswrapper[4810]: I1201 16:02:02.972881 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:02:02 crc kubenswrapper[4810]: I1201 16:02:02.973232 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:02:02 crc kubenswrapper[4810]: I1201 16:02:02.973286 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 16:02:02 crc kubenswrapper[4810]: I1201 16:02:02.974053 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:02:02 crc kubenswrapper[4810]: I1201 16:02:02.974124 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" gracePeriod=600 Dec 01 16:02:03 crc kubenswrapper[4810]: I1201 16:02:03.009914 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Dec 01 16:02:03 crc kubenswrapper[4810]: W1201 16:02:03.014204 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ff88757_ea48_4e78_9452_d9f89ac909df.slice/crio-7dacf92c084b0799c103d76e136f43e376ded8eb57efbf411bac8d29620a5f58 WatchSource:0}: Error finding container 7dacf92c084b0799c103d76e136f43e376ded8eb57efbf411bac8d29620a5f58: Status 404 returned error can't find the container with id 7dacf92c084b0799c103d76e136f43e376ded8eb57efbf411bac8d29620a5f58 Dec 01 16:02:03 crc kubenswrapper[4810]: I1201 16:02:03.476994 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" exitCode=0 Dec 01 16:02:03 crc kubenswrapper[4810]: I1201 16:02:03.477064 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706"} Dec 01 16:02:03 crc kubenswrapper[4810]: I1201 16:02:03.477412 4810 scope.go:117] "RemoveContainer" containerID="27dffee837d69d3c7e4cdc8a30f709eb2082d61dd91e53fa5540a464840c57c8" Dec 01 16:02:03 crc kubenswrapper[4810]: I1201 16:02:03.479179 4810 generic.go:334] "Generic (PLEG): container finished" podID="3ff88757-ea48-4e78-9452-d9f89ac909df" containerID="289328ff53e4549ae2ffb4684348ec80845c060d1c68c9dacde24aa29bac7b01" exitCode=0 Dec 01 16:02:03 crc kubenswrapper[4810]: I1201 16:02:03.479217 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"3ff88757-ea48-4e78-9452-d9f89ac909df","Type":"ContainerDied","Data":"289328ff53e4549ae2ffb4684348ec80845c060d1c68c9dacde24aa29bac7b01"} Dec 01 16:02:03 crc kubenswrapper[4810]: I1201 16:02:03.479244 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"3ff88757-ea48-4e78-9452-d9f89ac909df","Type":"ContainerStarted","Data":"7dacf92c084b0799c103d76e136f43e376ded8eb57efbf411bac8d29620a5f58"} Dec 01 16:02:03 crc kubenswrapper[4810]: E1201 16:02:03.604344 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:02:04 crc kubenswrapper[4810]: I1201 16:02:04.489528 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:02:04 crc kubenswrapper[4810]: E1201 16:02:04.489853 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:02:04 crc kubenswrapper[4810]: I1201 16:02:04.877204 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 01 16:02:04 crc kubenswrapper[4810]: I1201 16:02:04.899320 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_3ff88757-ea48-4e78-9452-d9f89ac909df/mariadb-client-1/0.log" Dec 01 16:02:04 crc kubenswrapper[4810]: I1201 16:02:04.929700 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Dec 01 16:02:04 crc kubenswrapper[4810]: I1201 16:02:04.933237 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Dec 01 16:02:04 crc kubenswrapper[4810]: I1201 16:02:04.960769 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44497\" (UniqueName: \"kubernetes.io/projected/3ff88757-ea48-4e78-9452-d9f89ac909df-kube-api-access-44497\") pod \"3ff88757-ea48-4e78-9452-d9f89ac909df\" (UID: \"3ff88757-ea48-4e78-9452-d9f89ac909df\") " Dec 01 16:02:04 crc kubenswrapper[4810]: I1201 16:02:04.975556 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ff88757-ea48-4e78-9452-d9f89ac909df-kube-api-access-44497" (OuterVolumeSpecName: "kube-api-access-44497") pod "3ff88757-ea48-4e78-9452-d9f89ac909df" (UID: "3ff88757-ea48-4e78-9452-d9f89ac909df"). InnerVolumeSpecName "kube-api-access-44497". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:02:05 crc kubenswrapper[4810]: I1201 16:02:05.069295 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44497\" (UniqueName: \"kubernetes.io/projected/3ff88757-ea48-4e78-9452-d9f89ac909df-kube-api-access-44497\") on node \"crc\" DevicePath \"\"" Dec 01 16:02:05 crc kubenswrapper[4810]: I1201 16:02:05.383027 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Dec 01 16:02:05 crc kubenswrapper[4810]: E1201 16:02:05.383578 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ff88757-ea48-4e78-9452-d9f89ac909df" containerName="mariadb-client-1" Dec 01 16:02:05 crc kubenswrapper[4810]: I1201 16:02:05.383667 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ff88757-ea48-4e78-9452-d9f89ac909df" containerName="mariadb-client-1" Dec 01 16:02:05 crc kubenswrapper[4810]: I1201 16:02:05.383906 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ff88757-ea48-4e78-9452-d9f89ac909df" containerName="mariadb-client-1" Dec 01 16:02:05 crc kubenswrapper[4810]: I1201 16:02:05.384577 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 01 16:02:05 crc kubenswrapper[4810]: I1201 16:02:05.389727 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 01 16:02:05 crc kubenswrapper[4810]: I1201 16:02:05.476555 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjq4k\" (UniqueName: \"kubernetes.io/projected/2ae0d8b0-cd9a-48a2-b567-16a944f9fe83-kube-api-access-hjq4k\") pod \"mariadb-client-4-default\" (UID: \"2ae0d8b0-cd9a-48a2-b567-16a944f9fe83\") " pod="openstack/mariadb-client-4-default" Dec 01 16:02:05 crc kubenswrapper[4810]: I1201 16:02:05.496542 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dacf92c084b0799c103d76e136f43e376ded8eb57efbf411bac8d29620a5f58" Dec 01 16:02:05 crc kubenswrapper[4810]: I1201 16:02:05.496583 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 01 16:02:05 crc kubenswrapper[4810]: I1201 16:02:05.577528 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjq4k\" (UniqueName: \"kubernetes.io/projected/2ae0d8b0-cd9a-48a2-b567-16a944f9fe83-kube-api-access-hjq4k\") pod \"mariadb-client-4-default\" (UID: \"2ae0d8b0-cd9a-48a2-b567-16a944f9fe83\") " pod="openstack/mariadb-client-4-default" Dec 01 16:02:05 crc kubenswrapper[4810]: I1201 16:02:05.597907 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjq4k\" (UniqueName: \"kubernetes.io/projected/2ae0d8b0-cd9a-48a2-b567-16a944f9fe83-kube-api-access-hjq4k\") pod \"mariadb-client-4-default\" (UID: \"2ae0d8b0-cd9a-48a2-b567-16a944f9fe83\") " pod="openstack/mariadb-client-4-default" Dec 01 16:02:05 crc kubenswrapper[4810]: I1201 16:02:05.708568 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 01 16:02:06 crc kubenswrapper[4810]: I1201 16:02:06.203206 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 01 16:02:06 crc kubenswrapper[4810]: I1201 16:02:06.500226 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ff88757-ea48-4e78-9452-d9f89ac909df" path="/var/lib/kubelet/pods/3ff88757-ea48-4e78-9452-d9f89ac909df/volumes" Dec 01 16:02:06 crc kubenswrapper[4810]: I1201 16:02:06.504419 4810 generic.go:334] "Generic (PLEG): container finished" podID="2ae0d8b0-cd9a-48a2-b567-16a944f9fe83" containerID="bd2ff2845d11b8de2f20849361b01186e4eccb39e9c53b023721d0a21f95933c" exitCode=0 Dec 01 16:02:06 crc kubenswrapper[4810]: I1201 16:02:06.504463 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"2ae0d8b0-cd9a-48a2-b567-16a944f9fe83","Type":"ContainerDied","Data":"bd2ff2845d11b8de2f20849361b01186e4eccb39e9c53b023721d0a21f95933c"} Dec 01 16:02:06 crc kubenswrapper[4810]: I1201 16:02:06.504518 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"2ae0d8b0-cd9a-48a2-b567-16a944f9fe83","Type":"ContainerStarted","Data":"1c5a6be3f618c9fdc7fdc4323597617c81f0b0e2232e760688385379c13ed2d7"} Dec 01 16:02:07 crc kubenswrapper[4810]: I1201 16:02:07.833835 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 01 16:02:07 crc kubenswrapper[4810]: I1201 16:02:07.849764 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_2ae0d8b0-cd9a-48a2-b567-16a944f9fe83/mariadb-client-4-default/0.log" Dec 01 16:02:07 crc kubenswrapper[4810]: I1201 16:02:07.878924 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 01 16:02:07 crc kubenswrapper[4810]: I1201 16:02:07.884932 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 01 16:02:08 crc kubenswrapper[4810]: I1201 16:02:08.010058 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjq4k\" (UniqueName: \"kubernetes.io/projected/2ae0d8b0-cd9a-48a2-b567-16a944f9fe83-kube-api-access-hjq4k\") pod \"2ae0d8b0-cd9a-48a2-b567-16a944f9fe83\" (UID: \"2ae0d8b0-cd9a-48a2-b567-16a944f9fe83\") " Dec 01 16:02:08 crc kubenswrapper[4810]: I1201 16:02:08.015774 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ae0d8b0-cd9a-48a2-b567-16a944f9fe83-kube-api-access-hjq4k" (OuterVolumeSpecName: "kube-api-access-hjq4k") pod "2ae0d8b0-cd9a-48a2-b567-16a944f9fe83" (UID: "2ae0d8b0-cd9a-48a2-b567-16a944f9fe83"). InnerVolumeSpecName "kube-api-access-hjq4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:02:08 crc kubenswrapper[4810]: I1201 16:02:08.111634 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjq4k\" (UniqueName: \"kubernetes.io/projected/2ae0d8b0-cd9a-48a2-b567-16a944f9fe83-kube-api-access-hjq4k\") on node \"crc\" DevicePath \"\"" Dec 01 16:02:08 crc kubenswrapper[4810]: I1201 16:02:08.503645 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ae0d8b0-cd9a-48a2-b567-16a944f9fe83" path="/var/lib/kubelet/pods/2ae0d8b0-cd9a-48a2-b567-16a944f9fe83/volumes" Dec 01 16:02:08 crc kubenswrapper[4810]: I1201 16:02:08.526610 4810 scope.go:117] "RemoveContainer" containerID="bd2ff2845d11b8de2f20849361b01186e4eccb39e9c53b023721d0a21f95933c" Dec 01 16:02:08 crc kubenswrapper[4810]: I1201 16:02:08.526647 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 01 16:02:11 crc kubenswrapper[4810]: I1201 16:02:11.728329 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Dec 01 16:02:11 crc kubenswrapper[4810]: E1201 16:02:11.729054 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ae0d8b0-cd9a-48a2-b567-16a944f9fe83" containerName="mariadb-client-4-default" Dec 01 16:02:11 crc kubenswrapper[4810]: I1201 16:02:11.729071 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ae0d8b0-cd9a-48a2-b567-16a944f9fe83" containerName="mariadb-client-4-default" Dec 01 16:02:11 crc kubenswrapper[4810]: I1201 16:02:11.729303 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ae0d8b0-cd9a-48a2-b567-16a944f9fe83" containerName="mariadb-client-4-default" Dec 01 16:02:11 crc kubenswrapper[4810]: I1201 16:02:11.729943 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 01 16:02:11 crc kubenswrapper[4810]: I1201 16:02:11.731973 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-xs2jm" Dec 01 16:02:11 crc kubenswrapper[4810]: I1201 16:02:11.735097 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 01 16:02:11 crc kubenswrapper[4810]: I1201 16:02:11.768928 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrgr5\" (UniqueName: \"kubernetes.io/projected/1e691bfd-8ed3-4cf5-bebc-007b5c1c7618-kube-api-access-nrgr5\") pod \"mariadb-client-5-default\" (UID: \"1e691bfd-8ed3-4cf5-bebc-007b5c1c7618\") " pod="openstack/mariadb-client-5-default" Dec 01 16:02:11 crc kubenswrapper[4810]: I1201 16:02:11.871075 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrgr5\" (UniqueName: \"kubernetes.io/projected/1e691bfd-8ed3-4cf5-bebc-007b5c1c7618-kube-api-access-nrgr5\") pod \"mariadb-client-5-default\" (UID: \"1e691bfd-8ed3-4cf5-bebc-007b5c1c7618\") " pod="openstack/mariadb-client-5-default" Dec 01 16:02:11 crc kubenswrapper[4810]: I1201 16:02:11.891661 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrgr5\" (UniqueName: \"kubernetes.io/projected/1e691bfd-8ed3-4cf5-bebc-007b5c1c7618-kube-api-access-nrgr5\") pod \"mariadb-client-5-default\" (UID: \"1e691bfd-8ed3-4cf5-bebc-007b5c1c7618\") " pod="openstack/mariadb-client-5-default" Dec 01 16:02:12 crc kubenswrapper[4810]: I1201 16:02:12.057099 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 01 16:02:13 crc kubenswrapper[4810]: I1201 16:02:12.557488 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 01 16:02:13 crc kubenswrapper[4810]: I1201 16:02:13.566609 4810 generic.go:334] "Generic (PLEG): container finished" podID="1e691bfd-8ed3-4cf5-bebc-007b5c1c7618" containerID="a9db3594ccf17b32f09e39d23617e8750b25df5c2b02d901949530612937db9d" exitCode=0 Dec 01 16:02:13 crc kubenswrapper[4810]: I1201 16:02:13.566762 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"1e691bfd-8ed3-4cf5-bebc-007b5c1c7618","Type":"ContainerDied","Data":"a9db3594ccf17b32f09e39d23617e8750b25df5c2b02d901949530612937db9d"} Dec 01 16:02:13 crc kubenswrapper[4810]: I1201 16:02:13.567039 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"1e691bfd-8ed3-4cf5-bebc-007b5c1c7618","Type":"ContainerStarted","Data":"d99f2b9bf7e1dbf8f39bd7f54ef93520747a3613fc645b51f2ccc8e11fac99e6"} Dec 01 16:02:14 crc kubenswrapper[4810]: I1201 16:02:14.920959 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 01 16:02:14 crc kubenswrapper[4810]: I1201 16:02:14.943528 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_1e691bfd-8ed3-4cf5-bebc-007b5c1c7618/mariadb-client-5-default/0.log" Dec 01 16:02:14 crc kubenswrapper[4810]: I1201 16:02:14.984113 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 01 16:02:14 crc kubenswrapper[4810]: I1201 16:02:14.992165 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 01 16:02:15 crc kubenswrapper[4810]: I1201 16:02:15.115020 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrgr5\" (UniqueName: \"kubernetes.io/projected/1e691bfd-8ed3-4cf5-bebc-007b5c1c7618-kube-api-access-nrgr5\") pod \"1e691bfd-8ed3-4cf5-bebc-007b5c1c7618\" (UID: \"1e691bfd-8ed3-4cf5-bebc-007b5c1c7618\") " Dec 01 16:02:15 crc kubenswrapper[4810]: I1201 16:02:15.120590 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Dec 01 16:02:15 crc kubenswrapper[4810]: E1201 16:02:15.121256 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e691bfd-8ed3-4cf5-bebc-007b5c1c7618" containerName="mariadb-client-5-default" Dec 01 16:02:15 crc kubenswrapper[4810]: I1201 16:02:15.121269 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e691bfd-8ed3-4cf5-bebc-007b5c1c7618" containerName="mariadb-client-5-default" Dec 01 16:02:15 crc kubenswrapper[4810]: I1201 16:02:15.121425 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e691bfd-8ed3-4cf5-bebc-007b5c1c7618" containerName="mariadb-client-5-default" Dec 01 16:02:15 crc kubenswrapper[4810]: I1201 16:02:15.123283 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e691bfd-8ed3-4cf5-bebc-007b5c1c7618-kube-api-access-nrgr5" (OuterVolumeSpecName: "kube-api-access-nrgr5") pod "1e691bfd-8ed3-4cf5-bebc-007b5c1c7618" (UID: "1e691bfd-8ed3-4cf5-bebc-007b5c1c7618"). InnerVolumeSpecName "kube-api-access-nrgr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:02:15 crc kubenswrapper[4810]: I1201 16:02:15.125276 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 01 16:02:15 crc kubenswrapper[4810]: I1201 16:02:15.134054 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 01 16:02:15 crc kubenswrapper[4810]: I1201 16:02:15.217883 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrgr5\" (UniqueName: \"kubernetes.io/projected/1e691bfd-8ed3-4cf5-bebc-007b5c1c7618-kube-api-access-nrgr5\") on node \"crc\" DevicePath \"\"" Dec 01 16:02:15 crc kubenswrapper[4810]: I1201 16:02:15.319774 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddl2c\" (UniqueName: \"kubernetes.io/projected/fd9aa6d0-6911-4584-8516-102524026058-kube-api-access-ddl2c\") pod \"mariadb-client-6-default\" (UID: \"fd9aa6d0-6911-4584-8516-102524026058\") " pod="openstack/mariadb-client-6-default" Dec 01 16:02:15 crc kubenswrapper[4810]: I1201 16:02:15.421290 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddl2c\" (UniqueName: \"kubernetes.io/projected/fd9aa6d0-6911-4584-8516-102524026058-kube-api-access-ddl2c\") pod \"mariadb-client-6-default\" (UID: \"fd9aa6d0-6911-4584-8516-102524026058\") " pod="openstack/mariadb-client-6-default" Dec 01 16:02:15 crc kubenswrapper[4810]: I1201 16:02:15.442254 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddl2c\" (UniqueName: \"kubernetes.io/projected/fd9aa6d0-6911-4584-8516-102524026058-kube-api-access-ddl2c\") pod \"mariadb-client-6-default\" (UID: \"fd9aa6d0-6911-4584-8516-102524026058\") " pod="openstack/mariadb-client-6-default" Dec 01 16:02:15 crc kubenswrapper[4810]: I1201 16:02:15.464826 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 01 16:02:15 crc kubenswrapper[4810]: I1201 16:02:15.585228 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d99f2b9bf7e1dbf8f39bd7f54ef93520747a3613fc645b51f2ccc8e11fac99e6" Dec 01 16:02:15 crc kubenswrapper[4810]: I1201 16:02:15.585275 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 01 16:02:15 crc kubenswrapper[4810]: I1201 16:02:15.988715 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 01 16:02:16 crc kubenswrapper[4810]: I1201 16:02:16.500951 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e691bfd-8ed3-4cf5-bebc-007b5c1c7618" path="/var/lib/kubelet/pods/1e691bfd-8ed3-4cf5-bebc-007b5c1c7618/volumes" Dec 01 16:02:16 crc kubenswrapper[4810]: I1201 16:02:16.597968 4810 generic.go:334] "Generic (PLEG): container finished" podID="fd9aa6d0-6911-4584-8516-102524026058" containerID="315a93270f1468bbf3956545bf51a57e8057f79cde0fa05b40328124129d10da" exitCode=1 Dec 01 16:02:16 crc kubenswrapper[4810]: I1201 16:02:16.598010 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"fd9aa6d0-6911-4584-8516-102524026058","Type":"ContainerDied","Data":"315a93270f1468bbf3956545bf51a57e8057f79cde0fa05b40328124129d10da"} Dec 01 16:02:16 crc kubenswrapper[4810]: I1201 16:02:16.598051 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"fd9aa6d0-6911-4584-8516-102524026058","Type":"ContainerStarted","Data":"5aebabccf9fbec2d4523600a9b7a9086f13f53100c0aa6642c2c4a1be65d9895"} Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.016273 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.037644 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-6-default_fd9aa6d0-6911-4584-8516-102524026058/mariadb-client-6-default/0.log" Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.067085 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.073154 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.166784 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddl2c\" (UniqueName: \"kubernetes.io/projected/fd9aa6d0-6911-4584-8516-102524026058-kube-api-access-ddl2c\") pod \"fd9aa6d0-6911-4584-8516-102524026058\" (UID: \"fd9aa6d0-6911-4584-8516-102524026058\") " Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.171511 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd9aa6d0-6911-4584-8516-102524026058-kube-api-access-ddl2c" (OuterVolumeSpecName: "kube-api-access-ddl2c") pod "fd9aa6d0-6911-4584-8516-102524026058" (UID: "fd9aa6d0-6911-4584-8516-102524026058"). InnerVolumeSpecName "kube-api-access-ddl2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.240564 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Dec 01 16:02:18 crc kubenswrapper[4810]: E1201 16:02:18.241104 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd9aa6d0-6911-4584-8516-102524026058" containerName="mariadb-client-6-default" Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.241126 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd9aa6d0-6911-4584-8516-102524026058" containerName="mariadb-client-6-default" Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.241317 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd9aa6d0-6911-4584-8516-102524026058" containerName="mariadb-client-6-default" Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.241899 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.247593 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.268552 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddl2c\" (UniqueName: \"kubernetes.io/projected/fd9aa6d0-6911-4584-8516-102524026058-kube-api-access-ddl2c\") on node \"crc\" DevicePath \"\"" Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.369584 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x7gn\" (UniqueName: \"kubernetes.io/projected/17e151b4-331d-4868-a185-226ecea725aa-kube-api-access-4x7gn\") pod \"mariadb-client-7-default\" (UID: \"17e151b4-331d-4868-a185-226ecea725aa\") " pod="openstack/mariadb-client-7-default" Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.471489 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x7gn\" (UniqueName: \"kubernetes.io/projected/17e151b4-331d-4868-a185-226ecea725aa-kube-api-access-4x7gn\") pod \"mariadb-client-7-default\" (UID: \"17e151b4-331d-4868-a185-226ecea725aa\") " pod="openstack/mariadb-client-7-default" Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.492901 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:02:18 crc kubenswrapper[4810]: E1201 16:02:18.493110 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.502624 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x7gn\" (UniqueName: \"kubernetes.io/projected/17e151b4-331d-4868-a185-226ecea725aa-kube-api-access-4x7gn\") pod \"mariadb-client-7-default\" (UID: \"17e151b4-331d-4868-a185-226ecea725aa\") " pod="openstack/mariadb-client-7-default" Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.523654 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd9aa6d0-6911-4584-8516-102524026058" path="/var/lib/kubelet/pods/fd9aa6d0-6911-4584-8516-102524026058/volumes" Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.566546 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.616832 4810 scope.go:117] "RemoveContainer" containerID="315a93270f1468bbf3956545bf51a57e8057f79cde0fa05b40328124129d10da" Dec 01 16:02:18 crc kubenswrapper[4810]: I1201 16:02:18.616910 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 01 16:02:19 crc kubenswrapper[4810]: I1201 16:02:19.052369 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 01 16:02:19 crc kubenswrapper[4810]: W1201 16:02:19.055883 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17e151b4_331d_4868_a185_226ecea725aa.slice/crio-3435ed3fa463b98528b47b6e38eb9195b6d35153d1e06e6dfb33d88cf8ba4357 WatchSource:0}: Error finding container 3435ed3fa463b98528b47b6e38eb9195b6d35153d1e06e6dfb33d88cf8ba4357: Status 404 returned error can't find the container with id 3435ed3fa463b98528b47b6e38eb9195b6d35153d1e06e6dfb33d88cf8ba4357 Dec 01 16:02:19 crc kubenswrapper[4810]: I1201 16:02:19.626770 4810 generic.go:334] "Generic (PLEG): container finished" podID="17e151b4-331d-4868-a185-226ecea725aa" containerID="8e997bb64a2abd3d4d7209ac74ffb5c1b18e2c592335bfe3b3f06baac4e03f1d" exitCode=0 Dec 01 16:02:19 crc kubenswrapper[4810]: I1201 16:02:19.626828 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"17e151b4-331d-4868-a185-226ecea725aa","Type":"ContainerDied","Data":"8e997bb64a2abd3d4d7209ac74ffb5c1b18e2c592335bfe3b3f06baac4e03f1d"} Dec 01 16:02:19 crc kubenswrapper[4810]: I1201 16:02:19.627166 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"17e151b4-331d-4868-a185-226ecea725aa","Type":"ContainerStarted","Data":"3435ed3fa463b98528b47b6e38eb9195b6d35153d1e06e6dfb33d88cf8ba4357"} Dec 01 16:02:20 crc kubenswrapper[4810]: I1201 16:02:20.965598 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 01 16:02:20 crc kubenswrapper[4810]: I1201 16:02:20.986370 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_17e151b4-331d-4868-a185-226ecea725aa/mariadb-client-7-default/0.log" Dec 01 16:02:21 crc kubenswrapper[4810]: I1201 16:02:21.014784 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 01 16:02:21 crc kubenswrapper[4810]: I1201 16:02:21.021270 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 01 16:02:21 crc kubenswrapper[4810]: I1201 16:02:21.108257 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x7gn\" (UniqueName: \"kubernetes.io/projected/17e151b4-331d-4868-a185-226ecea725aa-kube-api-access-4x7gn\") pod \"17e151b4-331d-4868-a185-226ecea725aa\" (UID: \"17e151b4-331d-4868-a185-226ecea725aa\") " Dec 01 16:02:21 crc kubenswrapper[4810]: I1201 16:02:21.114131 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17e151b4-331d-4868-a185-226ecea725aa-kube-api-access-4x7gn" (OuterVolumeSpecName: "kube-api-access-4x7gn") pod "17e151b4-331d-4868-a185-226ecea725aa" (UID: "17e151b4-331d-4868-a185-226ecea725aa"). InnerVolumeSpecName "kube-api-access-4x7gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:02:21 crc kubenswrapper[4810]: I1201 16:02:21.163486 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Dec 01 16:02:21 crc kubenswrapper[4810]: E1201 16:02:21.164236 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17e151b4-331d-4868-a185-226ecea725aa" containerName="mariadb-client-7-default" Dec 01 16:02:21 crc kubenswrapper[4810]: I1201 16:02:21.164316 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="17e151b4-331d-4868-a185-226ecea725aa" containerName="mariadb-client-7-default" Dec 01 16:02:21 crc kubenswrapper[4810]: I1201 16:02:21.164546 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="17e151b4-331d-4868-a185-226ecea725aa" containerName="mariadb-client-7-default" Dec 01 16:02:21 crc kubenswrapper[4810]: I1201 16:02:21.165165 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 01 16:02:21 crc kubenswrapper[4810]: I1201 16:02:21.175109 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Dec 01 16:02:21 crc kubenswrapper[4810]: I1201 16:02:21.210122 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s247w\" (UniqueName: \"kubernetes.io/projected/ecbb5705-4266-4158-8c96-e251c59a0301-kube-api-access-s247w\") pod \"mariadb-client-2\" (UID: \"ecbb5705-4266-4158-8c96-e251c59a0301\") " pod="openstack/mariadb-client-2" Dec 01 16:02:21 crc kubenswrapper[4810]: I1201 16:02:21.210263 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x7gn\" (UniqueName: \"kubernetes.io/projected/17e151b4-331d-4868-a185-226ecea725aa-kube-api-access-4x7gn\") on node \"crc\" DevicePath \"\"" Dec 01 16:02:21 crc kubenswrapper[4810]: I1201 16:02:21.313239 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s247w\" (UniqueName: \"kubernetes.io/projected/ecbb5705-4266-4158-8c96-e251c59a0301-kube-api-access-s247w\") pod \"mariadb-client-2\" (UID: \"ecbb5705-4266-4158-8c96-e251c59a0301\") " pod="openstack/mariadb-client-2" Dec 01 16:02:21 crc kubenswrapper[4810]: I1201 16:02:21.330753 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s247w\" (UniqueName: \"kubernetes.io/projected/ecbb5705-4266-4158-8c96-e251c59a0301-kube-api-access-s247w\") pod \"mariadb-client-2\" (UID: \"ecbb5705-4266-4158-8c96-e251c59a0301\") " pod="openstack/mariadb-client-2" Dec 01 16:02:21 crc kubenswrapper[4810]: I1201 16:02:21.485106 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 01 16:02:21 crc kubenswrapper[4810]: I1201 16:02:21.652433 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3435ed3fa463b98528b47b6e38eb9195b6d35153d1e06e6dfb33d88cf8ba4357" Dec 01 16:02:21 crc kubenswrapper[4810]: I1201 16:02:21.652546 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 01 16:02:21 crc kubenswrapper[4810]: I1201 16:02:21.957755 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Dec 01 16:02:21 crc kubenswrapper[4810]: W1201 16:02:21.963426 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecbb5705_4266_4158_8c96_e251c59a0301.slice/crio-01c8a8805084041a11c4a38149852276e616a6492d4296147edc535fdbc6f5dc WatchSource:0}: Error finding container 01c8a8805084041a11c4a38149852276e616a6492d4296147edc535fdbc6f5dc: Status 404 returned error can't find the container with id 01c8a8805084041a11c4a38149852276e616a6492d4296147edc535fdbc6f5dc Dec 01 16:02:22 crc kubenswrapper[4810]: I1201 16:02:22.513779 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17e151b4-331d-4868-a185-226ecea725aa" path="/var/lib/kubelet/pods/17e151b4-331d-4868-a185-226ecea725aa/volumes" Dec 01 16:02:22 crc kubenswrapper[4810]: I1201 16:02:22.665760 4810 generic.go:334] "Generic (PLEG): container finished" podID="ecbb5705-4266-4158-8c96-e251c59a0301" containerID="b0ee5f6476ed06242ca25422592566210345a67e0ed68bcd06679f5b24703f8f" exitCode=0 Dec 01 16:02:22 crc kubenswrapper[4810]: I1201 16:02:22.665816 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"ecbb5705-4266-4158-8c96-e251c59a0301","Type":"ContainerDied","Data":"b0ee5f6476ed06242ca25422592566210345a67e0ed68bcd06679f5b24703f8f"} Dec 01 16:02:22 crc kubenswrapper[4810]: I1201 16:02:22.665846 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"ecbb5705-4266-4158-8c96-e251c59a0301","Type":"ContainerStarted","Data":"01c8a8805084041a11c4a38149852276e616a6492d4296147edc535fdbc6f5dc"} Dec 01 16:02:24 crc kubenswrapper[4810]: I1201 16:02:24.060498 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 01 16:02:24 crc kubenswrapper[4810]: I1201 16:02:24.077983 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_ecbb5705-4266-4158-8c96-e251c59a0301/mariadb-client-2/0.log" Dec 01 16:02:24 crc kubenswrapper[4810]: I1201 16:02:24.104142 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Dec 01 16:02:24 crc kubenswrapper[4810]: I1201 16:02:24.109669 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Dec 01 16:02:24 crc kubenswrapper[4810]: I1201 16:02:24.154928 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s247w\" (UniqueName: \"kubernetes.io/projected/ecbb5705-4266-4158-8c96-e251c59a0301-kube-api-access-s247w\") pod \"ecbb5705-4266-4158-8c96-e251c59a0301\" (UID: \"ecbb5705-4266-4158-8c96-e251c59a0301\") " Dec 01 16:02:24 crc kubenswrapper[4810]: I1201 16:02:24.159664 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecbb5705-4266-4158-8c96-e251c59a0301-kube-api-access-s247w" (OuterVolumeSpecName: "kube-api-access-s247w") pod "ecbb5705-4266-4158-8c96-e251c59a0301" (UID: "ecbb5705-4266-4158-8c96-e251c59a0301"). InnerVolumeSpecName "kube-api-access-s247w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:02:24 crc kubenswrapper[4810]: I1201 16:02:24.256889 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s247w\" (UniqueName: \"kubernetes.io/projected/ecbb5705-4266-4158-8c96-e251c59a0301-kube-api-access-s247w\") on node \"crc\" DevicePath \"\"" Dec 01 16:02:24 crc kubenswrapper[4810]: I1201 16:02:24.502951 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecbb5705-4266-4158-8c96-e251c59a0301" path="/var/lib/kubelet/pods/ecbb5705-4266-4158-8c96-e251c59a0301/volumes" Dec 01 16:02:24 crc kubenswrapper[4810]: I1201 16:02:24.679430 4810 scope.go:117] "RemoveContainer" containerID="b0ee5f6476ed06242ca25422592566210345a67e0ed68bcd06679f5b24703f8f" Dec 01 16:02:24 crc kubenswrapper[4810]: I1201 16:02:24.679529 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 01 16:02:33 crc kubenswrapper[4810]: I1201 16:02:33.490881 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:02:33 crc kubenswrapper[4810]: E1201 16:02:33.491801 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:02:48 crc kubenswrapper[4810]: I1201 16:02:48.492918 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:02:48 crc kubenswrapper[4810]: E1201 16:02:48.494270 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:03:00 crc kubenswrapper[4810]: I1201 16:03:00.491042 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:03:00 crc kubenswrapper[4810]: E1201 16:03:00.491700 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:03:12 crc kubenswrapper[4810]: I1201 16:03:12.490808 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:03:12 crc kubenswrapper[4810]: E1201 16:03:12.491640 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:03:24 crc kubenswrapper[4810]: I1201 16:03:24.496022 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:03:24 crc kubenswrapper[4810]: E1201 16:03:24.496850 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:03:37 crc kubenswrapper[4810]: I1201 16:03:37.490598 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:03:37 crc kubenswrapper[4810]: E1201 16:03:37.491308 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:03:52 crc kubenswrapper[4810]: I1201 16:03:52.491108 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:03:52 crc kubenswrapper[4810]: E1201 16:03:52.492681 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:04:03 crc kubenswrapper[4810]: I1201 16:04:03.492253 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:04:03 crc kubenswrapper[4810]: E1201 16:04:03.493211 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:04:08 crc kubenswrapper[4810]: I1201 16:04:08.971541 4810 scope.go:117] "RemoveContainer" containerID="79a31301bf0b55903f9ccb9d1d6ebee23f756fedc97f72fa4043821de323c517" Dec 01 16:04:14 crc kubenswrapper[4810]: I1201 16:04:14.496614 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:04:14 crc kubenswrapper[4810]: E1201 16:04:14.497531 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:04:25 crc kubenswrapper[4810]: I1201 16:04:25.492027 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:04:25 crc kubenswrapper[4810]: E1201 16:04:25.492915 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:04:36 crc kubenswrapper[4810]: I1201 16:04:36.491411 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:04:36 crc kubenswrapper[4810]: E1201 16:04:36.493087 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:04:48 crc kubenswrapper[4810]: I1201 16:04:48.491621 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:04:48 crc kubenswrapper[4810]: E1201 16:04:48.492566 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:05:02 crc kubenswrapper[4810]: I1201 16:05:02.491119 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:05:02 crc kubenswrapper[4810]: E1201 16:05:02.491963 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:05:15 crc kubenswrapper[4810]: I1201 16:05:15.491593 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:05:15 crc kubenswrapper[4810]: E1201 16:05:15.492392 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:05:19 crc kubenswrapper[4810]: I1201 16:05:19.102928 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-29nmh"] Dec 01 16:05:19 crc kubenswrapper[4810]: E1201 16:05:19.103521 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecbb5705-4266-4158-8c96-e251c59a0301" containerName="mariadb-client-2" Dec 01 16:05:19 crc kubenswrapper[4810]: I1201 16:05:19.103534 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecbb5705-4266-4158-8c96-e251c59a0301" containerName="mariadb-client-2" Dec 01 16:05:19 crc kubenswrapper[4810]: I1201 16:05:19.103697 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecbb5705-4266-4158-8c96-e251c59a0301" containerName="mariadb-client-2" Dec 01 16:05:19 crc kubenswrapper[4810]: I1201 16:05:19.104948 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-29nmh" Dec 01 16:05:19 crc kubenswrapper[4810]: I1201 16:05:19.119913 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-29nmh"] Dec 01 16:05:19 crc kubenswrapper[4810]: I1201 16:05:19.248262 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1-utilities\") pod \"certified-operators-29nmh\" (UID: \"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1\") " pod="openshift-marketplace/certified-operators-29nmh" Dec 01 16:05:19 crc kubenswrapper[4810]: I1201 16:05:19.248356 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1-catalog-content\") pod \"certified-operators-29nmh\" (UID: \"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1\") " pod="openshift-marketplace/certified-operators-29nmh" Dec 01 16:05:19 crc kubenswrapper[4810]: I1201 16:05:19.248414 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mr66\" (UniqueName: \"kubernetes.io/projected/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1-kube-api-access-5mr66\") pod \"certified-operators-29nmh\" (UID: \"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1\") " pod="openshift-marketplace/certified-operators-29nmh" Dec 01 16:05:19 crc kubenswrapper[4810]: I1201 16:05:19.350564 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1-utilities\") pod \"certified-operators-29nmh\" (UID: \"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1\") " pod="openshift-marketplace/certified-operators-29nmh" Dec 01 16:05:19 crc kubenswrapper[4810]: I1201 16:05:19.350680 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1-catalog-content\") pod \"certified-operators-29nmh\" (UID: \"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1\") " pod="openshift-marketplace/certified-operators-29nmh" Dec 01 16:05:19 crc kubenswrapper[4810]: I1201 16:05:19.350730 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mr66\" (UniqueName: \"kubernetes.io/projected/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1-kube-api-access-5mr66\") pod \"certified-operators-29nmh\" (UID: \"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1\") " pod="openshift-marketplace/certified-operators-29nmh" Dec 01 16:05:19 crc kubenswrapper[4810]: I1201 16:05:19.351201 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1-utilities\") pod \"certified-operators-29nmh\" (UID: \"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1\") " pod="openshift-marketplace/certified-operators-29nmh" Dec 01 16:05:19 crc kubenswrapper[4810]: I1201 16:05:19.351486 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1-catalog-content\") pod \"certified-operators-29nmh\" (UID: \"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1\") " pod="openshift-marketplace/certified-operators-29nmh" Dec 01 16:05:19 crc kubenswrapper[4810]: I1201 16:05:19.379388 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mr66\" (UniqueName: \"kubernetes.io/projected/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1-kube-api-access-5mr66\") pod \"certified-operators-29nmh\" (UID: \"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1\") " pod="openshift-marketplace/certified-operators-29nmh" Dec 01 16:05:19 crc kubenswrapper[4810]: I1201 16:05:19.422706 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-29nmh" Dec 01 16:05:19 crc kubenswrapper[4810]: I1201 16:05:19.941337 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-29nmh"] Dec 01 16:05:20 crc kubenswrapper[4810]: I1201 16:05:20.068296 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-29nmh" event={"ID":"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1","Type":"ContainerStarted","Data":"43cc65ef1838138176b250f8e4b3c4ba6d681ce83d06f455d2e8702c4d941228"} Dec 01 16:05:21 crc kubenswrapper[4810]: I1201 16:05:21.078769 4810 generic.go:334] "Generic (PLEG): container finished" podID="4d0ae1a6-1098-4ea0-96f7-9a6792f755a1" containerID="8d107bd704840792a9eedc1d43a251d3a9f235d678a93b042705ebdc556de9a9" exitCode=0 Dec 01 16:05:21 crc kubenswrapper[4810]: I1201 16:05:21.078851 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-29nmh" event={"ID":"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1","Type":"ContainerDied","Data":"8d107bd704840792a9eedc1d43a251d3a9f235d678a93b042705ebdc556de9a9"} Dec 01 16:05:21 crc kubenswrapper[4810]: I1201 16:05:21.080960 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:05:23 crc kubenswrapper[4810]: I1201 16:05:23.100182 4810 generic.go:334] "Generic (PLEG): container finished" podID="4d0ae1a6-1098-4ea0-96f7-9a6792f755a1" containerID="8389d87de14bf94b3aa124b1e2fce9e6f2ca72aeb7e4c84f31944e8b2d03ce46" exitCode=0 Dec 01 16:05:23 crc kubenswrapper[4810]: I1201 16:05:23.100233 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-29nmh" event={"ID":"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1","Type":"ContainerDied","Data":"8389d87de14bf94b3aa124b1e2fce9e6f2ca72aeb7e4c84f31944e8b2d03ce46"} Dec 01 16:05:24 crc kubenswrapper[4810]: I1201 16:05:24.110293 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-29nmh" event={"ID":"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1","Type":"ContainerStarted","Data":"4081390eb4034e7677e45bd5ac2232c29abf462f99e4060d9865edfc76153fd6"} Dec 01 16:05:24 crc kubenswrapper[4810]: I1201 16:05:24.132001 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-29nmh" podStartSLOduration=2.546130857 podStartE2EDuration="5.131982687s" podCreationTimestamp="2025-12-01 16:05:19 +0000 UTC" firstStartedPulling="2025-12-01 16:05:21.080668678 +0000 UTC m=+5486.844178301" lastFinishedPulling="2025-12-01 16:05:23.666520528 +0000 UTC m=+5489.430030131" observedRunningTime="2025-12-01 16:05:24.126462948 +0000 UTC m=+5489.889972551" watchObservedRunningTime="2025-12-01 16:05:24.131982687 +0000 UTC m=+5489.895492290" Dec 01 16:05:26 crc kubenswrapper[4810]: I1201 16:05:26.491661 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:05:26 crc kubenswrapper[4810]: E1201 16:05:26.492144 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:05:29 crc kubenswrapper[4810]: I1201 16:05:29.422918 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-29nmh" Dec 01 16:05:29 crc kubenswrapper[4810]: I1201 16:05:29.423298 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-29nmh" Dec 01 16:05:29 crc kubenswrapper[4810]: I1201 16:05:29.462224 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-29nmh" Dec 01 16:05:30 crc kubenswrapper[4810]: I1201 16:05:30.205444 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-29nmh" Dec 01 16:05:30 crc kubenswrapper[4810]: I1201 16:05:30.251485 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-29nmh"] Dec 01 16:05:32 crc kubenswrapper[4810]: I1201 16:05:32.174580 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-29nmh" podUID="4d0ae1a6-1098-4ea0-96f7-9a6792f755a1" containerName="registry-server" containerID="cri-o://4081390eb4034e7677e45bd5ac2232c29abf462f99e4060d9865edfc76153fd6" gracePeriod=2 Dec 01 16:05:32 crc kubenswrapper[4810]: I1201 16:05:32.580427 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-29nmh" Dec 01 16:05:32 crc kubenswrapper[4810]: I1201 16:05:32.746909 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mr66\" (UniqueName: \"kubernetes.io/projected/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1-kube-api-access-5mr66\") pod \"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1\" (UID: \"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1\") " Dec 01 16:05:32 crc kubenswrapper[4810]: I1201 16:05:32.746970 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1-catalog-content\") pod \"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1\" (UID: \"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1\") " Dec 01 16:05:32 crc kubenswrapper[4810]: I1201 16:05:32.747165 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1-utilities\") pod \"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1\" (UID: \"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1\") " Dec 01 16:05:32 crc kubenswrapper[4810]: I1201 16:05:32.748165 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1-utilities" (OuterVolumeSpecName: "utilities") pod "4d0ae1a6-1098-4ea0-96f7-9a6792f755a1" (UID: "4d0ae1a6-1098-4ea0-96f7-9a6792f755a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:05:32 crc kubenswrapper[4810]: I1201 16:05:32.752906 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1-kube-api-access-5mr66" (OuterVolumeSpecName: "kube-api-access-5mr66") pod "4d0ae1a6-1098-4ea0-96f7-9a6792f755a1" (UID: "4d0ae1a6-1098-4ea0-96f7-9a6792f755a1"). InnerVolumeSpecName "kube-api-access-5mr66". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:05:32 crc kubenswrapper[4810]: I1201 16:05:32.803519 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d0ae1a6-1098-4ea0-96f7-9a6792f755a1" (UID: "4d0ae1a6-1098-4ea0-96f7-9a6792f755a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:05:32 crc kubenswrapper[4810]: I1201 16:05:32.848685 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:05:32 crc kubenswrapper[4810]: I1201 16:05:32.848891 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mr66\" (UniqueName: \"kubernetes.io/projected/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1-kube-api-access-5mr66\") on node \"crc\" DevicePath \"\"" Dec 01 16:05:32 crc kubenswrapper[4810]: I1201 16:05:32.848944 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:05:33 crc kubenswrapper[4810]: I1201 16:05:33.225994 4810 generic.go:334] "Generic (PLEG): container finished" podID="4d0ae1a6-1098-4ea0-96f7-9a6792f755a1" containerID="4081390eb4034e7677e45bd5ac2232c29abf462f99e4060d9865edfc76153fd6" exitCode=0 Dec 01 16:05:33 crc kubenswrapper[4810]: I1201 16:05:33.226041 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-29nmh" event={"ID":"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1","Type":"ContainerDied","Data":"4081390eb4034e7677e45bd5ac2232c29abf462f99e4060d9865edfc76153fd6"} Dec 01 16:05:33 crc kubenswrapper[4810]: I1201 16:05:33.226068 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-29nmh" event={"ID":"4d0ae1a6-1098-4ea0-96f7-9a6792f755a1","Type":"ContainerDied","Data":"43cc65ef1838138176b250f8e4b3c4ba6d681ce83d06f455d2e8702c4d941228"} Dec 01 16:05:33 crc kubenswrapper[4810]: I1201 16:05:33.226085 4810 scope.go:117] "RemoveContainer" containerID="4081390eb4034e7677e45bd5ac2232c29abf462f99e4060d9865edfc76153fd6" Dec 01 16:05:33 crc kubenswrapper[4810]: I1201 16:05:33.226081 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-29nmh" Dec 01 16:05:33 crc kubenswrapper[4810]: I1201 16:05:33.258207 4810 scope.go:117] "RemoveContainer" containerID="8389d87de14bf94b3aa124b1e2fce9e6f2ca72aeb7e4c84f31944e8b2d03ce46" Dec 01 16:05:33 crc kubenswrapper[4810]: I1201 16:05:33.262966 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-29nmh"] Dec 01 16:05:33 crc kubenswrapper[4810]: I1201 16:05:33.269503 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-29nmh"] Dec 01 16:05:33 crc kubenswrapper[4810]: I1201 16:05:33.279208 4810 scope.go:117] "RemoveContainer" containerID="8d107bd704840792a9eedc1d43a251d3a9f235d678a93b042705ebdc556de9a9" Dec 01 16:05:33 crc kubenswrapper[4810]: I1201 16:05:33.306219 4810 scope.go:117] "RemoveContainer" containerID="4081390eb4034e7677e45bd5ac2232c29abf462f99e4060d9865edfc76153fd6" Dec 01 16:05:33 crc kubenswrapper[4810]: E1201 16:05:33.306631 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4081390eb4034e7677e45bd5ac2232c29abf462f99e4060d9865edfc76153fd6\": container with ID starting with 4081390eb4034e7677e45bd5ac2232c29abf462f99e4060d9865edfc76153fd6 not found: ID does not exist" containerID="4081390eb4034e7677e45bd5ac2232c29abf462f99e4060d9865edfc76153fd6" Dec 01 16:05:33 crc kubenswrapper[4810]: I1201 16:05:33.306659 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4081390eb4034e7677e45bd5ac2232c29abf462f99e4060d9865edfc76153fd6"} err="failed to get container status \"4081390eb4034e7677e45bd5ac2232c29abf462f99e4060d9865edfc76153fd6\": rpc error: code = NotFound desc = could not find container \"4081390eb4034e7677e45bd5ac2232c29abf462f99e4060d9865edfc76153fd6\": container with ID starting with 4081390eb4034e7677e45bd5ac2232c29abf462f99e4060d9865edfc76153fd6 not found: ID does not exist" Dec 01 16:05:33 crc kubenswrapper[4810]: I1201 16:05:33.306678 4810 scope.go:117] "RemoveContainer" containerID="8389d87de14bf94b3aa124b1e2fce9e6f2ca72aeb7e4c84f31944e8b2d03ce46" Dec 01 16:05:33 crc kubenswrapper[4810]: E1201 16:05:33.307061 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8389d87de14bf94b3aa124b1e2fce9e6f2ca72aeb7e4c84f31944e8b2d03ce46\": container with ID starting with 8389d87de14bf94b3aa124b1e2fce9e6f2ca72aeb7e4c84f31944e8b2d03ce46 not found: ID does not exist" containerID="8389d87de14bf94b3aa124b1e2fce9e6f2ca72aeb7e4c84f31944e8b2d03ce46" Dec 01 16:05:33 crc kubenswrapper[4810]: I1201 16:05:33.307087 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8389d87de14bf94b3aa124b1e2fce9e6f2ca72aeb7e4c84f31944e8b2d03ce46"} err="failed to get container status \"8389d87de14bf94b3aa124b1e2fce9e6f2ca72aeb7e4c84f31944e8b2d03ce46\": rpc error: code = NotFound desc = could not find container \"8389d87de14bf94b3aa124b1e2fce9e6f2ca72aeb7e4c84f31944e8b2d03ce46\": container with ID starting with 8389d87de14bf94b3aa124b1e2fce9e6f2ca72aeb7e4c84f31944e8b2d03ce46 not found: ID does not exist" Dec 01 16:05:33 crc kubenswrapper[4810]: I1201 16:05:33.307100 4810 scope.go:117] "RemoveContainer" containerID="8d107bd704840792a9eedc1d43a251d3a9f235d678a93b042705ebdc556de9a9" Dec 01 16:05:33 crc kubenswrapper[4810]: E1201 16:05:33.307315 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d107bd704840792a9eedc1d43a251d3a9f235d678a93b042705ebdc556de9a9\": container with ID starting with 8d107bd704840792a9eedc1d43a251d3a9f235d678a93b042705ebdc556de9a9 not found: ID does not exist" containerID="8d107bd704840792a9eedc1d43a251d3a9f235d678a93b042705ebdc556de9a9" Dec 01 16:05:33 crc kubenswrapper[4810]: I1201 16:05:33.307332 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d107bd704840792a9eedc1d43a251d3a9f235d678a93b042705ebdc556de9a9"} err="failed to get container status \"8d107bd704840792a9eedc1d43a251d3a9f235d678a93b042705ebdc556de9a9\": rpc error: code = NotFound desc = could not find container \"8d107bd704840792a9eedc1d43a251d3a9f235d678a93b042705ebdc556de9a9\": container with ID starting with 8d107bd704840792a9eedc1d43a251d3a9f235d678a93b042705ebdc556de9a9 not found: ID does not exist" Dec 01 16:05:34 crc kubenswrapper[4810]: I1201 16:05:34.501943 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d0ae1a6-1098-4ea0-96f7-9a6792f755a1" path="/var/lib/kubelet/pods/4d0ae1a6-1098-4ea0-96f7-9a6792f755a1/volumes" Dec 01 16:05:39 crc kubenswrapper[4810]: I1201 16:05:39.491709 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:05:39 crc kubenswrapper[4810]: E1201 16:05:39.492380 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:05:50 crc kubenswrapper[4810]: I1201 16:05:50.493348 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:05:50 crc kubenswrapper[4810]: E1201 16:05:50.494257 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:06:00 crc kubenswrapper[4810]: I1201 16:06:00.116676 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Dec 01 16:06:00 crc kubenswrapper[4810]: E1201 16:06:00.117676 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d0ae1a6-1098-4ea0-96f7-9a6792f755a1" containerName="registry-server" Dec 01 16:06:00 crc kubenswrapper[4810]: I1201 16:06:00.117694 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d0ae1a6-1098-4ea0-96f7-9a6792f755a1" containerName="registry-server" Dec 01 16:06:00 crc kubenswrapper[4810]: E1201 16:06:00.117720 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d0ae1a6-1098-4ea0-96f7-9a6792f755a1" containerName="extract-content" Dec 01 16:06:00 crc kubenswrapper[4810]: I1201 16:06:00.117729 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d0ae1a6-1098-4ea0-96f7-9a6792f755a1" containerName="extract-content" Dec 01 16:06:00 crc kubenswrapper[4810]: E1201 16:06:00.117742 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d0ae1a6-1098-4ea0-96f7-9a6792f755a1" containerName="extract-utilities" Dec 01 16:06:00 crc kubenswrapper[4810]: I1201 16:06:00.117750 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d0ae1a6-1098-4ea0-96f7-9a6792f755a1" containerName="extract-utilities" Dec 01 16:06:00 crc kubenswrapper[4810]: I1201 16:06:00.117948 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d0ae1a6-1098-4ea0-96f7-9a6792f755a1" containerName="registry-server" Dec 01 16:06:00 crc kubenswrapper[4810]: I1201 16:06:00.118578 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 01 16:06:00 crc kubenswrapper[4810]: I1201 16:06:00.120441 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-xs2jm" Dec 01 16:06:00 crc kubenswrapper[4810]: I1201 16:06:00.130107 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Dec 01 16:06:00 crc kubenswrapper[4810]: I1201 16:06:00.279855 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmn6h\" (UniqueName: \"kubernetes.io/projected/f8d6ddae-90bb-43a8-8b28-fb05dd494ba3-kube-api-access-jmn6h\") pod \"mariadb-copy-data\" (UID: \"f8d6ddae-90bb-43a8-8b28-fb05dd494ba3\") " pod="openstack/mariadb-copy-data" Dec 01 16:06:00 crc kubenswrapper[4810]: I1201 16:06:00.279919 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-42fa2d68-0280-474c-b021-0e7adb73eaf0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-42fa2d68-0280-474c-b021-0e7adb73eaf0\") pod \"mariadb-copy-data\" (UID: \"f8d6ddae-90bb-43a8-8b28-fb05dd494ba3\") " pod="openstack/mariadb-copy-data" Dec 01 16:06:00 crc kubenswrapper[4810]: I1201 16:06:00.381699 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmn6h\" (UniqueName: \"kubernetes.io/projected/f8d6ddae-90bb-43a8-8b28-fb05dd494ba3-kube-api-access-jmn6h\") pod \"mariadb-copy-data\" (UID: \"f8d6ddae-90bb-43a8-8b28-fb05dd494ba3\") " pod="openstack/mariadb-copy-data" Dec 01 16:06:00 crc kubenswrapper[4810]: I1201 16:06:00.381754 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-42fa2d68-0280-474c-b021-0e7adb73eaf0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-42fa2d68-0280-474c-b021-0e7adb73eaf0\") pod \"mariadb-copy-data\" (UID: \"f8d6ddae-90bb-43a8-8b28-fb05dd494ba3\") " pod="openstack/mariadb-copy-data" Dec 01 16:06:00 crc kubenswrapper[4810]: I1201 16:06:00.386054 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 16:06:00 crc kubenswrapper[4810]: I1201 16:06:00.386102 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-42fa2d68-0280-474c-b021-0e7adb73eaf0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-42fa2d68-0280-474c-b021-0e7adb73eaf0\") pod \"mariadb-copy-data\" (UID: \"f8d6ddae-90bb-43a8-8b28-fb05dd494ba3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8e1f6df62c16f7fab80fea40fb8dbc5b90c2b22005b3b2d12ff09dc7664e8d87/globalmount\"" pod="openstack/mariadb-copy-data" Dec 01 16:06:00 crc kubenswrapper[4810]: I1201 16:06:00.409495 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmn6h\" (UniqueName: \"kubernetes.io/projected/f8d6ddae-90bb-43a8-8b28-fb05dd494ba3-kube-api-access-jmn6h\") pod \"mariadb-copy-data\" (UID: \"f8d6ddae-90bb-43a8-8b28-fb05dd494ba3\") " pod="openstack/mariadb-copy-data" Dec 01 16:06:00 crc kubenswrapper[4810]: I1201 16:06:00.429208 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-42fa2d68-0280-474c-b021-0e7adb73eaf0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-42fa2d68-0280-474c-b021-0e7adb73eaf0\") pod \"mariadb-copy-data\" (UID: \"f8d6ddae-90bb-43a8-8b28-fb05dd494ba3\") " pod="openstack/mariadb-copy-data" Dec 01 16:06:00 crc kubenswrapper[4810]: I1201 16:06:00.447588 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 01 16:06:00 crc kubenswrapper[4810]: I1201 16:06:00.968200 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Dec 01 16:06:01 crc kubenswrapper[4810]: I1201 16:06:01.451431 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"f8d6ddae-90bb-43a8-8b28-fb05dd494ba3","Type":"ContainerStarted","Data":"053a9cd297dff6f3bd2ddc9ee7d252d1f236607a51ac59962c81e9455c5a0f2a"} Dec 01 16:06:01 crc kubenswrapper[4810]: I1201 16:06:01.451824 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"f8d6ddae-90bb-43a8-8b28-fb05dd494ba3","Type":"ContainerStarted","Data":"cb990b29946c7c4e3745ee064ef5f50627ee656980823b6c5cf70db855f18c3c"} Dec 01 16:06:04 crc kubenswrapper[4810]: I1201 16:06:04.681853 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=5.68183372 podStartE2EDuration="5.68183372s" podCreationTimestamp="2025-12-01 16:05:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:06:01.466627437 +0000 UTC m=+5527.230137040" watchObservedRunningTime="2025-12-01 16:06:04.68183372 +0000 UTC m=+5530.445343323" Dec 01 16:06:04 crc kubenswrapper[4810]: I1201 16:06:04.689363 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Dec 01 16:06:04 crc kubenswrapper[4810]: I1201 16:06:04.690352 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 01 16:06:04 crc kubenswrapper[4810]: I1201 16:06:04.695958 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 01 16:06:04 crc kubenswrapper[4810]: I1201 16:06:04.845393 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd7fr\" (UniqueName: \"kubernetes.io/projected/76b5695c-7889-4c67-9441-68053d417d7d-kube-api-access-xd7fr\") pod \"mariadb-client\" (UID: \"76b5695c-7889-4c67-9441-68053d417d7d\") " pod="openstack/mariadb-client" Dec 01 16:06:04 crc kubenswrapper[4810]: I1201 16:06:04.947333 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd7fr\" (UniqueName: \"kubernetes.io/projected/76b5695c-7889-4c67-9441-68053d417d7d-kube-api-access-xd7fr\") pod \"mariadb-client\" (UID: \"76b5695c-7889-4c67-9441-68053d417d7d\") " pod="openstack/mariadb-client" Dec 01 16:06:04 crc kubenswrapper[4810]: I1201 16:06:04.967972 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd7fr\" (UniqueName: \"kubernetes.io/projected/76b5695c-7889-4c67-9441-68053d417d7d-kube-api-access-xd7fr\") pod \"mariadb-client\" (UID: \"76b5695c-7889-4c67-9441-68053d417d7d\") " pod="openstack/mariadb-client" Dec 01 16:06:05 crc kubenswrapper[4810]: I1201 16:06:05.009881 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 01 16:06:05 crc kubenswrapper[4810]: I1201 16:06:05.425410 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 01 16:06:05 crc kubenswrapper[4810]: W1201 16:06:05.433735 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76b5695c_7889_4c67_9441_68053d417d7d.slice/crio-962763d086898a290033b14b6051856d079d2b605aabb818f219c04085bfe910 WatchSource:0}: Error finding container 962763d086898a290033b14b6051856d079d2b605aabb818f219c04085bfe910: Status 404 returned error can't find the container with id 962763d086898a290033b14b6051856d079d2b605aabb818f219c04085bfe910 Dec 01 16:06:05 crc kubenswrapper[4810]: I1201 16:06:05.481801 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"76b5695c-7889-4c67-9441-68053d417d7d","Type":"ContainerStarted","Data":"962763d086898a290033b14b6051856d079d2b605aabb818f219c04085bfe910"} Dec 01 16:06:05 crc kubenswrapper[4810]: I1201 16:06:05.491257 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:06:05 crc kubenswrapper[4810]: E1201 16:06:05.491489 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:06:06 crc kubenswrapper[4810]: I1201 16:06:06.490513 4810 generic.go:334] "Generic (PLEG): container finished" podID="76b5695c-7889-4c67-9441-68053d417d7d" containerID="57030bcd1c685bf36cb57fd2b533f165ece63a0ca0887c3dd33d223e31bd1dc0" exitCode=0 Dec 01 16:06:06 crc kubenswrapper[4810]: I1201 16:06:06.513270 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"76b5695c-7889-4c67-9441-68053d417d7d","Type":"ContainerDied","Data":"57030bcd1c685bf36cb57fd2b533f165ece63a0ca0887c3dd33d223e31bd1dc0"} Dec 01 16:06:07 crc kubenswrapper[4810]: I1201 16:06:07.839420 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 01 16:06:07 crc kubenswrapper[4810]: I1201 16:06:07.860128 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_76b5695c-7889-4c67-9441-68053d417d7d/mariadb-client/0.log" Dec 01 16:06:07 crc kubenswrapper[4810]: I1201 16:06:07.890050 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Dec 01 16:06:07 crc kubenswrapper[4810]: I1201 16:06:07.896464 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Dec 01 16:06:07 crc kubenswrapper[4810]: I1201 16:06:07.986093 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xd7fr\" (UniqueName: \"kubernetes.io/projected/76b5695c-7889-4c67-9441-68053d417d7d-kube-api-access-xd7fr\") pod \"76b5695c-7889-4c67-9441-68053d417d7d\" (UID: \"76b5695c-7889-4c67-9441-68053d417d7d\") " Dec 01 16:06:07 crc kubenswrapper[4810]: I1201 16:06:07.992810 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76b5695c-7889-4c67-9441-68053d417d7d-kube-api-access-xd7fr" (OuterVolumeSpecName: "kube-api-access-xd7fr") pod "76b5695c-7889-4c67-9441-68053d417d7d" (UID: "76b5695c-7889-4c67-9441-68053d417d7d"). InnerVolumeSpecName "kube-api-access-xd7fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:06:08 crc kubenswrapper[4810]: I1201 16:06:08.026176 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Dec 01 16:06:08 crc kubenswrapper[4810]: E1201 16:06:08.026559 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76b5695c-7889-4c67-9441-68053d417d7d" containerName="mariadb-client" Dec 01 16:06:08 crc kubenswrapper[4810]: I1201 16:06:08.026581 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="76b5695c-7889-4c67-9441-68053d417d7d" containerName="mariadb-client" Dec 01 16:06:08 crc kubenswrapper[4810]: I1201 16:06:08.026798 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="76b5695c-7889-4c67-9441-68053d417d7d" containerName="mariadb-client" Dec 01 16:06:08 crc kubenswrapper[4810]: I1201 16:06:08.027395 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 01 16:06:08 crc kubenswrapper[4810]: I1201 16:06:08.035568 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 01 16:06:08 crc kubenswrapper[4810]: I1201 16:06:08.088010 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngm4n\" (UniqueName: \"kubernetes.io/projected/e173f12d-189c-4bdb-897b-a3d897291360-kube-api-access-ngm4n\") pod \"mariadb-client\" (UID: \"e173f12d-189c-4bdb-897b-a3d897291360\") " pod="openstack/mariadb-client" Dec 01 16:06:08 crc kubenswrapper[4810]: I1201 16:06:08.088362 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xd7fr\" (UniqueName: \"kubernetes.io/projected/76b5695c-7889-4c67-9441-68053d417d7d-kube-api-access-xd7fr\") on node \"crc\" DevicePath \"\"" Dec 01 16:06:08 crc kubenswrapper[4810]: I1201 16:06:08.189682 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngm4n\" (UniqueName: \"kubernetes.io/projected/e173f12d-189c-4bdb-897b-a3d897291360-kube-api-access-ngm4n\") pod \"mariadb-client\" (UID: \"e173f12d-189c-4bdb-897b-a3d897291360\") " pod="openstack/mariadb-client" Dec 01 16:06:08 crc kubenswrapper[4810]: I1201 16:06:08.226807 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngm4n\" (UniqueName: \"kubernetes.io/projected/e173f12d-189c-4bdb-897b-a3d897291360-kube-api-access-ngm4n\") pod \"mariadb-client\" (UID: \"e173f12d-189c-4bdb-897b-a3d897291360\") " pod="openstack/mariadb-client" Dec 01 16:06:08 crc kubenswrapper[4810]: I1201 16:06:08.343823 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 01 16:06:08 crc kubenswrapper[4810]: I1201 16:06:08.508406 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76b5695c-7889-4c67-9441-68053d417d7d" path="/var/lib/kubelet/pods/76b5695c-7889-4c67-9441-68053d417d7d/volumes" Dec 01 16:06:08 crc kubenswrapper[4810]: I1201 16:06:08.517866 4810 scope.go:117] "RemoveContainer" containerID="57030bcd1c685bf36cb57fd2b533f165ece63a0ca0887c3dd33d223e31bd1dc0" Dec 01 16:06:08 crc kubenswrapper[4810]: I1201 16:06:08.518143 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 01 16:06:08 crc kubenswrapper[4810]: I1201 16:06:08.574319 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 01 16:06:08 crc kubenswrapper[4810]: W1201 16:06:08.579383 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode173f12d_189c_4bdb_897b_a3d897291360.slice/crio-92181fd2587c5b3eb4abcedb8107db19f300791a09851ce0af56ddc58235fa0f WatchSource:0}: Error finding container 92181fd2587c5b3eb4abcedb8107db19f300791a09851ce0af56ddc58235fa0f: Status 404 returned error can't find the container with id 92181fd2587c5b3eb4abcedb8107db19f300791a09851ce0af56ddc58235fa0f Dec 01 16:06:09 crc kubenswrapper[4810]: I1201 16:06:09.531372 4810 generic.go:334] "Generic (PLEG): container finished" podID="e173f12d-189c-4bdb-897b-a3d897291360" containerID="3f95994d6c1e8562e5f518ad159a1d053739f1f742b61c865a4bab6d18909ca7" exitCode=0 Dec 01 16:06:09 crc kubenswrapper[4810]: I1201 16:06:09.531431 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"e173f12d-189c-4bdb-897b-a3d897291360","Type":"ContainerDied","Data":"3f95994d6c1e8562e5f518ad159a1d053739f1f742b61c865a4bab6d18909ca7"} Dec 01 16:06:09 crc kubenswrapper[4810]: I1201 16:06:09.531744 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"e173f12d-189c-4bdb-897b-a3d897291360","Type":"ContainerStarted","Data":"92181fd2587c5b3eb4abcedb8107db19f300791a09851ce0af56ddc58235fa0f"} Dec 01 16:06:10 crc kubenswrapper[4810]: I1201 16:06:10.849461 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 01 16:06:10 crc kubenswrapper[4810]: I1201 16:06:10.871392 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_e173f12d-189c-4bdb-897b-a3d897291360/mariadb-client/0.log" Dec 01 16:06:10 crc kubenswrapper[4810]: I1201 16:06:10.895777 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Dec 01 16:06:10 crc kubenswrapper[4810]: I1201 16:06:10.900833 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Dec 01 16:06:10 crc kubenswrapper[4810]: I1201 16:06:10.940679 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngm4n\" (UniqueName: \"kubernetes.io/projected/e173f12d-189c-4bdb-897b-a3d897291360-kube-api-access-ngm4n\") pod \"e173f12d-189c-4bdb-897b-a3d897291360\" (UID: \"e173f12d-189c-4bdb-897b-a3d897291360\") " Dec 01 16:06:10 crc kubenswrapper[4810]: I1201 16:06:10.946405 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e173f12d-189c-4bdb-897b-a3d897291360-kube-api-access-ngm4n" (OuterVolumeSpecName: "kube-api-access-ngm4n") pod "e173f12d-189c-4bdb-897b-a3d897291360" (UID: "e173f12d-189c-4bdb-897b-a3d897291360"). InnerVolumeSpecName "kube-api-access-ngm4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:06:11 crc kubenswrapper[4810]: I1201 16:06:11.042316 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngm4n\" (UniqueName: \"kubernetes.io/projected/e173f12d-189c-4bdb-897b-a3d897291360-kube-api-access-ngm4n\") on node \"crc\" DevicePath \"\"" Dec 01 16:06:11 crc kubenswrapper[4810]: I1201 16:06:11.546001 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92181fd2587c5b3eb4abcedb8107db19f300791a09851ce0af56ddc58235fa0f" Dec 01 16:06:11 crc kubenswrapper[4810]: I1201 16:06:11.546239 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 01 16:06:12 crc kubenswrapper[4810]: I1201 16:06:12.499406 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e173f12d-189c-4bdb-897b-a3d897291360" path="/var/lib/kubelet/pods/e173f12d-189c-4bdb-897b-a3d897291360/volumes" Dec 01 16:06:18 crc kubenswrapper[4810]: I1201 16:06:18.491260 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:06:18 crc kubenswrapper[4810]: E1201 16:06:18.492058 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:06:31 crc kubenswrapper[4810]: I1201 16:06:31.490853 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:06:31 crc kubenswrapper[4810]: E1201 16:06:31.491625 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.230355 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 16:06:38 crc kubenswrapper[4810]: E1201 16:06:38.231115 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e173f12d-189c-4bdb-897b-a3d897291360" containerName="mariadb-client" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.231128 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e173f12d-189c-4bdb-897b-a3d897291360" containerName="mariadb-client" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.231266 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e173f12d-189c-4bdb-897b-a3d897291360" containerName="mariadb-client" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.232068 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.234413 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.234517 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.234576 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.235075 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-ss874" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.236510 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.246449 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.248280 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.255227 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.273437 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.274736 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.279900 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.304630 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.375263 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.375314 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-config\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.375347 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2edd15df-7752-4d42-b3e9-cfd79df85565-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.375367 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.375429 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1bcdc584-8478-437c-ba9d-ec55814a4293\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1bcdc584-8478-437c-ba9d-ec55814a4293\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.375495 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.375523 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.375559 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.375578 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2edd15df-7752-4d42-b3e9-cfd79df85565-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.375600 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2edd15df-7752-4d42-b3e9-cfd79df85565-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.375714 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-config\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.375765 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.375858 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e1637301-6f30-4ef5-a59a-375951711a40\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e1637301-6f30-4ef5-a59a-375951711a40\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.375920 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2edd15df-7752-4d42-b3e9-cfd79df85565-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.375945 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sfvj\" (UniqueName: \"kubernetes.io/projected/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-kube-api-access-5sfvj\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.375972 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2edd15df-7752-4d42-b3e9-cfd79df85565-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.375996 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2cf2\" (UniqueName: \"kubernetes.io/projected/2edd15df-7752-4d42-b3e9-cfd79df85565-kube-api-access-t2cf2\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.376033 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.376092 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.376132 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2edd15df-7752-4d42-b3e9-cfd79df85565-config\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.376192 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.376250 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-22e1d38b-8279-46f7-8b21-e242917a6893\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-22e1d38b-8279-46f7-8b21-e242917a6893\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.376283 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdhbz\" (UniqueName: \"kubernetes.io/projected/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-kube-api-access-hdhbz\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.376298 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.477848 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-config\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.478126 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2edd15df-7752-4d42-b3e9-cfd79df85565-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.478225 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.478354 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1bcdc584-8478-437c-ba9d-ec55814a4293\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1bcdc584-8478-437c-ba9d-ec55814a4293\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.478485 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.478600 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.478709 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.478791 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2edd15df-7752-4d42-b3e9-cfd79df85565-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.478875 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2edd15df-7752-4d42-b3e9-cfd79df85565-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.478967 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-config\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.479053 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.479124 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.479290 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-config\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.479372 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e1637301-6f30-4ef5-a59a-375951711a40\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e1637301-6f30-4ef5-a59a-375951711a40\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.479480 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2edd15df-7752-4d42-b3e9-cfd79df85565-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.479511 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sfvj\" (UniqueName: \"kubernetes.io/projected/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-kube-api-access-5sfvj\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.479533 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2edd15df-7752-4d42-b3e9-cfd79df85565-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.479552 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2cf2\" (UniqueName: \"kubernetes.io/projected/2edd15df-7752-4d42-b3e9-cfd79df85565-kube-api-access-t2cf2\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.479570 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.479591 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.479619 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2edd15df-7752-4d42-b3e9-cfd79df85565-config\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.479644 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.479669 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-22e1d38b-8279-46f7-8b21-e242917a6893\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-22e1d38b-8279-46f7-8b21-e242917a6893\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.479704 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdhbz\" (UniqueName: \"kubernetes.io/projected/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-kube-api-access-hdhbz\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.479725 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.479747 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2edd15df-7752-4d42-b3e9-cfd79df85565-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.479763 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.480181 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-config\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.480819 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2edd15df-7752-4d42-b3e9-cfd79df85565-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.481384 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.482022 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2edd15df-7752-4d42-b3e9-cfd79df85565-config\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.482170 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.482586 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.486682 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.487405 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.490604 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.490907 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e1637301-6f30-4ef5-a59a-375951711a40\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e1637301-6f30-4ef5-a59a-375951711a40\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c6015a1eea2677cd3e4bf5ef563874b4ff697b9942000a4d0c67b954e75b178e/globalmount\"" pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.490608 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.490983 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1bcdc584-8478-437c-ba9d-ec55814a4293\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1bcdc584-8478-437c-ba9d-ec55814a4293\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/71b35011e05ab93367710b27c9f01da8672e019c3268916f44f0ffc334754c2d/globalmount\"" pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.490820 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.491083 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-22e1d38b-8279-46f7-8b21-e242917a6893\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-22e1d38b-8279-46f7-8b21-e242917a6893\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e708df1c3916c4c347673bd48e70a5a9e3e77c5f5b58939010f2c9f9c00d44b3/globalmount\"" pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.491374 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2edd15df-7752-4d42-b3e9-cfd79df85565-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.491441 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.493186 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.496096 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2edd15df-7752-4d42-b3e9-cfd79df85565-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.497101 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.498614 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2edd15df-7752-4d42-b3e9-cfd79df85565-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.498750 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.503498 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdhbz\" (UniqueName: \"kubernetes.io/projected/418356d0-f609-4fa5-a31a-2b3c83a6f5d7-kube-api-access-hdhbz\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.503895 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sfvj\" (UniqueName: \"kubernetes.io/projected/d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05-kube-api-access-5sfvj\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.506263 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2cf2\" (UniqueName: \"kubernetes.io/projected/2edd15df-7752-4d42-b3e9-cfd79df85565-kube-api-access-t2cf2\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.536058 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1bcdc584-8478-437c-ba9d-ec55814a4293\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1bcdc584-8478-437c-ba9d-ec55814a4293\") pod \"ovsdbserver-nb-1\" (UID: \"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05\") " pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.544444 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e1637301-6f30-4ef5-a59a-375951711a40\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e1637301-6f30-4ef5-a59a-375951711a40\") pod \"ovsdbserver-nb-2\" (UID: \"418356d0-f609-4fa5-a31a-2b3c83a6f5d7\") " pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.544922 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-22e1d38b-8279-46f7-8b21-e242917a6893\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-22e1d38b-8279-46f7-8b21-e242917a6893\") pod \"ovsdbserver-nb-0\" (UID: \"2edd15df-7752-4d42-b3e9-cfd79df85565\") " pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.553623 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.585164 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:38 crc kubenswrapper[4810]: I1201 16:06:38.606761 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:39 crc kubenswrapper[4810]: I1201 16:06:39.158853 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 16:06:39 crc kubenswrapper[4810]: W1201 16:06:39.177619 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2edd15df_7752_4d42_b3e9_cfd79df85565.slice/crio-b47cd6442ecc6bd8681ca45ba435a8149007a6cd65439824b6dc22f80b99588d WatchSource:0}: Error finding container b47cd6442ecc6bd8681ca45ba435a8149007a6cd65439824b6dc22f80b99588d: Status 404 returned error can't find the container with id b47cd6442ecc6bd8681ca45ba435a8149007a6cd65439824b6dc22f80b99588d Dec 01 16:06:39 crc kubenswrapper[4810]: I1201 16:06:39.271584 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 01 16:06:39 crc kubenswrapper[4810]: W1201 16:06:39.282785 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7d2b7f9_4a49_45e2_abfd_a55a3f5d3a05.slice/crio-2f39ec39fae772addfd9d3c9c8846718f47b27b83d4a0c1af2a3f60899350f29 WatchSource:0}: Error finding container 2f39ec39fae772addfd9d3c9c8846718f47b27b83d4a0c1af2a3f60899350f29: Status 404 returned error can't find the container with id 2f39ec39fae772addfd9d3c9c8846718f47b27b83d4a0c1af2a3f60899350f29 Dec 01 16:06:39 crc kubenswrapper[4810]: I1201 16:06:39.775355 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"2edd15df-7752-4d42-b3e9-cfd79df85565","Type":"ContainerStarted","Data":"3ec638b64bb76284434c4c714db4710160df008cbf1fda46692ab99ee6f856bd"} Dec 01 16:06:39 crc kubenswrapper[4810]: I1201 16:06:39.775655 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"2edd15df-7752-4d42-b3e9-cfd79df85565","Type":"ContainerStarted","Data":"c085b505b55a9ae1f799ca367b00bf49357de3d8639af8481b55abd56292bf62"} Dec 01 16:06:39 crc kubenswrapper[4810]: I1201 16:06:39.775669 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"2edd15df-7752-4d42-b3e9-cfd79df85565","Type":"ContainerStarted","Data":"b47cd6442ecc6bd8681ca45ba435a8149007a6cd65439824b6dc22f80b99588d"} Dec 01 16:06:39 crc kubenswrapper[4810]: I1201 16:06:39.785245 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05","Type":"ContainerStarted","Data":"8c27136ef923b855f9d3448477ae591c95da77b747fc0452a618273785611384"} Dec 01 16:06:39 crc kubenswrapper[4810]: I1201 16:06:39.785409 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05","Type":"ContainerStarted","Data":"2f39ec39fae772addfd9d3c9c8846718f47b27b83d4a0c1af2a3f60899350f29"} Dec 01 16:06:39 crc kubenswrapper[4810]: I1201 16:06:39.799141 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=2.799119065 podStartE2EDuration="2.799119065s" podCreationTimestamp="2025-12-01 16:06:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:06:39.797743287 +0000 UTC m=+5565.561252910" watchObservedRunningTime="2025-12-01 16:06:39.799119065 +0000 UTC m=+5565.562628668" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.231062 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 01 16:06:40 crc kubenswrapper[4810]: W1201 16:06:40.231129 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod418356d0_f609_4fa5_a31a_2b3c83a6f5d7.slice/crio-5fe367a238e514ba0821cfc1823ed89e932b44247e220b9397333f1665d5618c WatchSource:0}: Error finding container 5fe367a238e514ba0821cfc1823ed89e932b44247e220b9397333f1665d5618c: Status 404 returned error can't find the container with id 5fe367a238e514ba0821cfc1823ed89e932b44247e220b9397333f1665d5618c Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.574155 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.576131 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.577988 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-hk9c4" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.578928 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.579728 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.581131 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.589788 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.604979 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.610811 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.611276 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.612726 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.616026 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2db4252-6bc3-40d8-a375-03660d2aeae0-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.616065 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d2db4252-6bc3-40d8-a375-03660d2aeae0-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.616095 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2db4252-6bc3-40d8-a375-03660d2aeae0-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.616126 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2db4252-6bc3-40d8-a375-03660d2aeae0-config\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.616149 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5fa9b747-90e3-4dcc-9fe0-1463f9747484\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5fa9b747-90e3-4dcc-9fe0-1463f9747484\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.616185 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2db4252-6bc3-40d8-a375-03660d2aeae0-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.616255 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2db4252-6bc3-40d8-a375-03660d2aeae0-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.616292 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7k25\" (UniqueName: \"kubernetes.io/projected/d2db4252-6bc3-40d8-a375-03660d2aeae0-kube-api-access-t7k25\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.620162 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.637093 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.717949 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b15e71bf-2d3e-47b1-9013-a051eb580331-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718038 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bktk\" (UniqueName: \"kubernetes.io/projected/b15e71bf-2d3e-47b1-9013-a051eb580331-kube-api-access-8bktk\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718085 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2db4252-6bc3-40d8-a375-03660d2aeae0-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718106 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b15e71bf-2d3e-47b1-9013-a051eb580331-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718155 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b15e71bf-2d3e-47b1-9013-a051eb580331-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718185 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7k25\" (UniqueName: \"kubernetes.io/projected/d2db4252-6bc3-40d8-a375-03660d2aeae0-kube-api-access-t7k25\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718216 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd18cf90-4008-4cf1-953e-d0caf3382dd2-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718238 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-51aad043-88b0-43df-8ca7-826e23c919a2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-51aad043-88b0-43df-8ca7-826e23c919a2\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718261 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd18cf90-4008-4cf1-953e-d0caf3382dd2-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718278 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b15e71bf-2d3e-47b1-9013-a051eb580331-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718301 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhmwz\" (UniqueName: \"kubernetes.io/projected/fd18cf90-4008-4cf1-953e-d0caf3382dd2-kube-api-access-xhmwz\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718333 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd18cf90-4008-4cf1-953e-d0caf3382dd2-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718645 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b15e71bf-2d3e-47b1-9013-a051eb580331-config\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718679 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-16be9be8-9e35-4ab3-8340-572f2e4b43da\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16be9be8-9e35-4ab3-8340-572f2e4b43da\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718777 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b15e71bf-2d3e-47b1-9013-a051eb580331-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718900 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2db4252-6bc3-40d8-a375-03660d2aeae0-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718930 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd18cf90-4008-4cf1-953e-d0caf3382dd2-config\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718946 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd18cf90-4008-4cf1-953e-d0caf3382dd2-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718966 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d2db4252-6bc3-40d8-a375-03660d2aeae0-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.718989 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2db4252-6bc3-40d8-a375-03660d2aeae0-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.719012 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fd18cf90-4008-4cf1-953e-d0caf3382dd2-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.719042 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2db4252-6bc3-40d8-a375-03660d2aeae0-config\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.719061 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5fa9b747-90e3-4dcc-9fe0-1463f9747484\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5fa9b747-90e3-4dcc-9fe0-1463f9747484\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.719098 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2db4252-6bc3-40d8-a375-03660d2aeae0-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.719639 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d2db4252-6bc3-40d8-a375-03660d2aeae0-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.720062 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2db4252-6bc3-40d8-a375-03660d2aeae0-config\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.720205 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2db4252-6bc3-40d8-a375-03660d2aeae0-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.722183 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.722211 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5fa9b747-90e3-4dcc-9fe0-1463f9747484\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5fa9b747-90e3-4dcc-9fe0-1463f9747484\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a4c5b8618fee614654a0f1bf5bb4f91673b65b8914b448fdde69596f72bd0c60/globalmount\"" pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.726107 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2db4252-6bc3-40d8-a375-03660d2aeae0-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.726324 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2db4252-6bc3-40d8-a375-03660d2aeae0-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.730229 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2db4252-6bc3-40d8-a375-03660d2aeae0-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.734348 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7k25\" (UniqueName: \"kubernetes.io/projected/d2db4252-6bc3-40d8-a375-03660d2aeae0-kube-api-access-t7k25\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.756214 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5fa9b747-90e3-4dcc-9fe0-1463f9747484\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5fa9b747-90e3-4dcc-9fe0-1463f9747484\") pod \"ovsdbserver-sb-0\" (UID: \"d2db4252-6bc3-40d8-a375-03660d2aeae0\") " pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.801694 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"418356d0-f609-4fa5-a31a-2b3c83a6f5d7","Type":"ContainerStarted","Data":"9fcd818cd1485ab82f7322254442d5ac0847344de032f22eea4684d324436a7d"} Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.801737 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"418356d0-f609-4fa5-a31a-2b3c83a6f5d7","Type":"ContainerStarted","Data":"b7ccec35a85646271352983afa1392d6699db05262d52a6c9a91630cc1dd0ec2"} Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.801748 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"418356d0-f609-4fa5-a31a-2b3c83a6f5d7","Type":"ContainerStarted","Data":"5fe367a238e514ba0821cfc1823ed89e932b44247e220b9397333f1665d5618c"} Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.805720 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05","Type":"ContainerStarted","Data":"7ce5298ba5bd5c79c8f734eff8cb236659da330a1a21c8c263d996623ca4ff01"} Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.820424 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b15e71bf-2d3e-47b1-9013-a051eb580331-config\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.820486 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-16be9be8-9e35-4ab3-8340-572f2e4b43da\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16be9be8-9e35-4ab3-8340-572f2e4b43da\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.820521 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b15e71bf-2d3e-47b1-9013-a051eb580331-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.821214 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd18cf90-4008-4cf1-953e-d0caf3382dd2-config\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.821232 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd18cf90-4008-4cf1-953e-d0caf3382dd2-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.821277 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fd18cf90-4008-4cf1-953e-d0caf3382dd2-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.821325 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b15e71bf-2d3e-47b1-9013-a051eb580331-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.821374 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bktk\" (UniqueName: \"kubernetes.io/projected/b15e71bf-2d3e-47b1-9013-a051eb580331-kube-api-access-8bktk\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.821422 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b15e71bf-2d3e-47b1-9013-a051eb580331-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.821445 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b15e71bf-2d3e-47b1-9013-a051eb580331-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.821502 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd18cf90-4008-4cf1-953e-d0caf3382dd2-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.821520 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-51aad043-88b0-43df-8ca7-826e23c919a2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-51aad043-88b0-43df-8ca7-826e23c919a2\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.821536 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd18cf90-4008-4cf1-953e-d0caf3382dd2-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.821551 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b15e71bf-2d3e-47b1-9013-a051eb580331-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.825051 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhmwz\" (UniqueName: \"kubernetes.io/projected/fd18cf90-4008-4cf1-953e-d0caf3382dd2-kube-api-access-xhmwz\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.822346 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd18cf90-4008-4cf1-953e-d0caf3382dd2-config\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.825092 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd18cf90-4008-4cf1-953e-d0caf3382dd2-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.823114 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b15e71bf-2d3e-47b1-9013-a051eb580331-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.824041 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b15e71bf-2d3e-47b1-9013-a051eb580331-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.824063 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b15e71bf-2d3e-47b1-9013-a051eb580331-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.821334 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b15e71bf-2d3e-47b1-9013-a051eb580331-config\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.822797 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.825208 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-16be9be8-9e35-4ab3-8340-572f2e4b43da\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16be9be8-9e35-4ab3-8340-572f2e4b43da\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a27e72d59b2c4ca2616ac492e6d380eb7ee2aa883f9cea9edf7854a37b7db81c/globalmount\"" pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.822978 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fd18cf90-4008-4cf1-953e-d0caf3382dd2-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.826345 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd18cf90-4008-4cf1-953e-d0caf3382dd2-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.826978 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b15e71bf-2d3e-47b1-9013-a051eb580331-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.831351 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.831626 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-51aad043-88b0-43df-8ca7-826e23c919a2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-51aad043-88b0-43df-8ca7-826e23c919a2\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5584517e0de1dcf7088edd44b5592a14a76d93fc07f844329a827632e9c3a7e5/globalmount\"" pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.832767 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd18cf90-4008-4cf1-953e-d0caf3382dd2-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.839267 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=3.83924244 podStartE2EDuration="3.83924244s" podCreationTimestamp="2025-12-01 16:06:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:06:40.822435003 +0000 UTC m=+5566.585944606" watchObservedRunningTime="2025-12-01 16:06:40.83924244 +0000 UTC m=+5566.602752043" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.839785 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b15e71bf-2d3e-47b1-9013-a051eb580331-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.840965 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd18cf90-4008-4cf1-953e-d0caf3382dd2-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.842761 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd18cf90-4008-4cf1-953e-d0caf3382dd2-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.842971 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhmwz\" (UniqueName: \"kubernetes.io/projected/fd18cf90-4008-4cf1-953e-d0caf3382dd2-kube-api-access-xhmwz\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.847728 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bktk\" (UniqueName: \"kubernetes.io/projected/b15e71bf-2d3e-47b1-9013-a051eb580331-kube-api-access-8bktk\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.853498 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=3.8534597059999998 podStartE2EDuration="3.853459706s" podCreationTimestamp="2025-12-01 16:06:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:06:40.848189673 +0000 UTC m=+5566.611699276" watchObservedRunningTime="2025-12-01 16:06:40.853459706 +0000 UTC m=+5566.616969309" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.869577 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-51aad043-88b0-43df-8ca7-826e23c919a2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-51aad043-88b0-43df-8ca7-826e23c919a2\") pod \"ovsdbserver-sb-1\" (UID: \"fd18cf90-4008-4cf1-953e-d0caf3382dd2\") " pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.872983 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-16be9be8-9e35-4ab3-8340-572f2e4b43da\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16be9be8-9e35-4ab3-8340-572f2e4b43da\") pod \"ovsdbserver-sb-2\" (UID: \"b15e71bf-2d3e-47b1-9013-a051eb580331\") " pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.910771 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.934336 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:40 crc kubenswrapper[4810]: I1201 16:06:40.943279 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:41 crc kubenswrapper[4810]: I1201 16:06:41.486235 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 16:06:41 crc kubenswrapper[4810]: W1201 16:06:41.522154 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2db4252_6bc3_40d8_a375_03660d2aeae0.slice/crio-85f7a4ffe6d8e196e1666e878b8db501e7dc319bd2a268ecd764a232dcc45070 WatchSource:0}: Error finding container 85f7a4ffe6d8e196e1666e878b8db501e7dc319bd2a268ecd764a232dcc45070: Status 404 returned error can't find the container with id 85f7a4ffe6d8e196e1666e878b8db501e7dc319bd2a268ecd764a232dcc45070 Dec 01 16:06:41 crc kubenswrapper[4810]: I1201 16:06:41.554449 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:41 crc kubenswrapper[4810]: I1201 16:06:41.590576 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 01 16:06:41 crc kubenswrapper[4810]: I1201 16:06:41.590796 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:41 crc kubenswrapper[4810]: W1201 16:06:41.593528 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd18cf90_4008_4cf1_953e_d0caf3382dd2.slice/crio-4bcf722336f755e386fd2d1105e3a1f224a7977a1cde45823814aa07c2c3a82d WatchSource:0}: Error finding container 4bcf722336f755e386fd2d1105e3a1f224a7977a1cde45823814aa07c2c3a82d: Status 404 returned error can't find the container with id 4bcf722336f755e386fd2d1105e3a1f224a7977a1cde45823814aa07c2c3a82d Dec 01 16:06:41 crc kubenswrapper[4810]: I1201 16:06:41.607280 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:41 crc kubenswrapper[4810]: I1201 16:06:41.816375 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"fd18cf90-4008-4cf1-953e-d0caf3382dd2","Type":"ContainerStarted","Data":"e32ceffadbce3af64829f7012f08f70472b90951fc93edb8d193625935fa0283"} Dec 01 16:06:41 crc kubenswrapper[4810]: I1201 16:06:41.816429 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"fd18cf90-4008-4cf1-953e-d0caf3382dd2","Type":"ContainerStarted","Data":"4bcf722336f755e386fd2d1105e3a1f224a7977a1cde45823814aa07c2c3a82d"} Dec 01 16:06:41 crc kubenswrapper[4810]: I1201 16:06:41.818214 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d2db4252-6bc3-40d8-a375-03660d2aeae0","Type":"ContainerStarted","Data":"3f5f23f2d344c4a39f60edf40adc527e283f2436ad2500a6c96eb2eda1cda1b2"} Dec 01 16:06:41 crc kubenswrapper[4810]: I1201 16:06:41.818260 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d2db4252-6bc3-40d8-a375-03660d2aeae0","Type":"ContainerStarted","Data":"85f7a4ffe6d8e196e1666e878b8db501e7dc319bd2a268ecd764a232dcc45070"} Dec 01 16:06:42 crc kubenswrapper[4810]: I1201 16:06:42.304588 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 01 16:06:42 crc kubenswrapper[4810]: W1201 16:06:42.312088 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb15e71bf_2d3e_47b1_9013_a051eb580331.slice/crio-cbe16c81127dbca8c3b8c398ed53b9e4ac15eab664df089d3b1c0f5ab7c6dbda WatchSource:0}: Error finding container cbe16c81127dbca8c3b8c398ed53b9e4ac15eab664df089d3b1c0f5ab7c6dbda: Status 404 returned error can't find the container with id cbe16c81127dbca8c3b8c398ed53b9e4ac15eab664df089d3b1c0f5ab7c6dbda Dec 01 16:06:42 crc kubenswrapper[4810]: I1201 16:06:42.827393 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"b15e71bf-2d3e-47b1-9013-a051eb580331","Type":"ContainerStarted","Data":"b39532731fed6ed763685e69ea6e67f9ef6d02bfd849f6a0b40130140d2da2b9"} Dec 01 16:06:42 crc kubenswrapper[4810]: I1201 16:06:42.827455 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"b15e71bf-2d3e-47b1-9013-a051eb580331","Type":"ContainerStarted","Data":"069ff20f6b764e64efa646f09de670c07c45ca528cc295c713d550c8aaa34341"} Dec 01 16:06:42 crc kubenswrapper[4810]: I1201 16:06:42.827497 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"b15e71bf-2d3e-47b1-9013-a051eb580331","Type":"ContainerStarted","Data":"cbe16c81127dbca8c3b8c398ed53b9e4ac15eab664df089d3b1c0f5ab7c6dbda"} Dec 01 16:06:42 crc kubenswrapper[4810]: I1201 16:06:42.830571 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"fd18cf90-4008-4cf1-953e-d0caf3382dd2","Type":"ContainerStarted","Data":"e88c18eb67313ce94a6f702cfbcc9c03bce468e46efe66de5ed6b869e33bf45f"} Dec 01 16:06:42 crc kubenswrapper[4810]: I1201 16:06:42.832334 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d2db4252-6bc3-40d8-a375-03660d2aeae0","Type":"ContainerStarted","Data":"7318f7fdfe2f4c2f680271c7697ea43d8b846665a7f6bbab4ebd8d3c4eb97222"} Dec 01 16:06:42 crc kubenswrapper[4810]: I1201 16:06:42.859436 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=3.859416818 podStartE2EDuration="3.859416818s" podCreationTimestamp="2025-12-01 16:06:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:06:42.850430163 +0000 UTC m=+5568.613939776" watchObservedRunningTime="2025-12-01 16:06:42.859416818 +0000 UTC m=+5568.622926421" Dec 01 16:06:42 crc kubenswrapper[4810]: I1201 16:06:42.874734 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=3.874713893 podStartE2EDuration="3.874713893s" podCreationTimestamp="2025-12-01 16:06:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:06:42.86870916 +0000 UTC m=+5568.632218763" watchObservedRunningTime="2025-12-01 16:06:42.874713893 +0000 UTC m=+5568.638223496" Dec 01 16:06:42 crc kubenswrapper[4810]: I1201 16:06:42.890641 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.890623796 podStartE2EDuration="3.890623796s" podCreationTimestamp="2025-12-01 16:06:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:06:42.88821229 +0000 UTC m=+5568.651721893" watchObservedRunningTime="2025-12-01 16:06:42.890623796 +0000 UTC m=+5568.654133399" Dec 01 16:06:43 crc kubenswrapper[4810]: I1201 16:06:43.554865 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:43 crc kubenswrapper[4810]: I1201 16:06:43.586570 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:43 crc kubenswrapper[4810]: I1201 16:06:43.607247 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:43 crc kubenswrapper[4810]: I1201 16:06:43.912049 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:43 crc kubenswrapper[4810]: I1201 16:06:43.935398 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:43 crc kubenswrapper[4810]: I1201 16:06:43.943859 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.590774 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.626951 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.632798 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.648486 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.820282 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85d9c479c-zpq5m"] Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.822114 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.827305 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.833847 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef742dba-5fbf-499f-9e26-1af4eeb10c28-dns-svc\") pod \"dnsmasq-dns-85d9c479c-zpq5m\" (UID: \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\") " pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.833926 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef742dba-5fbf-499f-9e26-1af4eeb10c28-ovsdbserver-nb\") pod \"dnsmasq-dns-85d9c479c-zpq5m\" (UID: \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\") " pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.833956 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgwmh\" (UniqueName: \"kubernetes.io/projected/ef742dba-5fbf-499f-9e26-1af4eeb10c28-kube-api-access-sgwmh\") pod \"dnsmasq-dns-85d9c479c-zpq5m\" (UID: \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\") " pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.834017 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef742dba-5fbf-499f-9e26-1af4eeb10c28-config\") pod \"dnsmasq-dns-85d9c479c-zpq5m\" (UID: \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\") " pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.841339 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85d9c479c-zpq5m"] Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.918557 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.935853 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef742dba-5fbf-499f-9e26-1af4eeb10c28-config\") pod \"dnsmasq-dns-85d9c479c-zpq5m\" (UID: \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\") " pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.935984 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef742dba-5fbf-499f-9e26-1af4eeb10c28-dns-svc\") pod \"dnsmasq-dns-85d9c479c-zpq5m\" (UID: \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\") " pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.936046 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef742dba-5fbf-499f-9e26-1af4eeb10c28-ovsdbserver-nb\") pod \"dnsmasq-dns-85d9c479c-zpq5m\" (UID: \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\") " pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.936074 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgwmh\" (UniqueName: \"kubernetes.io/projected/ef742dba-5fbf-499f-9e26-1af4eeb10c28-kube-api-access-sgwmh\") pod \"dnsmasq-dns-85d9c479c-zpq5m\" (UID: \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\") " pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.937127 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef742dba-5fbf-499f-9e26-1af4eeb10c28-config\") pod \"dnsmasq-dns-85d9c479c-zpq5m\" (UID: \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\") " pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.937173 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef742dba-5fbf-499f-9e26-1af4eeb10c28-ovsdbserver-nb\") pod \"dnsmasq-dns-85d9c479c-zpq5m\" (UID: \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\") " pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.939010 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef742dba-5fbf-499f-9e26-1af4eeb10c28-dns-svc\") pod \"dnsmasq-dns-85d9c479c-zpq5m\" (UID: \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\") " pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" Dec 01 16:06:44 crc kubenswrapper[4810]: I1201 16:06:44.974793 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgwmh\" (UniqueName: \"kubernetes.io/projected/ef742dba-5fbf-499f-9e26-1af4eeb10c28-kube-api-access-sgwmh\") pod \"dnsmasq-dns-85d9c479c-zpq5m\" (UID: \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\") " pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" Dec 01 16:06:45 crc kubenswrapper[4810]: I1201 16:06:45.179533 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" Dec 01 16:06:45 crc kubenswrapper[4810]: I1201 16:06:45.493944 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:06:45 crc kubenswrapper[4810]: E1201 16:06:45.496065 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:06:45 crc kubenswrapper[4810]: I1201 16:06:45.603227 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85d9c479c-zpq5m"] Dec 01 16:06:45 crc kubenswrapper[4810]: I1201 16:06:45.863317 4810 generic.go:334] "Generic (PLEG): container finished" podID="ef742dba-5fbf-499f-9e26-1af4eeb10c28" containerID="96a2a8f5a0629a212e30106038bd3d9973d0253cd2077dc18f9e59165d979cae" exitCode=0 Dec 01 16:06:45 crc kubenswrapper[4810]: I1201 16:06:45.863942 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" event={"ID":"ef742dba-5fbf-499f-9e26-1af4eeb10c28","Type":"ContainerDied","Data":"96a2a8f5a0629a212e30106038bd3d9973d0253cd2077dc18f9e59165d979cae"} Dec 01 16:06:45 crc kubenswrapper[4810]: I1201 16:06:45.864021 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" event={"ID":"ef742dba-5fbf-499f-9e26-1af4eeb10c28","Type":"ContainerStarted","Data":"71c94453bbe9afab16aa3dd8dd36b88300d6bf2e47bbae513a399bc44f32c40e"} Dec 01 16:06:45 crc kubenswrapper[4810]: I1201 16:06:45.911532 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:45 crc kubenswrapper[4810]: I1201 16:06:45.934490 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:45 crc kubenswrapper[4810]: I1201 16:06:45.943634 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:46 crc kubenswrapper[4810]: I1201 16:06:46.876528 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" event={"ID":"ef742dba-5fbf-499f-9e26-1af4eeb10c28","Type":"ContainerStarted","Data":"8a6c1d58ec2175031db5f9f96da1828b53f2c0f7cc5b88e628dab6b902ec69f8"} Dec 01 16:06:46 crc kubenswrapper[4810]: I1201 16:06:46.877133 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" Dec 01 16:06:46 crc kubenswrapper[4810]: I1201 16:06:46.913703 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" podStartSLOduration=2.913678601 podStartE2EDuration="2.913678601s" podCreationTimestamp="2025-12-01 16:06:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:06:46.899741222 +0000 UTC m=+5572.663250825" watchObservedRunningTime="2025-12-01 16:06:46.913678601 +0000 UTC m=+5572.677188204" Dec 01 16:06:46 crc kubenswrapper[4810]: I1201 16:06:46.956110 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:46 crc kubenswrapper[4810]: I1201 16:06:46.979770 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:46 crc kubenswrapper[4810]: I1201 16:06:46.992646 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:46 crc kubenswrapper[4810]: I1201 16:06:46.997413 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.035435 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.324726 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85d9c479c-zpq5m"] Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.364678 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69f88df96f-gfr2k"] Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.366434 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.368964 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.377275 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69f88df96f-gfr2k"] Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.482965 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-ovsdbserver-sb\") pod \"dnsmasq-dns-69f88df96f-gfr2k\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.483187 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-ovsdbserver-nb\") pod \"dnsmasq-dns-69f88df96f-gfr2k\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.483316 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-dns-svc\") pod \"dnsmasq-dns-69f88df96f-gfr2k\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.483358 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-config\") pod \"dnsmasq-dns-69f88df96f-gfr2k\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.483396 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9p2z\" (UniqueName: \"kubernetes.io/projected/a55b0531-5dbb-4b9e-9c06-685e84258744-kube-api-access-t9p2z\") pod \"dnsmasq-dns-69f88df96f-gfr2k\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.584821 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-ovsdbserver-nb\") pod \"dnsmasq-dns-69f88df96f-gfr2k\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.584894 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-dns-svc\") pod \"dnsmasq-dns-69f88df96f-gfr2k\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.584918 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-config\") pod \"dnsmasq-dns-69f88df96f-gfr2k\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.584945 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9p2z\" (UniqueName: \"kubernetes.io/projected/a55b0531-5dbb-4b9e-9c06-685e84258744-kube-api-access-t9p2z\") pod \"dnsmasq-dns-69f88df96f-gfr2k\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.584986 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-ovsdbserver-sb\") pod \"dnsmasq-dns-69f88df96f-gfr2k\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.586138 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-ovsdbserver-nb\") pod \"dnsmasq-dns-69f88df96f-gfr2k\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.586151 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-dns-svc\") pod \"dnsmasq-dns-69f88df96f-gfr2k\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.586263 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-ovsdbserver-sb\") pod \"dnsmasq-dns-69f88df96f-gfr2k\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.586333 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-config\") pod \"dnsmasq-dns-69f88df96f-gfr2k\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.606356 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9p2z\" (UniqueName: \"kubernetes.io/projected/a55b0531-5dbb-4b9e-9c06-685e84258744-kube-api-access-t9p2z\") pod \"dnsmasq-dns-69f88df96f-gfr2k\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:47 crc kubenswrapper[4810]: I1201 16:06:47.684853 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:48 crc kubenswrapper[4810]: I1201 16:06:48.112500 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69f88df96f-gfr2k"] Dec 01 16:06:48 crc kubenswrapper[4810]: W1201 16:06:48.119675 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda55b0531_5dbb_4b9e_9c06_685e84258744.slice/crio-e9de7f1dd2ad6c232b237a0d73348dc0411e446f27af357e82990d4a43c4cb03 WatchSource:0}: Error finding container e9de7f1dd2ad6c232b237a0d73348dc0411e446f27af357e82990d4a43c4cb03: Status 404 returned error can't find the container with id e9de7f1dd2ad6c232b237a0d73348dc0411e446f27af357e82990d4a43c4cb03 Dec 01 16:06:48 crc kubenswrapper[4810]: I1201 16:06:48.636212 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Dec 01 16:06:48 crc kubenswrapper[4810]: I1201 16:06:48.894786 4810 generic.go:334] "Generic (PLEG): container finished" podID="a55b0531-5dbb-4b9e-9c06-685e84258744" containerID="bf56c647e2ac9118ac042c77ec638d514850c3f9df8ba66a82e4860a1a88091d" exitCode=0 Dec 01 16:06:48 crc kubenswrapper[4810]: I1201 16:06:48.894851 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" event={"ID":"a55b0531-5dbb-4b9e-9c06-685e84258744","Type":"ContainerDied","Data":"bf56c647e2ac9118ac042c77ec638d514850c3f9df8ba66a82e4860a1a88091d"} Dec 01 16:06:48 crc kubenswrapper[4810]: I1201 16:06:48.894880 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" event={"ID":"a55b0531-5dbb-4b9e-9c06-685e84258744","Type":"ContainerStarted","Data":"e9de7f1dd2ad6c232b237a0d73348dc0411e446f27af357e82990d4a43c4cb03"} Dec 01 16:06:48 crc kubenswrapper[4810]: I1201 16:06:48.895219 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" podUID="ef742dba-5fbf-499f-9e26-1af4eeb10c28" containerName="dnsmasq-dns" containerID="cri-o://8a6c1d58ec2175031db5f9f96da1828b53f2c0f7cc5b88e628dab6b902ec69f8" gracePeriod=10 Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.264083 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.418314 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgwmh\" (UniqueName: \"kubernetes.io/projected/ef742dba-5fbf-499f-9e26-1af4eeb10c28-kube-api-access-sgwmh\") pod \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\" (UID: \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\") " Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.418431 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef742dba-5fbf-499f-9e26-1af4eeb10c28-dns-svc\") pod \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\" (UID: \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\") " Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.418646 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef742dba-5fbf-499f-9e26-1af4eeb10c28-ovsdbserver-nb\") pod \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\" (UID: \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\") " Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.418690 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef742dba-5fbf-499f-9e26-1af4eeb10c28-config\") pod \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\" (UID: \"ef742dba-5fbf-499f-9e26-1af4eeb10c28\") " Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.422880 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef742dba-5fbf-499f-9e26-1af4eeb10c28-kube-api-access-sgwmh" (OuterVolumeSpecName: "kube-api-access-sgwmh") pod "ef742dba-5fbf-499f-9e26-1af4eeb10c28" (UID: "ef742dba-5fbf-499f-9e26-1af4eeb10c28"). InnerVolumeSpecName "kube-api-access-sgwmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.457345 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef742dba-5fbf-499f-9e26-1af4eeb10c28-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ef742dba-5fbf-499f-9e26-1af4eeb10c28" (UID: "ef742dba-5fbf-499f-9e26-1af4eeb10c28"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.457992 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef742dba-5fbf-499f-9e26-1af4eeb10c28-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ef742dba-5fbf-499f-9e26-1af4eeb10c28" (UID: "ef742dba-5fbf-499f-9e26-1af4eeb10c28"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.458287 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef742dba-5fbf-499f-9e26-1af4eeb10c28-config" (OuterVolumeSpecName: "config") pod "ef742dba-5fbf-499f-9e26-1af4eeb10c28" (UID: "ef742dba-5fbf-499f-9e26-1af4eeb10c28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.521339 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef742dba-5fbf-499f-9e26-1af4eeb10c28-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.521829 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef742dba-5fbf-499f-9e26-1af4eeb10c28-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.521903 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef742dba-5fbf-499f-9e26-1af4eeb10c28-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.521960 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgwmh\" (UniqueName: \"kubernetes.io/projected/ef742dba-5fbf-499f-9e26-1af4eeb10c28-kube-api-access-sgwmh\") on node \"crc\" DevicePath \"\"" Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.903613 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" event={"ID":"a55b0531-5dbb-4b9e-9c06-685e84258744","Type":"ContainerStarted","Data":"8d04cf00357fea3fd3f057931b4085fa3294bc435cb5fb2c9f5a217406d808bc"} Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.904033 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.906478 4810 generic.go:334] "Generic (PLEG): container finished" podID="ef742dba-5fbf-499f-9e26-1af4eeb10c28" containerID="8a6c1d58ec2175031db5f9f96da1828b53f2c0f7cc5b88e628dab6b902ec69f8" exitCode=0 Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.906523 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.906522 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" event={"ID":"ef742dba-5fbf-499f-9e26-1af4eeb10c28","Type":"ContainerDied","Data":"8a6c1d58ec2175031db5f9f96da1828b53f2c0f7cc5b88e628dab6b902ec69f8"} Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.906650 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85d9c479c-zpq5m" event={"ID":"ef742dba-5fbf-499f-9e26-1af4eeb10c28","Type":"ContainerDied","Data":"71c94453bbe9afab16aa3dd8dd36b88300d6bf2e47bbae513a399bc44f32c40e"} Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.906729 4810 scope.go:117] "RemoveContainer" containerID="8a6c1d58ec2175031db5f9f96da1828b53f2c0f7cc5b88e628dab6b902ec69f8" Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.925000 4810 scope.go:117] "RemoveContainer" containerID="96a2a8f5a0629a212e30106038bd3d9973d0253cd2077dc18f9e59165d979cae" Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.928850 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" podStartSLOduration=2.928834358 podStartE2EDuration="2.928834358s" podCreationTimestamp="2025-12-01 16:06:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:06:49.925308731 +0000 UTC m=+5575.688818334" watchObservedRunningTime="2025-12-01 16:06:49.928834358 +0000 UTC m=+5575.692343961" Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.945507 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85d9c479c-zpq5m"] Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.952854 4810 scope.go:117] "RemoveContainer" containerID="8a6c1d58ec2175031db5f9f96da1828b53f2c0f7cc5b88e628dab6b902ec69f8" Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.953960 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85d9c479c-zpq5m"] Dec 01 16:06:49 crc kubenswrapper[4810]: E1201 16:06:49.953962 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a6c1d58ec2175031db5f9f96da1828b53f2c0f7cc5b88e628dab6b902ec69f8\": container with ID starting with 8a6c1d58ec2175031db5f9f96da1828b53f2c0f7cc5b88e628dab6b902ec69f8 not found: ID does not exist" containerID="8a6c1d58ec2175031db5f9f96da1828b53f2c0f7cc5b88e628dab6b902ec69f8" Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.954019 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a6c1d58ec2175031db5f9f96da1828b53f2c0f7cc5b88e628dab6b902ec69f8"} err="failed to get container status \"8a6c1d58ec2175031db5f9f96da1828b53f2c0f7cc5b88e628dab6b902ec69f8\": rpc error: code = NotFound desc = could not find container \"8a6c1d58ec2175031db5f9f96da1828b53f2c0f7cc5b88e628dab6b902ec69f8\": container with ID starting with 8a6c1d58ec2175031db5f9f96da1828b53f2c0f7cc5b88e628dab6b902ec69f8 not found: ID does not exist" Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.954045 4810 scope.go:117] "RemoveContainer" containerID="96a2a8f5a0629a212e30106038bd3d9973d0253cd2077dc18f9e59165d979cae" Dec 01 16:06:49 crc kubenswrapper[4810]: E1201 16:06:49.954371 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96a2a8f5a0629a212e30106038bd3d9973d0253cd2077dc18f9e59165d979cae\": container with ID starting with 96a2a8f5a0629a212e30106038bd3d9973d0253cd2077dc18f9e59165d979cae not found: ID does not exist" containerID="96a2a8f5a0629a212e30106038bd3d9973d0253cd2077dc18f9e59165d979cae" Dec 01 16:06:49 crc kubenswrapper[4810]: I1201 16:06:49.954406 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96a2a8f5a0629a212e30106038bd3d9973d0253cd2077dc18f9e59165d979cae"} err="failed to get container status \"96a2a8f5a0629a212e30106038bd3d9973d0253cd2077dc18f9e59165d979cae\": rpc error: code = NotFound desc = could not find container \"96a2a8f5a0629a212e30106038bd3d9973d0253cd2077dc18f9e59165d979cae\": container with ID starting with 96a2a8f5a0629a212e30106038bd3d9973d0253cd2077dc18f9e59165d979cae not found: ID does not exist" Dec 01 16:06:50 crc kubenswrapper[4810]: I1201 16:06:50.501766 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef742dba-5fbf-499f-9e26-1af4eeb10c28" path="/var/lib/kubelet/pods/ef742dba-5fbf-499f-9e26-1af4eeb10c28/volumes" Dec 01 16:06:50 crc kubenswrapper[4810]: I1201 16:06:50.976574 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.563365 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Dec 01 16:06:53 crc kubenswrapper[4810]: E1201 16:06:53.564167 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef742dba-5fbf-499f-9e26-1af4eeb10c28" containerName="dnsmasq-dns" Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.564184 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef742dba-5fbf-499f-9e26-1af4eeb10c28" containerName="dnsmasq-dns" Dec 01 16:06:53 crc kubenswrapper[4810]: E1201 16:06:53.564205 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef742dba-5fbf-499f-9e26-1af4eeb10c28" containerName="init" Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.564211 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef742dba-5fbf-499f-9e26-1af4eeb10c28" containerName="init" Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.564363 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef742dba-5fbf-499f-9e26-1af4eeb10c28" containerName="dnsmasq-dns" Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.564983 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.567760 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.571109 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.698380 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-47574dfa-681b-43d9-a251-554a92055dd4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-47574dfa-681b-43d9-a251-554a92055dd4\") pod \"ovn-copy-data\" (UID: \"00f95a04-d522-4fd8-ab09-ed6dd151d2c3\") " pod="openstack/ovn-copy-data" Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.698502 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vh2m\" (UniqueName: \"kubernetes.io/projected/00f95a04-d522-4fd8-ab09-ed6dd151d2c3-kube-api-access-7vh2m\") pod \"ovn-copy-data\" (UID: \"00f95a04-d522-4fd8-ab09-ed6dd151d2c3\") " pod="openstack/ovn-copy-data" Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.698603 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/00f95a04-d522-4fd8-ab09-ed6dd151d2c3-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"00f95a04-d522-4fd8-ab09-ed6dd151d2c3\") " pod="openstack/ovn-copy-data" Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.800219 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/00f95a04-d522-4fd8-ab09-ed6dd151d2c3-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"00f95a04-d522-4fd8-ab09-ed6dd151d2c3\") " pod="openstack/ovn-copy-data" Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.800297 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-47574dfa-681b-43d9-a251-554a92055dd4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-47574dfa-681b-43d9-a251-554a92055dd4\") pod \"ovn-copy-data\" (UID: \"00f95a04-d522-4fd8-ab09-ed6dd151d2c3\") " pod="openstack/ovn-copy-data" Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.800357 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vh2m\" (UniqueName: \"kubernetes.io/projected/00f95a04-d522-4fd8-ab09-ed6dd151d2c3-kube-api-access-7vh2m\") pod \"ovn-copy-data\" (UID: \"00f95a04-d522-4fd8-ab09-ed6dd151d2c3\") " pod="openstack/ovn-copy-data" Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.803574 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.803623 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-47574dfa-681b-43d9-a251-554a92055dd4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-47574dfa-681b-43d9-a251-554a92055dd4\") pod \"ovn-copy-data\" (UID: \"00f95a04-d522-4fd8-ab09-ed6dd151d2c3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/be1bf660e1355350b2cee142062e8dbb7a30116a232b98803faa1f1e840026fb/globalmount\"" pod="openstack/ovn-copy-data" Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.807338 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/00f95a04-d522-4fd8-ab09-ed6dd151d2c3-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"00f95a04-d522-4fd8-ab09-ed6dd151d2c3\") " pod="openstack/ovn-copy-data" Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.820001 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vh2m\" (UniqueName: \"kubernetes.io/projected/00f95a04-d522-4fd8-ab09-ed6dd151d2c3-kube-api-access-7vh2m\") pod \"ovn-copy-data\" (UID: \"00f95a04-d522-4fd8-ab09-ed6dd151d2c3\") " pod="openstack/ovn-copy-data" Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.839680 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-47574dfa-681b-43d9-a251-554a92055dd4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-47574dfa-681b-43d9-a251-554a92055dd4\") pod \"ovn-copy-data\" (UID: \"00f95a04-d522-4fd8-ab09-ed6dd151d2c3\") " pod="openstack/ovn-copy-data" Dec 01 16:06:53 crc kubenswrapper[4810]: I1201 16:06:53.881189 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 01 16:06:54 crc kubenswrapper[4810]: I1201 16:06:54.390342 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Dec 01 16:06:54 crc kubenswrapper[4810]: W1201 16:06:54.393327 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00f95a04_d522_4fd8_ab09_ed6dd151d2c3.slice/crio-e95a4f9ea8e60dda9356ea03e859f39102c6c65d31447f7b73a5af43b4bc7c32 WatchSource:0}: Error finding container e95a4f9ea8e60dda9356ea03e859f39102c6c65d31447f7b73a5af43b4bc7c32: Status 404 returned error can't find the container with id e95a4f9ea8e60dda9356ea03e859f39102c6c65d31447f7b73a5af43b4bc7c32 Dec 01 16:06:54 crc kubenswrapper[4810]: I1201 16:06:54.943706 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"00f95a04-d522-4fd8-ab09-ed6dd151d2c3","Type":"ContainerStarted","Data":"e95a4f9ea8e60dda9356ea03e859f39102c6c65d31447f7b73a5af43b4bc7c32"} Dec 01 16:06:57 crc kubenswrapper[4810]: I1201 16:06:57.491008 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:06:57 crc kubenswrapper[4810]: E1201 16:06:57.491876 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:06:57 crc kubenswrapper[4810]: I1201 16:06:57.686774 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:06:57 crc kubenswrapper[4810]: I1201 16:06:57.752782 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-964767c4f-mxwfj"] Dec 01 16:06:57 crc kubenswrapper[4810]: I1201 16:06:57.753039 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-964767c4f-mxwfj" podUID="47f4dca5-a8bd-4017-948f-8add5492e78f" containerName="dnsmasq-dns" containerID="cri-o://a91cbfa70cc72acc400e8cb6eb1913ac06cbf54150c2af7da400cd56e29191c6" gracePeriod=10 Dec 01 16:06:57 crc kubenswrapper[4810]: I1201 16:06:57.978651 4810 generic.go:334] "Generic (PLEG): container finished" podID="47f4dca5-a8bd-4017-948f-8add5492e78f" containerID="a91cbfa70cc72acc400e8cb6eb1913ac06cbf54150c2af7da400cd56e29191c6" exitCode=0 Dec 01 16:06:57 crc kubenswrapper[4810]: I1201 16:06:57.978716 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-964767c4f-mxwfj" event={"ID":"47f4dca5-a8bd-4017-948f-8add5492e78f","Type":"ContainerDied","Data":"a91cbfa70cc72acc400e8cb6eb1913ac06cbf54150c2af7da400cd56e29191c6"} Dec 01 16:06:57 crc kubenswrapper[4810]: I1201 16:06:57.980511 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"00f95a04-d522-4fd8-ab09-ed6dd151d2c3","Type":"ContainerStarted","Data":"bf278a2f671ea4e4f6c001b1f357946db1ebd7816507ed7aae899253c8dac572"} Dec 01 16:06:57 crc kubenswrapper[4810]: I1201 16:06:57.998796 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=2.996168621 podStartE2EDuration="5.998775866s" podCreationTimestamp="2025-12-01 16:06:52 +0000 UTC" firstStartedPulling="2025-12-01 16:06:54.395541159 +0000 UTC m=+5580.159050772" lastFinishedPulling="2025-12-01 16:06:57.398148414 +0000 UTC m=+5583.161658017" observedRunningTime="2025-12-01 16:06:57.992805414 +0000 UTC m=+5583.756315017" watchObservedRunningTime="2025-12-01 16:06:57.998775866 +0000 UTC m=+5583.762285469" Dec 01 16:06:58 crc kubenswrapper[4810]: I1201 16:06:58.207857 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-964767c4f-mxwfj" Dec 01 16:06:58 crc kubenswrapper[4810]: I1201 16:06:58.277820 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjpqd\" (UniqueName: \"kubernetes.io/projected/47f4dca5-a8bd-4017-948f-8add5492e78f-kube-api-access-wjpqd\") pod \"47f4dca5-a8bd-4017-948f-8add5492e78f\" (UID: \"47f4dca5-a8bd-4017-948f-8add5492e78f\") " Dec 01 16:06:58 crc kubenswrapper[4810]: I1201 16:06:58.278064 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47f4dca5-a8bd-4017-948f-8add5492e78f-dns-svc\") pod \"47f4dca5-a8bd-4017-948f-8add5492e78f\" (UID: \"47f4dca5-a8bd-4017-948f-8add5492e78f\") " Dec 01 16:06:58 crc kubenswrapper[4810]: I1201 16:06:58.278119 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47f4dca5-a8bd-4017-948f-8add5492e78f-config\") pod \"47f4dca5-a8bd-4017-948f-8add5492e78f\" (UID: \"47f4dca5-a8bd-4017-948f-8add5492e78f\") " Dec 01 16:06:58 crc kubenswrapper[4810]: I1201 16:06:58.283678 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47f4dca5-a8bd-4017-948f-8add5492e78f-kube-api-access-wjpqd" (OuterVolumeSpecName: "kube-api-access-wjpqd") pod "47f4dca5-a8bd-4017-948f-8add5492e78f" (UID: "47f4dca5-a8bd-4017-948f-8add5492e78f"). InnerVolumeSpecName "kube-api-access-wjpqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:06:58 crc kubenswrapper[4810]: I1201 16:06:58.318644 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47f4dca5-a8bd-4017-948f-8add5492e78f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "47f4dca5-a8bd-4017-948f-8add5492e78f" (UID: "47f4dca5-a8bd-4017-948f-8add5492e78f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:06:58 crc kubenswrapper[4810]: I1201 16:06:58.321320 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47f4dca5-a8bd-4017-948f-8add5492e78f-config" (OuterVolumeSpecName: "config") pod "47f4dca5-a8bd-4017-948f-8add5492e78f" (UID: "47f4dca5-a8bd-4017-948f-8add5492e78f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:06:58 crc kubenswrapper[4810]: I1201 16:06:58.380351 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47f4dca5-a8bd-4017-948f-8add5492e78f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 16:06:58 crc kubenswrapper[4810]: I1201 16:06:58.380704 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47f4dca5-a8bd-4017-948f-8add5492e78f-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:06:58 crc kubenswrapper[4810]: I1201 16:06:58.380719 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjpqd\" (UniqueName: \"kubernetes.io/projected/47f4dca5-a8bd-4017-948f-8add5492e78f-kube-api-access-wjpqd\") on node \"crc\" DevicePath \"\"" Dec 01 16:06:58 crc kubenswrapper[4810]: I1201 16:06:58.989816 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-964767c4f-mxwfj" event={"ID":"47f4dca5-a8bd-4017-948f-8add5492e78f","Type":"ContainerDied","Data":"9df1fd9da6a928e0dbe4defb38a9d56a565638f7e47cf0abe2a06563fe82d97e"} Dec 01 16:06:58 crc kubenswrapper[4810]: I1201 16:06:58.989870 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-964767c4f-mxwfj" Dec 01 16:06:58 crc kubenswrapper[4810]: I1201 16:06:58.989888 4810 scope.go:117] "RemoveContainer" containerID="a91cbfa70cc72acc400e8cb6eb1913ac06cbf54150c2af7da400cd56e29191c6" Dec 01 16:06:59 crc kubenswrapper[4810]: I1201 16:06:59.011829 4810 scope.go:117] "RemoveContainer" containerID="121e95f4c3189ba3c64f8dd383a65c3c03a106899a7d7c0e58558a88344f67d9" Dec 01 16:06:59 crc kubenswrapper[4810]: I1201 16:06:59.012519 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-964767c4f-mxwfj"] Dec 01 16:06:59 crc kubenswrapper[4810]: I1201 16:06:59.018747 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-964767c4f-mxwfj"] Dec 01 16:07:00 crc kubenswrapper[4810]: I1201 16:07:00.501248 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47f4dca5-a8bd-4017-948f-8add5492e78f" path="/var/lib/kubelet/pods/47f4dca5-a8bd-4017-948f-8add5492e78f/volumes" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.635287 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 01 16:07:03 crc kubenswrapper[4810]: E1201 16:07:03.635902 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47f4dca5-a8bd-4017-948f-8add5492e78f" containerName="init" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.635915 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="47f4dca5-a8bd-4017-948f-8add5492e78f" containerName="init" Dec 01 16:07:03 crc kubenswrapper[4810]: E1201 16:07:03.635934 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47f4dca5-a8bd-4017-948f-8add5492e78f" containerName="dnsmasq-dns" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.635940 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="47f4dca5-a8bd-4017-948f-8add5492e78f" containerName="dnsmasq-dns" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.636109 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="47f4dca5-a8bd-4017-948f-8add5492e78f" containerName="dnsmasq-dns" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.638937 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.649261 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.649445 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.652843 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-bdnpp" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.653333 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.663896 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.764154 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e8bace8-979e-4aa2-bacd-e9c45c301391-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.764232 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e8bace8-979e-4aa2-bacd-e9c45c301391-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.764518 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e8bace8-979e-4aa2-bacd-e9c45c301391-scripts\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.764568 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5e8bace8-979e-4aa2-bacd-e9c45c301391-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.764606 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl2s2\" (UniqueName: \"kubernetes.io/projected/5e8bace8-979e-4aa2-bacd-e9c45c301391-kube-api-access-hl2s2\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.764758 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e8bace8-979e-4aa2-bacd-e9c45c301391-config\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.764935 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e8bace8-979e-4aa2-bacd-e9c45c301391-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.866775 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e8bace8-979e-4aa2-bacd-e9c45c301391-scripts\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.866831 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5e8bace8-979e-4aa2-bacd-e9c45c301391-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.866858 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl2s2\" (UniqueName: \"kubernetes.io/projected/5e8bace8-979e-4aa2-bacd-e9c45c301391-kube-api-access-hl2s2\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.867401 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5e8bace8-979e-4aa2-bacd-e9c45c301391-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.867531 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e8bace8-979e-4aa2-bacd-e9c45c301391-config\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.867651 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e8bace8-979e-4aa2-bacd-e9c45c301391-scripts\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.868255 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e8bace8-979e-4aa2-bacd-e9c45c301391-config\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.868399 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e8bace8-979e-4aa2-bacd-e9c45c301391-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.869304 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e8bace8-979e-4aa2-bacd-e9c45c301391-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.869375 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e8bace8-979e-4aa2-bacd-e9c45c301391-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.874313 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e8bace8-979e-4aa2-bacd-e9c45c301391-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.874928 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e8bace8-979e-4aa2-bacd-e9c45c301391-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.885227 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl2s2\" (UniqueName: \"kubernetes.io/projected/5e8bace8-979e-4aa2-bacd-e9c45c301391-kube-api-access-hl2s2\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.889579 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e8bace8-979e-4aa2-bacd-e9c45c301391-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"5e8bace8-979e-4aa2-bacd-e9c45c301391\") " pod="openstack/ovn-northd-0" Dec 01 16:07:03 crc kubenswrapper[4810]: I1201 16:07:03.962584 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 01 16:07:04 crc kubenswrapper[4810]: I1201 16:07:04.387219 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 16:07:04 crc kubenswrapper[4810]: W1201 16:07:04.389418 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e8bace8_979e_4aa2_bacd_e9c45c301391.slice/crio-6fe3368d7170ed28f22d4f98d6edd6fc95d85cf3c55afc20e2f8ea9d6169da21 WatchSource:0}: Error finding container 6fe3368d7170ed28f22d4f98d6edd6fc95d85cf3c55afc20e2f8ea9d6169da21: Status 404 returned error can't find the container with id 6fe3368d7170ed28f22d4f98d6edd6fc95d85cf3c55afc20e2f8ea9d6169da21 Dec 01 16:07:05 crc kubenswrapper[4810]: I1201 16:07:05.040749 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5e8bace8-979e-4aa2-bacd-e9c45c301391","Type":"ContainerStarted","Data":"ad1fa7cbbf66aa6a1893e2d8b0cb3c77826431f78f62b94522da9eeccb9e1ddf"} Dec 01 16:07:05 crc kubenswrapper[4810]: I1201 16:07:05.041310 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5e8bace8-979e-4aa2-bacd-e9c45c301391","Type":"ContainerStarted","Data":"6fe3368d7170ed28f22d4f98d6edd6fc95d85cf3c55afc20e2f8ea9d6169da21"} Dec 01 16:07:06 crc kubenswrapper[4810]: I1201 16:07:06.049615 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5e8bace8-979e-4aa2-bacd-e9c45c301391","Type":"ContainerStarted","Data":"caa53117eeecf29edbc3a11cca5ca1d18acf736e489ca8ad055784f5f9ab9eb3"} Dec 01 16:07:06 crc kubenswrapper[4810]: I1201 16:07:06.049765 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 01 16:07:06 crc kubenswrapper[4810]: I1201 16:07:06.076660 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.0766139 podStartE2EDuration="3.0766139s" podCreationTimestamp="2025-12-01 16:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:07:06.074348898 +0000 UTC m=+5591.837858501" watchObservedRunningTime="2025-12-01 16:07:06.0766139 +0000 UTC m=+5591.840123503" Dec 01 16:07:08 crc kubenswrapper[4810]: I1201 16:07:08.491490 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.079742 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"a05157f175de8551ab2001ecba2c36ed1be28837c6a63ea1ad99ac3229a74f07"} Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.304740 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-nhj96"] Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.305834 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nhj96" Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.319625 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-nhj96"] Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.363585 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpqxg\" (UniqueName: \"kubernetes.io/projected/b0bb8241-0cd8-487e-86d0-7c89995f2632-kube-api-access-tpqxg\") pod \"keystone-db-create-nhj96\" (UID: \"b0bb8241-0cd8-487e-86d0-7c89995f2632\") " pod="openstack/keystone-db-create-nhj96" Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.363690 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0bb8241-0cd8-487e-86d0-7c89995f2632-operator-scripts\") pod \"keystone-db-create-nhj96\" (UID: \"b0bb8241-0cd8-487e-86d0-7c89995f2632\") " pod="openstack/keystone-db-create-nhj96" Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.407132 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5de1-account-create-update-79k2v"] Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.408184 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5de1-account-create-update-79k2v" Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.410268 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.426964 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5de1-account-create-update-79k2v"] Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.465113 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88b6fee6-788f-48c3-9d06-c5eaec035c49-operator-scripts\") pod \"keystone-5de1-account-create-update-79k2v\" (UID: \"88b6fee6-788f-48c3-9d06-c5eaec035c49\") " pod="openstack/keystone-5de1-account-create-update-79k2v" Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.465163 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpqxg\" (UniqueName: \"kubernetes.io/projected/b0bb8241-0cd8-487e-86d0-7c89995f2632-kube-api-access-tpqxg\") pod \"keystone-db-create-nhj96\" (UID: \"b0bb8241-0cd8-487e-86d0-7c89995f2632\") " pod="openstack/keystone-db-create-nhj96" Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.465200 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0bb8241-0cd8-487e-86d0-7c89995f2632-operator-scripts\") pod \"keystone-db-create-nhj96\" (UID: \"b0bb8241-0cd8-487e-86d0-7c89995f2632\") " pod="openstack/keystone-db-create-nhj96" Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.465261 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8qcp\" (UniqueName: \"kubernetes.io/projected/88b6fee6-788f-48c3-9d06-c5eaec035c49-kube-api-access-w8qcp\") pod \"keystone-5de1-account-create-update-79k2v\" (UID: \"88b6fee6-788f-48c3-9d06-c5eaec035c49\") " pod="openstack/keystone-5de1-account-create-update-79k2v" Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.465895 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0bb8241-0cd8-487e-86d0-7c89995f2632-operator-scripts\") pod \"keystone-db-create-nhj96\" (UID: \"b0bb8241-0cd8-487e-86d0-7c89995f2632\") " pod="openstack/keystone-db-create-nhj96" Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.491645 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpqxg\" (UniqueName: \"kubernetes.io/projected/b0bb8241-0cd8-487e-86d0-7c89995f2632-kube-api-access-tpqxg\") pod \"keystone-db-create-nhj96\" (UID: \"b0bb8241-0cd8-487e-86d0-7c89995f2632\") " pod="openstack/keystone-db-create-nhj96" Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.566782 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88b6fee6-788f-48c3-9d06-c5eaec035c49-operator-scripts\") pod \"keystone-5de1-account-create-update-79k2v\" (UID: \"88b6fee6-788f-48c3-9d06-c5eaec035c49\") " pod="openstack/keystone-5de1-account-create-update-79k2v" Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.566860 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8qcp\" (UniqueName: \"kubernetes.io/projected/88b6fee6-788f-48c3-9d06-c5eaec035c49-kube-api-access-w8qcp\") pod \"keystone-5de1-account-create-update-79k2v\" (UID: \"88b6fee6-788f-48c3-9d06-c5eaec035c49\") " pod="openstack/keystone-5de1-account-create-update-79k2v" Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.567696 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88b6fee6-788f-48c3-9d06-c5eaec035c49-operator-scripts\") pod \"keystone-5de1-account-create-update-79k2v\" (UID: \"88b6fee6-788f-48c3-9d06-c5eaec035c49\") " pod="openstack/keystone-5de1-account-create-update-79k2v" Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.585644 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8qcp\" (UniqueName: \"kubernetes.io/projected/88b6fee6-788f-48c3-9d06-c5eaec035c49-kube-api-access-w8qcp\") pod \"keystone-5de1-account-create-update-79k2v\" (UID: \"88b6fee6-788f-48c3-9d06-c5eaec035c49\") " pod="openstack/keystone-5de1-account-create-update-79k2v" Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.631302 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nhj96" Dec 01 16:07:09 crc kubenswrapper[4810]: I1201 16:07:09.722637 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5de1-account-create-update-79k2v" Dec 01 16:07:10 crc kubenswrapper[4810]: I1201 16:07:10.258034 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-nhj96"] Dec 01 16:07:10 crc kubenswrapper[4810]: I1201 16:07:10.323609 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5de1-account-create-update-79k2v"] Dec 01 16:07:10 crc kubenswrapper[4810]: W1201 16:07:10.326617 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88b6fee6_788f_48c3_9d06_c5eaec035c49.slice/crio-603d20e316da6ef817d4c034507a7dc7717296d1828a6295e5cd2505c1be3ca7 WatchSource:0}: Error finding container 603d20e316da6ef817d4c034507a7dc7717296d1828a6295e5cd2505c1be3ca7: Status 404 returned error can't find the container with id 603d20e316da6ef817d4c034507a7dc7717296d1828a6295e5cd2505c1be3ca7 Dec 01 16:07:11 crc kubenswrapper[4810]: I1201 16:07:11.094018 4810 generic.go:334] "Generic (PLEG): container finished" podID="b0bb8241-0cd8-487e-86d0-7c89995f2632" containerID="d31f441586726257a95607a056fe938965f790e148a7288e8d22984c6eaac50e" exitCode=0 Dec 01 16:07:11 crc kubenswrapper[4810]: I1201 16:07:11.094160 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nhj96" event={"ID":"b0bb8241-0cd8-487e-86d0-7c89995f2632","Type":"ContainerDied","Data":"d31f441586726257a95607a056fe938965f790e148a7288e8d22984c6eaac50e"} Dec 01 16:07:11 crc kubenswrapper[4810]: I1201 16:07:11.094953 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nhj96" event={"ID":"b0bb8241-0cd8-487e-86d0-7c89995f2632","Type":"ContainerStarted","Data":"24c6b506b96b7ea41b6751d04034527a9ac0ce49c3f62ec1b969d0cf45604d3a"} Dec 01 16:07:11 crc kubenswrapper[4810]: I1201 16:07:11.096349 4810 generic.go:334] "Generic (PLEG): container finished" podID="88b6fee6-788f-48c3-9d06-c5eaec035c49" containerID="990e5b1039da24fd77b75aa90bdccc2baa567f6f1e4927c0333286693b7e9a03" exitCode=0 Dec 01 16:07:11 crc kubenswrapper[4810]: I1201 16:07:11.096382 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5de1-account-create-update-79k2v" event={"ID":"88b6fee6-788f-48c3-9d06-c5eaec035c49","Type":"ContainerDied","Data":"990e5b1039da24fd77b75aa90bdccc2baa567f6f1e4927c0333286693b7e9a03"} Dec 01 16:07:11 crc kubenswrapper[4810]: I1201 16:07:11.096397 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5de1-account-create-update-79k2v" event={"ID":"88b6fee6-788f-48c3-9d06-c5eaec035c49","Type":"ContainerStarted","Data":"603d20e316da6ef817d4c034507a7dc7717296d1828a6295e5cd2505c1be3ca7"} Dec 01 16:07:12 crc kubenswrapper[4810]: I1201 16:07:12.511173 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5de1-account-create-update-79k2v" Dec 01 16:07:12 crc kubenswrapper[4810]: I1201 16:07:12.519172 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nhj96" Dec 01 16:07:12 crc kubenswrapper[4810]: I1201 16:07:12.620564 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8qcp\" (UniqueName: \"kubernetes.io/projected/88b6fee6-788f-48c3-9d06-c5eaec035c49-kube-api-access-w8qcp\") pod \"88b6fee6-788f-48c3-9d06-c5eaec035c49\" (UID: \"88b6fee6-788f-48c3-9d06-c5eaec035c49\") " Dec 01 16:07:12 crc kubenswrapper[4810]: I1201 16:07:12.620627 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0bb8241-0cd8-487e-86d0-7c89995f2632-operator-scripts\") pod \"b0bb8241-0cd8-487e-86d0-7c89995f2632\" (UID: \"b0bb8241-0cd8-487e-86d0-7c89995f2632\") " Dec 01 16:07:12 crc kubenswrapper[4810]: I1201 16:07:12.620691 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpqxg\" (UniqueName: \"kubernetes.io/projected/b0bb8241-0cd8-487e-86d0-7c89995f2632-kube-api-access-tpqxg\") pod \"b0bb8241-0cd8-487e-86d0-7c89995f2632\" (UID: \"b0bb8241-0cd8-487e-86d0-7c89995f2632\") " Dec 01 16:07:12 crc kubenswrapper[4810]: I1201 16:07:12.620814 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88b6fee6-788f-48c3-9d06-c5eaec035c49-operator-scripts\") pod \"88b6fee6-788f-48c3-9d06-c5eaec035c49\" (UID: \"88b6fee6-788f-48c3-9d06-c5eaec035c49\") " Dec 01 16:07:12 crc kubenswrapper[4810]: I1201 16:07:12.621499 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0bb8241-0cd8-487e-86d0-7c89995f2632-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b0bb8241-0cd8-487e-86d0-7c89995f2632" (UID: "b0bb8241-0cd8-487e-86d0-7c89995f2632"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:07:12 crc kubenswrapper[4810]: I1201 16:07:12.621754 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88b6fee6-788f-48c3-9d06-c5eaec035c49-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "88b6fee6-788f-48c3-9d06-c5eaec035c49" (UID: "88b6fee6-788f-48c3-9d06-c5eaec035c49"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:07:12 crc kubenswrapper[4810]: I1201 16:07:12.625946 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0bb8241-0cd8-487e-86d0-7c89995f2632-kube-api-access-tpqxg" (OuterVolumeSpecName: "kube-api-access-tpqxg") pod "b0bb8241-0cd8-487e-86d0-7c89995f2632" (UID: "b0bb8241-0cd8-487e-86d0-7c89995f2632"). InnerVolumeSpecName "kube-api-access-tpqxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:07:12 crc kubenswrapper[4810]: I1201 16:07:12.626067 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88b6fee6-788f-48c3-9d06-c5eaec035c49-kube-api-access-w8qcp" (OuterVolumeSpecName: "kube-api-access-w8qcp") pod "88b6fee6-788f-48c3-9d06-c5eaec035c49" (UID: "88b6fee6-788f-48c3-9d06-c5eaec035c49"). InnerVolumeSpecName "kube-api-access-w8qcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:07:12 crc kubenswrapper[4810]: I1201 16:07:12.723113 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8qcp\" (UniqueName: \"kubernetes.io/projected/88b6fee6-788f-48c3-9d06-c5eaec035c49-kube-api-access-w8qcp\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:12 crc kubenswrapper[4810]: I1201 16:07:12.723144 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0bb8241-0cd8-487e-86d0-7c89995f2632-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:12 crc kubenswrapper[4810]: I1201 16:07:12.723154 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpqxg\" (UniqueName: \"kubernetes.io/projected/b0bb8241-0cd8-487e-86d0-7c89995f2632-kube-api-access-tpqxg\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:12 crc kubenswrapper[4810]: I1201 16:07:12.723162 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88b6fee6-788f-48c3-9d06-c5eaec035c49-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:13 crc kubenswrapper[4810]: I1201 16:07:13.113288 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-nhj96" event={"ID":"b0bb8241-0cd8-487e-86d0-7c89995f2632","Type":"ContainerDied","Data":"24c6b506b96b7ea41b6751d04034527a9ac0ce49c3f62ec1b969d0cf45604d3a"} Dec 01 16:07:13 crc kubenswrapper[4810]: I1201 16:07:13.113336 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24c6b506b96b7ea41b6751d04034527a9ac0ce49c3f62ec1b969d0cf45604d3a" Dec 01 16:07:13 crc kubenswrapper[4810]: I1201 16:07:13.113311 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-nhj96" Dec 01 16:07:13 crc kubenswrapper[4810]: I1201 16:07:13.115515 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5de1-account-create-update-79k2v" event={"ID":"88b6fee6-788f-48c3-9d06-c5eaec035c49","Type":"ContainerDied","Data":"603d20e316da6ef817d4c034507a7dc7717296d1828a6295e5cd2505c1be3ca7"} Dec 01 16:07:13 crc kubenswrapper[4810]: I1201 16:07:13.115554 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5de1-account-create-update-79k2v" Dec 01 16:07:13 crc kubenswrapper[4810]: I1201 16:07:13.115565 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="603d20e316da6ef817d4c034507a7dc7717296d1828a6295e5cd2505c1be3ca7" Dec 01 16:07:14 crc kubenswrapper[4810]: I1201 16:07:14.840880 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-vt57p"] Dec 01 16:07:14 crc kubenswrapper[4810]: E1201 16:07:14.841564 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0bb8241-0cd8-487e-86d0-7c89995f2632" containerName="mariadb-database-create" Dec 01 16:07:14 crc kubenswrapper[4810]: I1201 16:07:14.841577 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0bb8241-0cd8-487e-86d0-7c89995f2632" containerName="mariadb-database-create" Dec 01 16:07:14 crc kubenswrapper[4810]: E1201 16:07:14.841586 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88b6fee6-788f-48c3-9d06-c5eaec035c49" containerName="mariadb-account-create-update" Dec 01 16:07:14 crc kubenswrapper[4810]: I1201 16:07:14.841591 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="88b6fee6-788f-48c3-9d06-c5eaec035c49" containerName="mariadb-account-create-update" Dec 01 16:07:14 crc kubenswrapper[4810]: I1201 16:07:14.841754 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0bb8241-0cd8-487e-86d0-7c89995f2632" containerName="mariadb-database-create" Dec 01 16:07:14 crc kubenswrapper[4810]: I1201 16:07:14.841770 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="88b6fee6-788f-48c3-9d06-c5eaec035c49" containerName="mariadb-account-create-update" Dec 01 16:07:14 crc kubenswrapper[4810]: I1201 16:07:14.842342 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vt57p" Dec 01 16:07:14 crc kubenswrapper[4810]: I1201 16:07:14.845674 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 16:07:14 crc kubenswrapper[4810]: I1201 16:07:14.845770 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vl9m4" Dec 01 16:07:14 crc kubenswrapper[4810]: I1201 16:07:14.845670 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 16:07:14 crc kubenswrapper[4810]: I1201 16:07:14.847020 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 16:07:14 crc kubenswrapper[4810]: I1201 16:07:14.853933 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-vt57p"] Dec 01 16:07:14 crc kubenswrapper[4810]: I1201 16:07:14.963403 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cef9541-a570-4638-9360-8ecc123c441a-combined-ca-bundle\") pod \"keystone-db-sync-vt57p\" (UID: \"4cef9541-a570-4638-9360-8ecc123c441a\") " pod="openstack/keystone-db-sync-vt57p" Dec 01 16:07:14 crc kubenswrapper[4810]: I1201 16:07:14.963737 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q4fl\" (UniqueName: \"kubernetes.io/projected/4cef9541-a570-4638-9360-8ecc123c441a-kube-api-access-7q4fl\") pod \"keystone-db-sync-vt57p\" (UID: \"4cef9541-a570-4638-9360-8ecc123c441a\") " pod="openstack/keystone-db-sync-vt57p" Dec 01 16:07:14 crc kubenswrapper[4810]: I1201 16:07:14.963859 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cef9541-a570-4638-9360-8ecc123c441a-config-data\") pod \"keystone-db-sync-vt57p\" (UID: \"4cef9541-a570-4638-9360-8ecc123c441a\") " pod="openstack/keystone-db-sync-vt57p" Dec 01 16:07:15 crc kubenswrapper[4810]: I1201 16:07:15.065394 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q4fl\" (UniqueName: \"kubernetes.io/projected/4cef9541-a570-4638-9360-8ecc123c441a-kube-api-access-7q4fl\") pod \"keystone-db-sync-vt57p\" (UID: \"4cef9541-a570-4638-9360-8ecc123c441a\") " pod="openstack/keystone-db-sync-vt57p" Dec 01 16:07:15 crc kubenswrapper[4810]: I1201 16:07:15.065492 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cef9541-a570-4638-9360-8ecc123c441a-config-data\") pod \"keystone-db-sync-vt57p\" (UID: \"4cef9541-a570-4638-9360-8ecc123c441a\") " pod="openstack/keystone-db-sync-vt57p" Dec 01 16:07:15 crc kubenswrapper[4810]: I1201 16:07:15.065538 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cef9541-a570-4638-9360-8ecc123c441a-combined-ca-bundle\") pod \"keystone-db-sync-vt57p\" (UID: \"4cef9541-a570-4638-9360-8ecc123c441a\") " pod="openstack/keystone-db-sync-vt57p" Dec 01 16:07:15 crc kubenswrapper[4810]: I1201 16:07:15.071032 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cef9541-a570-4638-9360-8ecc123c441a-combined-ca-bundle\") pod \"keystone-db-sync-vt57p\" (UID: \"4cef9541-a570-4638-9360-8ecc123c441a\") " pod="openstack/keystone-db-sync-vt57p" Dec 01 16:07:15 crc kubenswrapper[4810]: I1201 16:07:15.071238 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cef9541-a570-4638-9360-8ecc123c441a-config-data\") pod \"keystone-db-sync-vt57p\" (UID: \"4cef9541-a570-4638-9360-8ecc123c441a\") " pod="openstack/keystone-db-sync-vt57p" Dec 01 16:07:15 crc kubenswrapper[4810]: I1201 16:07:15.083944 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q4fl\" (UniqueName: \"kubernetes.io/projected/4cef9541-a570-4638-9360-8ecc123c441a-kube-api-access-7q4fl\") pod \"keystone-db-sync-vt57p\" (UID: \"4cef9541-a570-4638-9360-8ecc123c441a\") " pod="openstack/keystone-db-sync-vt57p" Dec 01 16:07:15 crc kubenswrapper[4810]: I1201 16:07:15.158934 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vt57p" Dec 01 16:07:15 crc kubenswrapper[4810]: I1201 16:07:15.565702 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-vt57p"] Dec 01 16:07:16 crc kubenswrapper[4810]: I1201 16:07:16.135632 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vt57p" event={"ID":"4cef9541-a570-4638-9360-8ecc123c441a","Type":"ContainerStarted","Data":"c1119418971e85144659b81c9e22e1e466e311c7122e4a09f4cbbaf559b73f61"} Dec 01 16:07:16 crc kubenswrapper[4810]: I1201 16:07:16.135906 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vt57p" event={"ID":"4cef9541-a570-4638-9360-8ecc123c441a","Type":"ContainerStarted","Data":"616acea1bef9a0fd0c34c2e49c96f3ca1b23e9b0b80df38acc20da62f557923d"} Dec 01 16:07:16 crc kubenswrapper[4810]: I1201 16:07:16.154192 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-vt57p" podStartSLOduration=2.154170815 podStartE2EDuration="2.154170815s" podCreationTimestamp="2025-12-01 16:07:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:07:16.148874231 +0000 UTC m=+5601.912383834" watchObservedRunningTime="2025-12-01 16:07:16.154170815 +0000 UTC m=+5601.917680418" Dec 01 16:07:18 crc kubenswrapper[4810]: I1201 16:07:18.152567 4810 generic.go:334] "Generic (PLEG): container finished" podID="4cef9541-a570-4638-9360-8ecc123c441a" containerID="c1119418971e85144659b81c9e22e1e466e311c7122e4a09f4cbbaf559b73f61" exitCode=0 Dec 01 16:07:18 crc kubenswrapper[4810]: I1201 16:07:18.152672 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vt57p" event={"ID":"4cef9541-a570-4638-9360-8ecc123c441a","Type":"ContainerDied","Data":"c1119418971e85144659b81c9e22e1e466e311c7122e4a09f4cbbaf559b73f61"} Dec 01 16:07:19 crc kubenswrapper[4810]: I1201 16:07:19.021814 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 01 16:07:19 crc kubenswrapper[4810]: I1201 16:07:19.484879 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vt57p" Dec 01 16:07:19 crc kubenswrapper[4810]: I1201 16:07:19.553178 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cef9541-a570-4638-9360-8ecc123c441a-combined-ca-bundle\") pod \"4cef9541-a570-4638-9360-8ecc123c441a\" (UID: \"4cef9541-a570-4638-9360-8ecc123c441a\") " Dec 01 16:07:19 crc kubenswrapper[4810]: I1201 16:07:19.553225 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q4fl\" (UniqueName: \"kubernetes.io/projected/4cef9541-a570-4638-9360-8ecc123c441a-kube-api-access-7q4fl\") pod \"4cef9541-a570-4638-9360-8ecc123c441a\" (UID: \"4cef9541-a570-4638-9360-8ecc123c441a\") " Dec 01 16:07:19 crc kubenswrapper[4810]: I1201 16:07:19.553259 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cef9541-a570-4638-9360-8ecc123c441a-config-data\") pod \"4cef9541-a570-4638-9360-8ecc123c441a\" (UID: \"4cef9541-a570-4638-9360-8ecc123c441a\") " Dec 01 16:07:19 crc kubenswrapper[4810]: I1201 16:07:19.559603 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cef9541-a570-4638-9360-8ecc123c441a-kube-api-access-7q4fl" (OuterVolumeSpecName: "kube-api-access-7q4fl") pod "4cef9541-a570-4638-9360-8ecc123c441a" (UID: "4cef9541-a570-4638-9360-8ecc123c441a"). InnerVolumeSpecName "kube-api-access-7q4fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:07:19 crc kubenswrapper[4810]: I1201 16:07:19.580169 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cef9541-a570-4638-9360-8ecc123c441a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4cef9541-a570-4638-9360-8ecc123c441a" (UID: "4cef9541-a570-4638-9360-8ecc123c441a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:07:19 crc kubenswrapper[4810]: I1201 16:07:19.591316 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cef9541-a570-4638-9360-8ecc123c441a-config-data" (OuterVolumeSpecName: "config-data") pod "4cef9541-a570-4638-9360-8ecc123c441a" (UID: "4cef9541-a570-4638-9360-8ecc123c441a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:07:19 crc kubenswrapper[4810]: I1201 16:07:19.654766 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cef9541-a570-4638-9360-8ecc123c441a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:19 crc kubenswrapper[4810]: I1201 16:07:19.654797 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q4fl\" (UniqueName: \"kubernetes.io/projected/4cef9541-a570-4638-9360-8ecc123c441a-kube-api-access-7q4fl\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:19 crc kubenswrapper[4810]: I1201 16:07:19.654808 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cef9541-a570-4638-9360-8ecc123c441a-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.173428 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vt57p" event={"ID":"4cef9541-a570-4638-9360-8ecc123c441a","Type":"ContainerDied","Data":"616acea1bef9a0fd0c34c2e49c96f3ca1b23e9b0b80df38acc20da62f557923d"} Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.173797 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="616acea1bef9a0fd0c34c2e49c96f3ca1b23e9b0b80df38acc20da62f557923d" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.173487 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vt57p" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.398993 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64578c6499-mvcxj"] Dec 01 16:07:20 crc kubenswrapper[4810]: E1201 16:07:20.399318 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cef9541-a570-4638-9360-8ecc123c441a" containerName="keystone-db-sync" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.399334 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cef9541-a570-4638-9360-8ecc123c441a" containerName="keystone-db-sync" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.399515 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cef9541-a570-4638-9360-8ecc123c441a" containerName="keystone-db-sync" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.400289 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.436453 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64578c6499-mvcxj"] Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.467294 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-dns-svc\") pod \"dnsmasq-dns-64578c6499-mvcxj\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.467392 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-ovsdbserver-sb\") pod \"dnsmasq-dns-64578c6499-mvcxj\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.467419 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-config\") pod \"dnsmasq-dns-64578c6499-mvcxj\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.467437 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-ovsdbserver-nb\") pod \"dnsmasq-dns-64578c6499-mvcxj\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.467521 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r69qd\" (UniqueName: \"kubernetes.io/projected/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-kube-api-access-r69qd\") pod \"dnsmasq-dns-64578c6499-mvcxj\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.535426 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-rv49g"] Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.537869 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.540614 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.540836 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.540933 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vl9m4" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.541090 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.545521 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rv49g"] Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.546283 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.570915 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-ovsdbserver-sb\") pod \"dnsmasq-dns-64578c6499-mvcxj\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.569840 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-ovsdbserver-sb\") pod \"dnsmasq-dns-64578c6499-mvcxj\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.571308 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-config\") pod \"dnsmasq-dns-64578c6499-mvcxj\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.571946 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-ovsdbserver-nb\") pod \"dnsmasq-dns-64578c6499-mvcxj\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.571895 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-config\") pod \"dnsmasq-dns-64578c6499-mvcxj\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.572826 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-ovsdbserver-nb\") pod \"dnsmasq-dns-64578c6499-mvcxj\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.573618 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r69qd\" (UniqueName: \"kubernetes.io/projected/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-kube-api-access-r69qd\") pod \"dnsmasq-dns-64578c6499-mvcxj\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.575682 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-dns-svc\") pod \"dnsmasq-dns-64578c6499-mvcxj\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.576510 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-dns-svc\") pod \"dnsmasq-dns-64578c6499-mvcxj\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.592197 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r69qd\" (UniqueName: \"kubernetes.io/projected/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-kube-api-access-r69qd\") pod \"dnsmasq-dns-64578c6499-mvcxj\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.678114 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-config-data\") pod \"keystone-bootstrap-rv49g\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.678190 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5g2t\" (UniqueName: \"kubernetes.io/projected/3114e619-39fd-4bd4-88ae-444eb5c26cad-kube-api-access-c5g2t\") pod \"keystone-bootstrap-rv49g\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.678228 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-combined-ca-bundle\") pod \"keystone-bootstrap-rv49g\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.678251 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-scripts\") pod \"keystone-bootstrap-rv49g\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.678449 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-fernet-keys\") pod \"keystone-bootstrap-rv49g\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.678592 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-credential-keys\") pod \"keystone-bootstrap-rv49g\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.719811 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.779747 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5g2t\" (UniqueName: \"kubernetes.io/projected/3114e619-39fd-4bd4-88ae-444eb5c26cad-kube-api-access-c5g2t\") pod \"keystone-bootstrap-rv49g\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.780084 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-combined-ca-bundle\") pod \"keystone-bootstrap-rv49g\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.780121 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-scripts\") pod \"keystone-bootstrap-rv49g\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.780186 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-fernet-keys\") pod \"keystone-bootstrap-rv49g\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.780222 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-credential-keys\") pod \"keystone-bootstrap-rv49g\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.780312 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-config-data\") pod \"keystone-bootstrap-rv49g\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.784341 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-combined-ca-bundle\") pod \"keystone-bootstrap-rv49g\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.784523 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-config-data\") pod \"keystone-bootstrap-rv49g\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.784593 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-credential-keys\") pod \"keystone-bootstrap-rv49g\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.792599 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-scripts\") pod \"keystone-bootstrap-rv49g\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.798344 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-fernet-keys\") pod \"keystone-bootstrap-rv49g\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.806397 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5g2t\" (UniqueName: \"kubernetes.io/projected/3114e619-39fd-4bd4-88ae-444eb5c26cad-kube-api-access-c5g2t\") pod \"keystone-bootstrap-rv49g\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:20 crc kubenswrapper[4810]: I1201 16:07:20.853212 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:21 crc kubenswrapper[4810]: I1201 16:07:21.168774 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64578c6499-mvcxj"] Dec 01 16:07:21 crc kubenswrapper[4810]: I1201 16:07:21.182383 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64578c6499-mvcxj" event={"ID":"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e","Type":"ContainerStarted","Data":"6976574d391688ab9b2088891b29e0513bbb2ec2f539b4b7a0c65535b685c87f"} Dec 01 16:07:21 crc kubenswrapper[4810]: I1201 16:07:21.339093 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rv49g"] Dec 01 16:07:22 crc kubenswrapper[4810]: I1201 16:07:22.191496 4810 generic.go:334] "Generic (PLEG): container finished" podID="a2b19a52-ce05-49a9-bebd-b3e8a3c9363e" containerID="3c31fb968a8d869c9a460270298ca303afadc6ce67273d2992398035c3848aa8" exitCode=0 Dec 01 16:07:22 crc kubenswrapper[4810]: I1201 16:07:22.191821 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64578c6499-mvcxj" event={"ID":"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e","Type":"ContainerDied","Data":"3c31fb968a8d869c9a460270298ca303afadc6ce67273d2992398035c3848aa8"} Dec 01 16:07:22 crc kubenswrapper[4810]: I1201 16:07:22.197023 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rv49g" event={"ID":"3114e619-39fd-4bd4-88ae-444eb5c26cad","Type":"ContainerStarted","Data":"03b4a9f23ef2c19d78d55ba428b184c75e1f07608755280953ab0aa6589596a5"} Dec 01 16:07:22 crc kubenswrapper[4810]: I1201 16:07:22.197066 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rv49g" event={"ID":"3114e619-39fd-4bd4-88ae-444eb5c26cad","Type":"ContainerStarted","Data":"1c3105ee25bf51b9befdb9ee4c7cdd800988c78ce3a6db6a50631dc67f78c936"} Dec 01 16:07:22 crc kubenswrapper[4810]: I1201 16:07:22.242724 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-rv49g" podStartSLOduration=2.24270259 podStartE2EDuration="2.24270259s" podCreationTimestamp="2025-12-01 16:07:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:07:22.234963089 +0000 UTC m=+5607.998472722" watchObservedRunningTime="2025-12-01 16:07:22.24270259 +0000 UTC m=+5608.006212193" Dec 01 16:07:23 crc kubenswrapper[4810]: I1201 16:07:23.206956 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64578c6499-mvcxj" event={"ID":"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e","Type":"ContainerStarted","Data":"e888a8bce60f9a4907d307433c7ad9b53378dd0ff8bb280328d9ef5bc7d41566"} Dec 01 16:07:23 crc kubenswrapper[4810]: I1201 16:07:23.223611 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64578c6499-mvcxj" podStartSLOduration=3.223591375 podStartE2EDuration="3.223591375s" podCreationTimestamp="2025-12-01 16:07:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:07:23.221875839 +0000 UTC m=+5608.985385442" watchObservedRunningTime="2025-12-01 16:07:23.223591375 +0000 UTC m=+5608.987100978" Dec 01 16:07:24 crc kubenswrapper[4810]: I1201 16:07:24.214251 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:25 crc kubenswrapper[4810]: I1201 16:07:25.222705 4810 generic.go:334] "Generic (PLEG): container finished" podID="3114e619-39fd-4bd4-88ae-444eb5c26cad" containerID="03b4a9f23ef2c19d78d55ba428b184c75e1f07608755280953ab0aa6589596a5" exitCode=0 Dec 01 16:07:25 crc kubenswrapper[4810]: I1201 16:07:25.222801 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rv49g" event={"ID":"3114e619-39fd-4bd4-88ae-444eb5c26cad","Type":"ContainerDied","Data":"03b4a9f23ef2c19d78d55ba428b184c75e1f07608755280953ab0aa6589596a5"} Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.618023 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.700059 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-config-data\") pod \"3114e619-39fd-4bd4-88ae-444eb5c26cad\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.700151 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5g2t\" (UniqueName: \"kubernetes.io/projected/3114e619-39fd-4bd4-88ae-444eb5c26cad-kube-api-access-c5g2t\") pod \"3114e619-39fd-4bd4-88ae-444eb5c26cad\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.700248 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-fernet-keys\") pod \"3114e619-39fd-4bd4-88ae-444eb5c26cad\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.700329 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-credential-keys\") pod \"3114e619-39fd-4bd4-88ae-444eb5c26cad\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.701227 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-scripts\") pod \"3114e619-39fd-4bd4-88ae-444eb5c26cad\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.701303 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-combined-ca-bundle\") pod \"3114e619-39fd-4bd4-88ae-444eb5c26cad\" (UID: \"3114e619-39fd-4bd4-88ae-444eb5c26cad\") " Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.706505 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-scripts" (OuterVolumeSpecName: "scripts") pod "3114e619-39fd-4bd4-88ae-444eb5c26cad" (UID: "3114e619-39fd-4bd4-88ae-444eb5c26cad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.710734 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3114e619-39fd-4bd4-88ae-444eb5c26cad-kube-api-access-c5g2t" (OuterVolumeSpecName: "kube-api-access-c5g2t") pod "3114e619-39fd-4bd4-88ae-444eb5c26cad" (UID: "3114e619-39fd-4bd4-88ae-444eb5c26cad"). InnerVolumeSpecName "kube-api-access-c5g2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.711001 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "3114e619-39fd-4bd4-88ae-444eb5c26cad" (UID: "3114e619-39fd-4bd4-88ae-444eb5c26cad"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.711140 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "3114e619-39fd-4bd4-88ae-444eb5c26cad" (UID: "3114e619-39fd-4bd4-88ae-444eb5c26cad"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.726674 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-config-data" (OuterVolumeSpecName: "config-data") pod "3114e619-39fd-4bd4-88ae-444eb5c26cad" (UID: "3114e619-39fd-4bd4-88ae-444eb5c26cad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.733565 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3114e619-39fd-4bd4-88ae-444eb5c26cad" (UID: "3114e619-39fd-4bd4-88ae-444eb5c26cad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.802941 4810 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.803008 4810 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.803021 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.803029 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.803038 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3114e619-39fd-4bd4-88ae-444eb5c26cad-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:26 crc kubenswrapper[4810]: I1201 16:07:26.803047 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5g2t\" (UniqueName: \"kubernetes.io/projected/3114e619-39fd-4bd4-88ae-444eb5c26cad-kube-api-access-c5g2t\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.239242 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rv49g" event={"ID":"3114e619-39fd-4bd4-88ae-444eb5c26cad","Type":"ContainerDied","Data":"1c3105ee25bf51b9befdb9ee4c7cdd800988c78ce3a6db6a50631dc67f78c936"} Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.239279 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c3105ee25bf51b9befdb9ee4c7cdd800988c78ce3a6db6a50631dc67f78c936" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.239298 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rv49g" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.403587 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-rv49g"] Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.410104 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-rv49g"] Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.503199 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-nbhcs"] Dec 01 16:07:27 crc kubenswrapper[4810]: E1201 16:07:27.503672 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3114e619-39fd-4bd4-88ae-444eb5c26cad" containerName="keystone-bootstrap" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.503702 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3114e619-39fd-4bd4-88ae-444eb5c26cad" containerName="keystone-bootstrap" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.503910 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3114e619-39fd-4bd4-88ae-444eb5c26cad" containerName="keystone-bootstrap" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.504692 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.506590 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.506697 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.506798 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.506846 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vl9m4" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.507083 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.518276 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nbhcs"] Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.618611 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-config-data\") pod \"keystone-bootstrap-nbhcs\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.618710 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-scripts\") pod \"keystone-bootstrap-nbhcs\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.618730 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-combined-ca-bundle\") pod \"keystone-bootstrap-nbhcs\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.618787 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wr58\" (UniqueName: \"kubernetes.io/projected/230f81a0-af9d-4d25-8319-4c9990813ffa-kube-api-access-2wr58\") pod \"keystone-bootstrap-nbhcs\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.618806 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-credential-keys\") pod \"keystone-bootstrap-nbhcs\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.618845 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-fernet-keys\") pod \"keystone-bootstrap-nbhcs\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.719952 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-scripts\") pod \"keystone-bootstrap-nbhcs\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.720315 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-combined-ca-bundle\") pod \"keystone-bootstrap-nbhcs\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.720463 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wr58\" (UniqueName: \"kubernetes.io/projected/230f81a0-af9d-4d25-8319-4c9990813ffa-kube-api-access-2wr58\") pod \"keystone-bootstrap-nbhcs\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.720586 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-credential-keys\") pod \"keystone-bootstrap-nbhcs\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.720692 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-fernet-keys\") pod \"keystone-bootstrap-nbhcs\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.720798 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-config-data\") pod \"keystone-bootstrap-nbhcs\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.724610 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-credential-keys\") pod \"keystone-bootstrap-nbhcs\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.724883 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-scripts\") pod \"keystone-bootstrap-nbhcs\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.727102 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-fernet-keys\") pod \"keystone-bootstrap-nbhcs\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.727193 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-config-data\") pod \"keystone-bootstrap-nbhcs\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.727349 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-combined-ca-bundle\") pod \"keystone-bootstrap-nbhcs\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.740291 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wr58\" (UniqueName: \"kubernetes.io/projected/230f81a0-af9d-4d25-8319-4c9990813ffa-kube-api-access-2wr58\") pod \"keystone-bootstrap-nbhcs\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:27 crc kubenswrapper[4810]: I1201 16:07:27.833408 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:28 crc kubenswrapper[4810]: I1201 16:07:28.318043 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-nbhcs"] Dec 01 16:07:28 crc kubenswrapper[4810]: I1201 16:07:28.502447 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3114e619-39fd-4bd4-88ae-444eb5c26cad" path="/var/lib/kubelet/pods/3114e619-39fd-4bd4-88ae-444eb5c26cad/volumes" Dec 01 16:07:29 crc kubenswrapper[4810]: I1201 16:07:29.253876 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nbhcs" event={"ID":"230f81a0-af9d-4d25-8319-4c9990813ffa","Type":"ContainerStarted","Data":"2bd399de42e3ec006036ea6b138ef04cd65b13f1f8ef227bdf1dc3b4046c24bc"} Dec 01 16:07:29 crc kubenswrapper[4810]: I1201 16:07:29.254192 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nbhcs" event={"ID":"230f81a0-af9d-4d25-8319-4c9990813ffa","Type":"ContainerStarted","Data":"4f3dc6b8f9b0d2492913252dfdbc1d013cb34d8b1efc9fa23b0117a3afe5dbcf"} Dec 01 16:07:29 crc kubenswrapper[4810]: I1201 16:07:29.276390 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-nbhcs" podStartSLOduration=2.276372088 podStartE2EDuration="2.276372088s" podCreationTimestamp="2025-12-01 16:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:07:29.272144273 +0000 UTC m=+5615.035653886" watchObservedRunningTime="2025-12-01 16:07:29.276372088 +0000 UTC m=+5615.039881691" Dec 01 16:07:30 crc kubenswrapper[4810]: I1201 16:07:30.721651 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:07:30 crc kubenswrapper[4810]: I1201 16:07:30.779754 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69f88df96f-gfr2k"] Dec 01 16:07:30 crc kubenswrapper[4810]: I1201 16:07:30.779980 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" podUID="a55b0531-5dbb-4b9e-9c06-685e84258744" containerName="dnsmasq-dns" containerID="cri-o://8d04cf00357fea3fd3f057931b4085fa3294bc435cb5fb2c9f5a217406d808bc" gracePeriod=10 Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.255900 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.275739 4810 generic.go:334] "Generic (PLEG): container finished" podID="230f81a0-af9d-4d25-8319-4c9990813ffa" containerID="2bd399de42e3ec006036ea6b138ef04cd65b13f1f8ef227bdf1dc3b4046c24bc" exitCode=0 Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.275821 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nbhcs" event={"ID":"230f81a0-af9d-4d25-8319-4c9990813ffa","Type":"ContainerDied","Data":"2bd399de42e3ec006036ea6b138ef04cd65b13f1f8ef227bdf1dc3b4046c24bc"} Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.282257 4810 generic.go:334] "Generic (PLEG): container finished" podID="a55b0531-5dbb-4b9e-9c06-685e84258744" containerID="8d04cf00357fea3fd3f057931b4085fa3294bc435cb5fb2c9f5a217406d808bc" exitCode=0 Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.282299 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.282313 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" event={"ID":"a55b0531-5dbb-4b9e-9c06-685e84258744","Type":"ContainerDied","Data":"8d04cf00357fea3fd3f057931b4085fa3294bc435cb5fb2c9f5a217406d808bc"} Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.282345 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69f88df96f-gfr2k" event={"ID":"a55b0531-5dbb-4b9e-9c06-685e84258744","Type":"ContainerDied","Data":"e9de7f1dd2ad6c232b237a0d73348dc0411e446f27af357e82990d4a43c4cb03"} Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.282366 4810 scope.go:117] "RemoveContainer" containerID="8d04cf00357fea3fd3f057931b4085fa3294bc435cb5fb2c9f5a217406d808bc" Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.326855 4810 scope.go:117] "RemoveContainer" containerID="bf56c647e2ac9118ac042c77ec638d514850c3f9df8ba66a82e4860a1a88091d" Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.348455 4810 scope.go:117] "RemoveContainer" containerID="8d04cf00357fea3fd3f057931b4085fa3294bc435cb5fb2c9f5a217406d808bc" Dec 01 16:07:31 crc kubenswrapper[4810]: E1201 16:07:31.348978 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d04cf00357fea3fd3f057931b4085fa3294bc435cb5fb2c9f5a217406d808bc\": container with ID starting with 8d04cf00357fea3fd3f057931b4085fa3294bc435cb5fb2c9f5a217406d808bc not found: ID does not exist" containerID="8d04cf00357fea3fd3f057931b4085fa3294bc435cb5fb2c9f5a217406d808bc" Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.349012 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d04cf00357fea3fd3f057931b4085fa3294bc435cb5fb2c9f5a217406d808bc"} err="failed to get container status \"8d04cf00357fea3fd3f057931b4085fa3294bc435cb5fb2c9f5a217406d808bc\": rpc error: code = NotFound desc = could not find container \"8d04cf00357fea3fd3f057931b4085fa3294bc435cb5fb2c9f5a217406d808bc\": container with ID starting with 8d04cf00357fea3fd3f057931b4085fa3294bc435cb5fb2c9f5a217406d808bc not found: ID does not exist" Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.349038 4810 scope.go:117] "RemoveContainer" containerID="bf56c647e2ac9118ac042c77ec638d514850c3f9df8ba66a82e4860a1a88091d" Dec 01 16:07:31 crc kubenswrapper[4810]: E1201 16:07:31.350453 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf56c647e2ac9118ac042c77ec638d514850c3f9df8ba66a82e4860a1a88091d\": container with ID starting with bf56c647e2ac9118ac042c77ec638d514850c3f9df8ba66a82e4860a1a88091d not found: ID does not exist" containerID="bf56c647e2ac9118ac042c77ec638d514850c3f9df8ba66a82e4860a1a88091d" Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.350502 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf56c647e2ac9118ac042c77ec638d514850c3f9df8ba66a82e4860a1a88091d"} err="failed to get container status \"bf56c647e2ac9118ac042c77ec638d514850c3f9df8ba66a82e4860a1a88091d\": rpc error: code = NotFound desc = could not find container \"bf56c647e2ac9118ac042c77ec638d514850c3f9df8ba66a82e4860a1a88091d\": container with ID starting with bf56c647e2ac9118ac042c77ec638d514850c3f9df8ba66a82e4860a1a88091d not found: ID does not exist" Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.437327 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-config\") pod \"a55b0531-5dbb-4b9e-9c06-685e84258744\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.437415 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9p2z\" (UniqueName: \"kubernetes.io/projected/a55b0531-5dbb-4b9e-9c06-685e84258744-kube-api-access-t9p2z\") pod \"a55b0531-5dbb-4b9e-9c06-685e84258744\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.437465 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-dns-svc\") pod \"a55b0531-5dbb-4b9e-9c06-685e84258744\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.437505 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-ovsdbserver-sb\") pod \"a55b0531-5dbb-4b9e-9c06-685e84258744\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.437539 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-ovsdbserver-nb\") pod \"a55b0531-5dbb-4b9e-9c06-685e84258744\" (UID: \"a55b0531-5dbb-4b9e-9c06-685e84258744\") " Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.444754 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a55b0531-5dbb-4b9e-9c06-685e84258744-kube-api-access-t9p2z" (OuterVolumeSpecName: "kube-api-access-t9p2z") pod "a55b0531-5dbb-4b9e-9c06-685e84258744" (UID: "a55b0531-5dbb-4b9e-9c06-685e84258744"). InnerVolumeSpecName "kube-api-access-t9p2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.474327 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-config" (OuterVolumeSpecName: "config") pod "a55b0531-5dbb-4b9e-9c06-685e84258744" (UID: "a55b0531-5dbb-4b9e-9c06-685e84258744"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.475555 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a55b0531-5dbb-4b9e-9c06-685e84258744" (UID: "a55b0531-5dbb-4b9e-9c06-685e84258744"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.480615 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a55b0531-5dbb-4b9e-9c06-685e84258744" (UID: "a55b0531-5dbb-4b9e-9c06-685e84258744"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.480830 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a55b0531-5dbb-4b9e-9c06-685e84258744" (UID: "a55b0531-5dbb-4b9e-9c06-685e84258744"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.539226 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.539273 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.539287 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9p2z\" (UniqueName: \"kubernetes.io/projected/a55b0531-5dbb-4b9e-9c06-685e84258744-kube-api-access-t9p2z\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.539297 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.539305 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a55b0531-5dbb-4b9e-9c06-685e84258744-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.615072 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69f88df96f-gfr2k"] Dec 01 16:07:31 crc kubenswrapper[4810]: I1201 16:07:31.631279 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-69f88df96f-gfr2k"] Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.511214 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a55b0531-5dbb-4b9e-9c06-685e84258744" path="/var/lib/kubelet/pods/a55b0531-5dbb-4b9e-9c06-685e84258744/volumes" Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.674065 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.856976 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-scripts\") pod \"230f81a0-af9d-4d25-8319-4c9990813ffa\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.857032 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wr58\" (UniqueName: \"kubernetes.io/projected/230f81a0-af9d-4d25-8319-4c9990813ffa-kube-api-access-2wr58\") pod \"230f81a0-af9d-4d25-8319-4c9990813ffa\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.857064 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-fernet-keys\") pod \"230f81a0-af9d-4d25-8319-4c9990813ffa\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.857159 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-combined-ca-bundle\") pod \"230f81a0-af9d-4d25-8319-4c9990813ffa\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.857194 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-config-data\") pod \"230f81a0-af9d-4d25-8319-4c9990813ffa\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.857239 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-credential-keys\") pod \"230f81a0-af9d-4d25-8319-4c9990813ffa\" (UID: \"230f81a0-af9d-4d25-8319-4c9990813ffa\") " Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.861781 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "230f81a0-af9d-4d25-8319-4c9990813ffa" (UID: "230f81a0-af9d-4d25-8319-4c9990813ffa"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.862819 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-scripts" (OuterVolumeSpecName: "scripts") pod "230f81a0-af9d-4d25-8319-4c9990813ffa" (UID: "230f81a0-af9d-4d25-8319-4c9990813ffa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.862973 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/230f81a0-af9d-4d25-8319-4c9990813ffa-kube-api-access-2wr58" (OuterVolumeSpecName: "kube-api-access-2wr58") pod "230f81a0-af9d-4d25-8319-4c9990813ffa" (UID: "230f81a0-af9d-4d25-8319-4c9990813ffa"). InnerVolumeSpecName "kube-api-access-2wr58". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.864061 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "230f81a0-af9d-4d25-8319-4c9990813ffa" (UID: "230f81a0-af9d-4d25-8319-4c9990813ffa"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.883421 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-config-data" (OuterVolumeSpecName: "config-data") pod "230f81a0-af9d-4d25-8319-4c9990813ffa" (UID: "230f81a0-af9d-4d25-8319-4c9990813ffa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.887041 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "230f81a0-af9d-4d25-8319-4c9990813ffa" (UID: "230f81a0-af9d-4d25-8319-4c9990813ffa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.959911 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.959942 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wr58\" (UniqueName: \"kubernetes.io/projected/230f81a0-af9d-4d25-8319-4c9990813ffa-kube-api-access-2wr58\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.959956 4810 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.959968 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.959982 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:32 crc kubenswrapper[4810]: I1201 16:07:32.959992 4810 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/230f81a0-af9d-4d25-8319-4c9990813ffa-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.304675 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-nbhcs" event={"ID":"230f81a0-af9d-4d25-8319-4c9990813ffa","Type":"ContainerDied","Data":"4f3dc6b8f9b0d2492913252dfdbc1d013cb34d8b1efc9fa23b0117a3afe5dbcf"} Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.304710 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f3dc6b8f9b0d2492913252dfdbc1d013cb34d8b1efc9fa23b0117a3afe5dbcf" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.304738 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-nbhcs" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.396562 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7d4469b5c6-z5bkt"] Dec 01 16:07:33 crc kubenswrapper[4810]: E1201 16:07:33.396877 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a55b0531-5dbb-4b9e-9c06-685e84258744" containerName="dnsmasq-dns" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.396893 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a55b0531-5dbb-4b9e-9c06-685e84258744" containerName="dnsmasq-dns" Dec 01 16:07:33 crc kubenswrapper[4810]: E1201 16:07:33.396908 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a55b0531-5dbb-4b9e-9c06-685e84258744" containerName="init" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.396914 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a55b0531-5dbb-4b9e-9c06-685e84258744" containerName="init" Dec 01 16:07:33 crc kubenswrapper[4810]: E1201 16:07:33.396929 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="230f81a0-af9d-4d25-8319-4c9990813ffa" containerName="keystone-bootstrap" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.396936 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="230f81a0-af9d-4d25-8319-4c9990813ffa" containerName="keystone-bootstrap" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.397093 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="230f81a0-af9d-4d25-8319-4c9990813ffa" containerName="keystone-bootstrap" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.397105 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a55b0531-5dbb-4b9e-9c06-685e84258744" containerName="dnsmasq-dns" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.397702 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.400966 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.401315 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.401539 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.401755 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.401943 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.408883 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vl9m4" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.411978 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7d4469b5c6-z5bkt"] Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.568733 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-credential-keys\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.568801 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-config-data\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.568828 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-combined-ca-bundle\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.568902 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-fernet-keys\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.568959 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-public-tls-certs\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.568996 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-scripts\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.569027 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pttww\" (UniqueName: \"kubernetes.io/projected/83a32a66-545a-4122-a47e-a7111a5d745f-kube-api-access-pttww\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.569060 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-internal-tls-certs\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.670967 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pttww\" (UniqueName: \"kubernetes.io/projected/83a32a66-545a-4122-a47e-a7111a5d745f-kube-api-access-pttww\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.671050 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-internal-tls-certs\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.671113 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-credential-keys\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.671141 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-config-data\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.671166 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-combined-ca-bundle\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.671198 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-fernet-keys\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.671824 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-public-tls-certs\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.671873 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-scripts\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.675172 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-fernet-keys\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.675169 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-credential-keys\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.675326 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-public-tls-certs\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.675655 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-combined-ca-bundle\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.675725 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-scripts\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.676733 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-internal-tls-certs\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.676818 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83a32a66-545a-4122-a47e-a7111a5d745f-config-data\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.686860 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pttww\" (UniqueName: \"kubernetes.io/projected/83a32a66-545a-4122-a47e-a7111a5d745f-kube-api-access-pttww\") pod \"keystone-7d4469b5c6-z5bkt\" (UID: \"83a32a66-545a-4122-a47e-a7111a5d745f\") " pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:33 crc kubenswrapper[4810]: I1201 16:07:33.731037 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:34 crc kubenswrapper[4810]: I1201 16:07:34.207503 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7d4469b5c6-z5bkt"] Dec 01 16:07:34 crc kubenswrapper[4810]: I1201 16:07:34.312895 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7d4469b5c6-z5bkt" event={"ID":"83a32a66-545a-4122-a47e-a7111a5d745f","Type":"ContainerStarted","Data":"c5ad5b1c466fc3ef2c8dff81069effd555285de7b10b60d270be69711ff286fd"} Dec 01 16:07:35 crc kubenswrapper[4810]: I1201 16:07:35.344967 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7d4469b5c6-z5bkt" event={"ID":"83a32a66-545a-4122-a47e-a7111a5d745f","Type":"ContainerStarted","Data":"64a4c55a816f3d63fb2a520ef2ec145848c554a92c24db422c52a235d927423f"} Dec 01 16:07:35 crc kubenswrapper[4810]: I1201 16:07:35.346624 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:07:35 crc kubenswrapper[4810]: I1201 16:07:35.386370 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7d4469b5c6-z5bkt" podStartSLOduration=2.386349336 podStartE2EDuration="2.386349336s" podCreationTimestamp="2025-12-01 16:07:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:07:35.37472044 +0000 UTC m=+5621.138230093" watchObservedRunningTime="2025-12-01 16:07:35.386349336 +0000 UTC m=+5621.149858939" Dec 01 16:08:05 crc kubenswrapper[4810]: I1201 16:08:05.296288 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7d4469b5c6-z5bkt" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.162458 4810 scope.go:117] "RemoveContainer" containerID="49ec02d5346e11e8f636f3169f4c6d7f580b7ea8185198a5cfc6b935e42c9ca3" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.191367 4810 scope.go:117] "RemoveContainer" containerID="289328ff53e4549ae2ffb4684348ec80845c060d1c68c9dacde24aa29bac7b01" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.245511 4810 scope.go:117] "RemoveContainer" containerID="d451256e7412bbd5f2eef5a3d23a3015d190d220b52d9b5f40ca6bd12e47d0cd" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.327741 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.329381 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.338657 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.338670 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.338937 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-7crq9" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.350538 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.464170 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vn6f\" (UniqueName: \"kubernetes.io/projected/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-kube-api-access-6vn6f\") pod \"openstackclient\" (UID: \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\") " pod="openstack/openstackclient" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.464536 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-openstack-config\") pod \"openstackclient\" (UID: \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\") " pod="openstack/openstackclient" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.464567 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\") " pod="openstack/openstackclient" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.464686 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-openstack-config-secret\") pod \"openstackclient\" (UID: \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\") " pod="openstack/openstackclient" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.566741 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-openstack-config\") pod \"openstackclient\" (UID: \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\") " pod="openstack/openstackclient" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.566781 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\") " pod="openstack/openstackclient" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.566844 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-openstack-config-secret\") pod \"openstackclient\" (UID: \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\") " pod="openstack/openstackclient" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.566933 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vn6f\" (UniqueName: \"kubernetes.io/projected/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-kube-api-access-6vn6f\") pod \"openstackclient\" (UID: \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\") " pod="openstack/openstackclient" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.567622 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-openstack-config\") pod \"openstackclient\" (UID: \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\") " pod="openstack/openstackclient" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.572694 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\") " pod="openstack/openstackclient" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.573306 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-openstack-config-secret\") pod \"openstackclient\" (UID: \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\") " pod="openstack/openstackclient" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.583697 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vn6f\" (UniqueName: \"kubernetes.io/projected/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-kube-api-access-6vn6f\") pod \"openstackclient\" (UID: \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\") " pod="openstack/openstackclient" Dec 01 16:08:09 crc kubenswrapper[4810]: I1201 16:08:09.673354 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 16:08:10 crc kubenswrapper[4810]: I1201 16:08:10.076786 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 16:08:10 crc kubenswrapper[4810]: I1201 16:08:10.612863 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b","Type":"ContainerStarted","Data":"73589ed04e6cb6c6075477260b43267dffa628c16a6dc292c1ae630bdd4bd7b0"} Dec 01 16:08:10 crc kubenswrapper[4810]: I1201 16:08:10.613174 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b","Type":"ContainerStarted","Data":"e0f6b6ec582dd4c76a7781935e4d366945bec7194fa207429b53c5083be4de02"} Dec 01 16:09:09 crc kubenswrapper[4810]: I1201 16:09:09.391405 4810 scope.go:117] "RemoveContainer" containerID="a9db3594ccf17b32f09e39d23617e8750b25df5c2b02d901949530612937db9d" Dec 01 16:09:09 crc kubenswrapper[4810]: I1201 16:09:09.415390 4810 scope.go:117] "RemoveContainer" containerID="8e997bb64a2abd3d4d7209ac74ffb5c1b18e2c592335bfe3b3f06baac4e03f1d" Dec 01 16:09:32 crc kubenswrapper[4810]: I1201 16:09:32.971990 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:09:32 crc kubenswrapper[4810]: I1201 16:09:32.972374 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.631183 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=93.631165858 podStartE2EDuration="1m33.631165858s" podCreationTimestamp="2025-12-01 16:08:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:08:10.626098819 +0000 UTC m=+5656.389608432" watchObservedRunningTime="2025-12-01 16:09:42.631165858 +0000 UTC m=+5748.394675461" Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.638294 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-t57h9"] Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.639391 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-t57h9" Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.650246 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-t57h9"] Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.671904 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9cg9\" (UniqueName: \"kubernetes.io/projected/ea8c6b74-11cf-4f7c-be88-5715592abf4b-kube-api-access-f9cg9\") pod \"barbican-db-create-t57h9\" (UID: \"ea8c6b74-11cf-4f7c-be88-5715592abf4b\") " pod="openstack/barbican-db-create-t57h9" Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.671964 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea8c6b74-11cf-4f7c-be88-5715592abf4b-operator-scripts\") pod \"barbican-db-create-t57h9\" (UID: \"ea8c6b74-11cf-4f7c-be88-5715592abf4b\") " pod="openstack/barbican-db-create-t57h9" Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.734322 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-fc84-account-create-update-wgmzd"] Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.735647 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fc84-account-create-update-wgmzd" Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.738133 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.744041 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-fc84-account-create-update-wgmzd"] Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.773823 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6z4h\" (UniqueName: \"kubernetes.io/projected/25572bab-6803-44f8-b3d9-98491d59fe7b-kube-api-access-h6z4h\") pod \"barbican-fc84-account-create-update-wgmzd\" (UID: \"25572bab-6803-44f8-b3d9-98491d59fe7b\") " pod="openstack/barbican-fc84-account-create-update-wgmzd" Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.773931 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25572bab-6803-44f8-b3d9-98491d59fe7b-operator-scripts\") pod \"barbican-fc84-account-create-update-wgmzd\" (UID: \"25572bab-6803-44f8-b3d9-98491d59fe7b\") " pod="openstack/barbican-fc84-account-create-update-wgmzd" Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.773976 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9cg9\" (UniqueName: \"kubernetes.io/projected/ea8c6b74-11cf-4f7c-be88-5715592abf4b-kube-api-access-f9cg9\") pod \"barbican-db-create-t57h9\" (UID: \"ea8c6b74-11cf-4f7c-be88-5715592abf4b\") " pod="openstack/barbican-db-create-t57h9" Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.774009 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea8c6b74-11cf-4f7c-be88-5715592abf4b-operator-scripts\") pod \"barbican-db-create-t57h9\" (UID: \"ea8c6b74-11cf-4f7c-be88-5715592abf4b\") " pod="openstack/barbican-db-create-t57h9" Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.774882 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea8c6b74-11cf-4f7c-be88-5715592abf4b-operator-scripts\") pod \"barbican-db-create-t57h9\" (UID: \"ea8c6b74-11cf-4f7c-be88-5715592abf4b\") " pod="openstack/barbican-db-create-t57h9" Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.795114 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9cg9\" (UniqueName: \"kubernetes.io/projected/ea8c6b74-11cf-4f7c-be88-5715592abf4b-kube-api-access-f9cg9\") pod \"barbican-db-create-t57h9\" (UID: \"ea8c6b74-11cf-4f7c-be88-5715592abf4b\") " pod="openstack/barbican-db-create-t57h9" Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.875627 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6z4h\" (UniqueName: \"kubernetes.io/projected/25572bab-6803-44f8-b3d9-98491d59fe7b-kube-api-access-h6z4h\") pod \"barbican-fc84-account-create-update-wgmzd\" (UID: \"25572bab-6803-44f8-b3d9-98491d59fe7b\") " pod="openstack/barbican-fc84-account-create-update-wgmzd" Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.875715 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25572bab-6803-44f8-b3d9-98491d59fe7b-operator-scripts\") pod \"barbican-fc84-account-create-update-wgmzd\" (UID: \"25572bab-6803-44f8-b3d9-98491d59fe7b\") " pod="openstack/barbican-fc84-account-create-update-wgmzd" Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.876521 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25572bab-6803-44f8-b3d9-98491d59fe7b-operator-scripts\") pod \"barbican-fc84-account-create-update-wgmzd\" (UID: \"25572bab-6803-44f8-b3d9-98491d59fe7b\") " pod="openstack/barbican-fc84-account-create-update-wgmzd" Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.892731 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6z4h\" (UniqueName: \"kubernetes.io/projected/25572bab-6803-44f8-b3d9-98491d59fe7b-kube-api-access-h6z4h\") pod \"barbican-fc84-account-create-update-wgmzd\" (UID: \"25572bab-6803-44f8-b3d9-98491d59fe7b\") " pod="openstack/barbican-fc84-account-create-update-wgmzd" Dec 01 16:09:42 crc kubenswrapper[4810]: I1201 16:09:42.966315 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-t57h9" Dec 01 16:09:43 crc kubenswrapper[4810]: I1201 16:09:43.057950 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fc84-account-create-update-wgmzd" Dec 01 16:09:43 crc kubenswrapper[4810]: I1201 16:09:43.351007 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-fc84-account-create-update-wgmzd"] Dec 01 16:09:43 crc kubenswrapper[4810]: I1201 16:09:43.381687 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-fc84-account-create-update-wgmzd" event={"ID":"25572bab-6803-44f8-b3d9-98491d59fe7b","Type":"ContainerStarted","Data":"3c856ac5095c0a7d0bc4ebfd78c7032af8a34305a09bc1f86a142b60a9795681"} Dec 01 16:09:43 crc kubenswrapper[4810]: W1201 16:09:43.382477 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea8c6b74_11cf_4f7c_be88_5715592abf4b.slice/crio-dc64d26051ff9f2dd31580aabda8ce5a287082ea9268248bb70a4185d9b02d0d WatchSource:0}: Error finding container dc64d26051ff9f2dd31580aabda8ce5a287082ea9268248bb70a4185d9b02d0d: Status 404 returned error can't find the container with id dc64d26051ff9f2dd31580aabda8ce5a287082ea9268248bb70a4185d9b02d0d Dec 01 16:09:43 crc kubenswrapper[4810]: I1201 16:09:43.385431 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-t57h9"] Dec 01 16:09:44 crc kubenswrapper[4810]: I1201 16:09:44.391984 4810 generic.go:334] "Generic (PLEG): container finished" podID="25572bab-6803-44f8-b3d9-98491d59fe7b" containerID="84f0047271368cf2223c8fbcd2ff1ab071bee50b8138c1d762f09707ef546dc5" exitCode=0 Dec 01 16:09:44 crc kubenswrapper[4810]: I1201 16:09:44.392060 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-fc84-account-create-update-wgmzd" event={"ID":"25572bab-6803-44f8-b3d9-98491d59fe7b","Type":"ContainerDied","Data":"84f0047271368cf2223c8fbcd2ff1ab071bee50b8138c1d762f09707ef546dc5"} Dec 01 16:09:44 crc kubenswrapper[4810]: I1201 16:09:44.393712 4810 generic.go:334] "Generic (PLEG): container finished" podID="ea8c6b74-11cf-4f7c-be88-5715592abf4b" containerID="7173470a69c19c6fcbcf89a00f13376b1f317809ba6047ff859d127b2a57ad6f" exitCode=0 Dec 01 16:09:44 crc kubenswrapper[4810]: I1201 16:09:44.393740 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-t57h9" event={"ID":"ea8c6b74-11cf-4f7c-be88-5715592abf4b","Type":"ContainerDied","Data":"7173470a69c19c6fcbcf89a00f13376b1f317809ba6047ff859d127b2a57ad6f"} Dec 01 16:09:44 crc kubenswrapper[4810]: I1201 16:09:44.393771 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-t57h9" event={"ID":"ea8c6b74-11cf-4f7c-be88-5715592abf4b","Type":"ContainerStarted","Data":"dc64d26051ff9f2dd31580aabda8ce5a287082ea9268248bb70a4185d9b02d0d"} Dec 01 16:09:45 crc kubenswrapper[4810]: I1201 16:09:45.789653 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-t57h9" Dec 01 16:09:45 crc kubenswrapper[4810]: I1201 16:09:45.799007 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fc84-account-create-update-wgmzd" Dec 01 16:09:45 crc kubenswrapper[4810]: I1201 16:09:45.967272 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25572bab-6803-44f8-b3d9-98491d59fe7b-operator-scripts\") pod \"25572bab-6803-44f8-b3d9-98491d59fe7b\" (UID: \"25572bab-6803-44f8-b3d9-98491d59fe7b\") " Dec 01 16:09:45 crc kubenswrapper[4810]: I1201 16:09:45.967343 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9cg9\" (UniqueName: \"kubernetes.io/projected/ea8c6b74-11cf-4f7c-be88-5715592abf4b-kube-api-access-f9cg9\") pod \"ea8c6b74-11cf-4f7c-be88-5715592abf4b\" (UID: \"ea8c6b74-11cf-4f7c-be88-5715592abf4b\") " Dec 01 16:09:45 crc kubenswrapper[4810]: I1201 16:09:45.967448 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6z4h\" (UniqueName: \"kubernetes.io/projected/25572bab-6803-44f8-b3d9-98491d59fe7b-kube-api-access-h6z4h\") pod \"25572bab-6803-44f8-b3d9-98491d59fe7b\" (UID: \"25572bab-6803-44f8-b3d9-98491d59fe7b\") " Dec 01 16:09:45 crc kubenswrapper[4810]: I1201 16:09:45.967529 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea8c6b74-11cf-4f7c-be88-5715592abf4b-operator-scripts\") pod \"ea8c6b74-11cf-4f7c-be88-5715592abf4b\" (UID: \"ea8c6b74-11cf-4f7c-be88-5715592abf4b\") " Dec 01 16:09:45 crc kubenswrapper[4810]: I1201 16:09:45.968221 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25572bab-6803-44f8-b3d9-98491d59fe7b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "25572bab-6803-44f8-b3d9-98491d59fe7b" (UID: "25572bab-6803-44f8-b3d9-98491d59fe7b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:09:45 crc kubenswrapper[4810]: I1201 16:09:45.968303 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea8c6b74-11cf-4f7c-be88-5715592abf4b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ea8c6b74-11cf-4f7c-be88-5715592abf4b" (UID: "ea8c6b74-11cf-4f7c-be88-5715592abf4b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:09:45 crc kubenswrapper[4810]: I1201 16:09:45.983812 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea8c6b74-11cf-4f7c-be88-5715592abf4b-kube-api-access-f9cg9" (OuterVolumeSpecName: "kube-api-access-f9cg9") pod "ea8c6b74-11cf-4f7c-be88-5715592abf4b" (UID: "ea8c6b74-11cf-4f7c-be88-5715592abf4b"). InnerVolumeSpecName "kube-api-access-f9cg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:09:45 crc kubenswrapper[4810]: I1201 16:09:45.986683 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25572bab-6803-44f8-b3d9-98491d59fe7b-kube-api-access-h6z4h" (OuterVolumeSpecName: "kube-api-access-h6z4h") pod "25572bab-6803-44f8-b3d9-98491d59fe7b" (UID: "25572bab-6803-44f8-b3d9-98491d59fe7b"). InnerVolumeSpecName "kube-api-access-h6z4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:09:46 crc kubenswrapper[4810]: I1201 16:09:46.069593 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25572bab-6803-44f8-b3d9-98491d59fe7b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:09:46 crc kubenswrapper[4810]: I1201 16:09:46.069889 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9cg9\" (UniqueName: \"kubernetes.io/projected/ea8c6b74-11cf-4f7c-be88-5715592abf4b-kube-api-access-f9cg9\") on node \"crc\" DevicePath \"\"" Dec 01 16:09:46 crc kubenswrapper[4810]: I1201 16:09:46.069900 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6z4h\" (UniqueName: \"kubernetes.io/projected/25572bab-6803-44f8-b3d9-98491d59fe7b-kube-api-access-h6z4h\") on node \"crc\" DevicePath \"\"" Dec 01 16:09:46 crc kubenswrapper[4810]: I1201 16:09:46.069909 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea8c6b74-11cf-4f7c-be88-5715592abf4b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:09:46 crc kubenswrapper[4810]: I1201 16:09:46.412426 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fc84-account-create-update-wgmzd" Dec 01 16:09:46 crc kubenswrapper[4810]: I1201 16:09:46.412453 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-fc84-account-create-update-wgmzd" event={"ID":"25572bab-6803-44f8-b3d9-98491d59fe7b","Type":"ContainerDied","Data":"3c856ac5095c0a7d0bc4ebfd78c7032af8a34305a09bc1f86a142b60a9795681"} Dec 01 16:09:46 crc kubenswrapper[4810]: I1201 16:09:46.412493 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c856ac5095c0a7d0bc4ebfd78c7032af8a34305a09bc1f86a142b60a9795681" Dec 01 16:09:46 crc kubenswrapper[4810]: I1201 16:09:46.418082 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-t57h9" event={"ID":"ea8c6b74-11cf-4f7c-be88-5715592abf4b","Type":"ContainerDied","Data":"dc64d26051ff9f2dd31580aabda8ce5a287082ea9268248bb70a4185d9b02d0d"} Dec 01 16:09:46 crc kubenswrapper[4810]: I1201 16:09:46.418103 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-t57h9" Dec 01 16:09:46 crc kubenswrapper[4810]: I1201 16:09:46.418117 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc64d26051ff9f2dd31580aabda8ce5a287082ea9268248bb70a4185d9b02d0d" Dec 01 16:09:47 crc kubenswrapper[4810]: I1201 16:09:47.992774 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-g8lkr"] Dec 01 16:09:47 crc kubenswrapper[4810]: E1201 16:09:47.993403 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea8c6b74-11cf-4f7c-be88-5715592abf4b" containerName="mariadb-database-create" Dec 01 16:09:47 crc kubenswrapper[4810]: I1201 16:09:47.993420 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea8c6b74-11cf-4f7c-be88-5715592abf4b" containerName="mariadb-database-create" Dec 01 16:09:47 crc kubenswrapper[4810]: E1201 16:09:47.993448 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25572bab-6803-44f8-b3d9-98491d59fe7b" containerName="mariadb-account-create-update" Dec 01 16:09:47 crc kubenswrapper[4810]: I1201 16:09:47.993457 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="25572bab-6803-44f8-b3d9-98491d59fe7b" containerName="mariadb-account-create-update" Dec 01 16:09:47 crc kubenswrapper[4810]: I1201 16:09:47.993686 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="25572bab-6803-44f8-b3d9-98491d59fe7b" containerName="mariadb-account-create-update" Dec 01 16:09:47 crc kubenswrapper[4810]: I1201 16:09:47.993715 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea8c6b74-11cf-4f7c-be88-5715592abf4b" containerName="mariadb-database-create" Dec 01 16:09:47 crc kubenswrapper[4810]: I1201 16:09:47.994365 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-g8lkr" Dec 01 16:09:48 crc kubenswrapper[4810]: I1201 16:09:47.999981 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 01 16:09:48 crc kubenswrapper[4810]: I1201 16:09:48.002053 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-kl86b" Dec 01 16:09:48 crc kubenswrapper[4810]: I1201 16:09:48.002708 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-g8lkr"] Dec 01 16:09:48 crc kubenswrapper[4810]: I1201 16:09:48.100445 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmc8t\" (UniqueName: \"kubernetes.io/projected/1dc4b16f-f399-4883-bb29-0ec81865287f-kube-api-access-nmc8t\") pod \"barbican-db-sync-g8lkr\" (UID: \"1dc4b16f-f399-4883-bb29-0ec81865287f\") " pod="openstack/barbican-db-sync-g8lkr" Dec 01 16:09:48 crc kubenswrapper[4810]: I1201 16:09:48.100569 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dc4b16f-f399-4883-bb29-0ec81865287f-combined-ca-bundle\") pod \"barbican-db-sync-g8lkr\" (UID: \"1dc4b16f-f399-4883-bb29-0ec81865287f\") " pod="openstack/barbican-db-sync-g8lkr" Dec 01 16:09:48 crc kubenswrapper[4810]: I1201 16:09:48.100656 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1dc4b16f-f399-4883-bb29-0ec81865287f-db-sync-config-data\") pod \"barbican-db-sync-g8lkr\" (UID: \"1dc4b16f-f399-4883-bb29-0ec81865287f\") " pod="openstack/barbican-db-sync-g8lkr" Dec 01 16:09:48 crc kubenswrapper[4810]: I1201 16:09:48.201840 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmc8t\" (UniqueName: \"kubernetes.io/projected/1dc4b16f-f399-4883-bb29-0ec81865287f-kube-api-access-nmc8t\") pod \"barbican-db-sync-g8lkr\" (UID: \"1dc4b16f-f399-4883-bb29-0ec81865287f\") " pod="openstack/barbican-db-sync-g8lkr" Dec 01 16:09:48 crc kubenswrapper[4810]: I1201 16:09:48.201914 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dc4b16f-f399-4883-bb29-0ec81865287f-combined-ca-bundle\") pod \"barbican-db-sync-g8lkr\" (UID: \"1dc4b16f-f399-4883-bb29-0ec81865287f\") " pod="openstack/barbican-db-sync-g8lkr" Dec 01 16:09:48 crc kubenswrapper[4810]: I1201 16:09:48.201999 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1dc4b16f-f399-4883-bb29-0ec81865287f-db-sync-config-data\") pod \"barbican-db-sync-g8lkr\" (UID: \"1dc4b16f-f399-4883-bb29-0ec81865287f\") " pod="openstack/barbican-db-sync-g8lkr" Dec 01 16:09:48 crc kubenswrapper[4810]: I1201 16:09:48.206360 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1dc4b16f-f399-4883-bb29-0ec81865287f-db-sync-config-data\") pod \"barbican-db-sync-g8lkr\" (UID: \"1dc4b16f-f399-4883-bb29-0ec81865287f\") " pod="openstack/barbican-db-sync-g8lkr" Dec 01 16:09:48 crc kubenswrapper[4810]: I1201 16:09:48.216140 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dc4b16f-f399-4883-bb29-0ec81865287f-combined-ca-bundle\") pod \"barbican-db-sync-g8lkr\" (UID: \"1dc4b16f-f399-4883-bb29-0ec81865287f\") " pod="openstack/barbican-db-sync-g8lkr" Dec 01 16:09:48 crc kubenswrapper[4810]: I1201 16:09:48.218815 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmc8t\" (UniqueName: \"kubernetes.io/projected/1dc4b16f-f399-4883-bb29-0ec81865287f-kube-api-access-nmc8t\") pod \"barbican-db-sync-g8lkr\" (UID: \"1dc4b16f-f399-4883-bb29-0ec81865287f\") " pod="openstack/barbican-db-sync-g8lkr" Dec 01 16:09:48 crc kubenswrapper[4810]: I1201 16:09:48.359703 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-g8lkr" Dec 01 16:09:48 crc kubenswrapper[4810]: I1201 16:09:48.792489 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-g8lkr"] Dec 01 16:09:49 crc kubenswrapper[4810]: I1201 16:09:49.443197 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-g8lkr" event={"ID":"1dc4b16f-f399-4883-bb29-0ec81865287f","Type":"ContainerStarted","Data":"51ce5819550849f51a2704ce44356a9f9400058434bc9d595bc1d317b876e573"} Dec 01 16:09:49 crc kubenswrapper[4810]: I1201 16:09:49.443563 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-g8lkr" event={"ID":"1dc4b16f-f399-4883-bb29-0ec81865287f","Type":"ContainerStarted","Data":"031cea30c04fb85e505c09c345fa59ec85f68072b9e8cea842be457ae9e63aec"} Dec 01 16:09:49 crc kubenswrapper[4810]: I1201 16:09:49.463021 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-g8lkr" podStartSLOduration=2.463000301 podStartE2EDuration="2.463000301s" podCreationTimestamp="2025-12-01 16:09:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:09:49.456461173 +0000 UTC m=+5755.219970786" watchObservedRunningTime="2025-12-01 16:09:49.463000301 +0000 UTC m=+5755.226509904" Dec 01 16:09:51 crc kubenswrapper[4810]: I1201 16:09:51.457397 4810 generic.go:334] "Generic (PLEG): container finished" podID="1dc4b16f-f399-4883-bb29-0ec81865287f" containerID="51ce5819550849f51a2704ce44356a9f9400058434bc9d595bc1d317b876e573" exitCode=0 Dec 01 16:09:51 crc kubenswrapper[4810]: I1201 16:09:51.457438 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-g8lkr" event={"ID":"1dc4b16f-f399-4883-bb29-0ec81865287f","Type":"ContainerDied","Data":"51ce5819550849f51a2704ce44356a9f9400058434bc9d595bc1d317b876e573"} Dec 01 16:09:52 crc kubenswrapper[4810]: I1201 16:09:52.796011 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-g8lkr" Dec 01 16:09:52 crc kubenswrapper[4810]: I1201 16:09:52.985887 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmc8t\" (UniqueName: \"kubernetes.io/projected/1dc4b16f-f399-4883-bb29-0ec81865287f-kube-api-access-nmc8t\") pod \"1dc4b16f-f399-4883-bb29-0ec81865287f\" (UID: \"1dc4b16f-f399-4883-bb29-0ec81865287f\") " Dec 01 16:09:52 crc kubenswrapper[4810]: I1201 16:09:52.985985 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dc4b16f-f399-4883-bb29-0ec81865287f-combined-ca-bundle\") pod \"1dc4b16f-f399-4883-bb29-0ec81865287f\" (UID: \"1dc4b16f-f399-4883-bb29-0ec81865287f\") " Dec 01 16:09:52 crc kubenswrapper[4810]: I1201 16:09:52.986080 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1dc4b16f-f399-4883-bb29-0ec81865287f-db-sync-config-data\") pod \"1dc4b16f-f399-4883-bb29-0ec81865287f\" (UID: \"1dc4b16f-f399-4883-bb29-0ec81865287f\") " Dec 01 16:09:52 crc kubenswrapper[4810]: I1201 16:09:52.991909 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dc4b16f-f399-4883-bb29-0ec81865287f-kube-api-access-nmc8t" (OuterVolumeSpecName: "kube-api-access-nmc8t") pod "1dc4b16f-f399-4883-bb29-0ec81865287f" (UID: "1dc4b16f-f399-4883-bb29-0ec81865287f"). InnerVolumeSpecName "kube-api-access-nmc8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:09:52 crc kubenswrapper[4810]: I1201 16:09:52.992030 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dc4b16f-f399-4883-bb29-0ec81865287f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "1dc4b16f-f399-4883-bb29-0ec81865287f" (UID: "1dc4b16f-f399-4883-bb29-0ec81865287f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.013809 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dc4b16f-f399-4883-bb29-0ec81865287f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1dc4b16f-f399-4883-bb29-0ec81865287f" (UID: "1dc4b16f-f399-4883-bb29-0ec81865287f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.087910 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dc4b16f-f399-4883-bb29-0ec81865287f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.087948 4810 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1dc4b16f-f399-4883-bb29-0ec81865287f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.087961 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmc8t\" (UniqueName: \"kubernetes.io/projected/1dc4b16f-f399-4883-bb29-0ec81865287f-kube-api-access-nmc8t\") on node \"crc\" DevicePath \"\"" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.475053 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-g8lkr" event={"ID":"1dc4b16f-f399-4883-bb29-0ec81865287f","Type":"ContainerDied","Data":"031cea30c04fb85e505c09c345fa59ec85f68072b9e8cea842be457ae9e63aec"} Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.475097 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="031cea30c04fb85e505c09c345fa59ec85f68072b9e8cea842be457ae9e63aec" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.475101 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-g8lkr" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.720700 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-855f768b54-b52dt"] Dec 01 16:09:53 crc kubenswrapper[4810]: E1201 16:09:53.721181 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dc4b16f-f399-4883-bb29-0ec81865287f" containerName="barbican-db-sync" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.721204 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dc4b16f-f399-4883-bb29-0ec81865287f" containerName="barbican-db-sync" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.721425 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dc4b16f-f399-4883-bb29-0ec81865287f" containerName="barbican-db-sync" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.722331 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-855f768b54-b52dt" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.731747 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-kl86b" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.731975 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.732102 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.744521 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-679666fcbc-r7dpm"] Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.746013 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-679666fcbc-r7dpm" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.755387 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.757559 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-855f768b54-b52dt"] Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.786603 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-679666fcbc-r7dpm"] Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.799629 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/835bf15c-547d-4bb3-89bd-767b9c7e8f94-logs\") pod \"barbican-worker-679666fcbc-r7dpm\" (UID: \"835bf15c-547d-4bb3-89bd-767b9c7e8f94\") " pod="openstack/barbican-worker-679666fcbc-r7dpm" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.799700 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67805ea2-2459-4c00-9b18-54d3e60d1281-config-data\") pod \"barbican-keystone-listener-855f768b54-b52dt\" (UID: \"67805ea2-2459-4c00-9b18-54d3e60d1281\") " pod="openstack/barbican-keystone-listener-855f768b54-b52dt" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.799737 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67805ea2-2459-4c00-9b18-54d3e60d1281-config-data-custom\") pod \"barbican-keystone-listener-855f768b54-b52dt\" (UID: \"67805ea2-2459-4c00-9b18-54d3e60d1281\") " pod="openstack/barbican-keystone-listener-855f768b54-b52dt" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.799768 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/835bf15c-547d-4bb3-89bd-767b9c7e8f94-combined-ca-bundle\") pod \"barbican-worker-679666fcbc-r7dpm\" (UID: \"835bf15c-547d-4bb3-89bd-767b9c7e8f94\") " pod="openstack/barbican-worker-679666fcbc-r7dpm" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.799844 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/835bf15c-547d-4bb3-89bd-767b9c7e8f94-config-data-custom\") pod \"barbican-worker-679666fcbc-r7dpm\" (UID: \"835bf15c-547d-4bb3-89bd-767b9c7e8f94\") " pod="openstack/barbican-worker-679666fcbc-r7dpm" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.799893 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/835bf15c-547d-4bb3-89bd-767b9c7e8f94-config-data\") pod \"barbican-worker-679666fcbc-r7dpm\" (UID: \"835bf15c-547d-4bb3-89bd-767b9c7e8f94\") " pod="openstack/barbican-worker-679666fcbc-r7dpm" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.799914 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x48mv\" (UniqueName: \"kubernetes.io/projected/835bf15c-547d-4bb3-89bd-767b9c7e8f94-kube-api-access-x48mv\") pod \"barbican-worker-679666fcbc-r7dpm\" (UID: \"835bf15c-547d-4bb3-89bd-767b9c7e8f94\") " pod="openstack/barbican-worker-679666fcbc-r7dpm" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.800003 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67805ea2-2459-4c00-9b18-54d3e60d1281-combined-ca-bundle\") pod \"barbican-keystone-listener-855f768b54-b52dt\" (UID: \"67805ea2-2459-4c00-9b18-54d3e60d1281\") " pod="openstack/barbican-keystone-listener-855f768b54-b52dt" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.800052 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67805ea2-2459-4c00-9b18-54d3e60d1281-logs\") pod \"barbican-keystone-listener-855f768b54-b52dt\" (UID: \"67805ea2-2459-4c00-9b18-54d3e60d1281\") " pod="openstack/barbican-keystone-listener-855f768b54-b52dt" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.800119 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7nmr\" (UniqueName: \"kubernetes.io/projected/67805ea2-2459-4c00-9b18-54d3e60d1281-kube-api-access-f7nmr\") pod \"barbican-keystone-listener-855f768b54-b52dt\" (UID: \"67805ea2-2459-4c00-9b18-54d3e60d1281\") " pod="openstack/barbican-keystone-listener-855f768b54-b52dt" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.821173 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84555f7879-rcsn2"] Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.824397 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.857896 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84555f7879-rcsn2"] Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.901825 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-config\") pod \"dnsmasq-dns-84555f7879-rcsn2\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.901900 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67805ea2-2459-4c00-9b18-54d3e60d1281-combined-ca-bundle\") pod \"barbican-keystone-listener-855f768b54-b52dt\" (UID: \"67805ea2-2459-4c00-9b18-54d3e60d1281\") " pod="openstack/barbican-keystone-listener-855f768b54-b52dt" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.901942 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67805ea2-2459-4c00-9b18-54d3e60d1281-logs\") pod \"barbican-keystone-listener-855f768b54-b52dt\" (UID: \"67805ea2-2459-4c00-9b18-54d3e60d1281\") " pod="openstack/barbican-keystone-listener-855f768b54-b52dt" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.901995 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7nmr\" (UniqueName: \"kubernetes.io/projected/67805ea2-2459-4c00-9b18-54d3e60d1281-kube-api-access-f7nmr\") pod \"barbican-keystone-listener-855f768b54-b52dt\" (UID: \"67805ea2-2459-4c00-9b18-54d3e60d1281\") " pod="openstack/barbican-keystone-listener-855f768b54-b52dt" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.902063 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtl7j\" (UniqueName: \"kubernetes.io/projected/029bd0e5-0d9e-4734-a15c-abae3e06df04-kube-api-access-qtl7j\") pod \"dnsmasq-dns-84555f7879-rcsn2\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.902088 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-ovsdbserver-sb\") pod \"dnsmasq-dns-84555f7879-rcsn2\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.902126 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/835bf15c-547d-4bb3-89bd-767b9c7e8f94-logs\") pod \"barbican-worker-679666fcbc-r7dpm\" (UID: \"835bf15c-547d-4bb3-89bd-767b9c7e8f94\") " pod="openstack/barbican-worker-679666fcbc-r7dpm" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.902158 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67805ea2-2459-4c00-9b18-54d3e60d1281-config-data\") pod \"barbican-keystone-listener-855f768b54-b52dt\" (UID: \"67805ea2-2459-4c00-9b18-54d3e60d1281\") " pod="openstack/barbican-keystone-listener-855f768b54-b52dt" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.902211 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67805ea2-2459-4c00-9b18-54d3e60d1281-config-data-custom\") pod \"barbican-keystone-listener-855f768b54-b52dt\" (UID: \"67805ea2-2459-4c00-9b18-54d3e60d1281\") " pod="openstack/barbican-keystone-listener-855f768b54-b52dt" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.902253 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-dns-svc\") pod \"dnsmasq-dns-84555f7879-rcsn2\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.902291 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/835bf15c-547d-4bb3-89bd-767b9c7e8f94-combined-ca-bundle\") pod \"barbican-worker-679666fcbc-r7dpm\" (UID: \"835bf15c-547d-4bb3-89bd-767b9c7e8f94\") " pod="openstack/barbican-worker-679666fcbc-r7dpm" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.902319 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/835bf15c-547d-4bb3-89bd-767b9c7e8f94-config-data-custom\") pod \"barbican-worker-679666fcbc-r7dpm\" (UID: \"835bf15c-547d-4bb3-89bd-767b9c7e8f94\") " pod="openstack/barbican-worker-679666fcbc-r7dpm" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.902348 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/835bf15c-547d-4bb3-89bd-767b9c7e8f94-config-data\") pod \"barbican-worker-679666fcbc-r7dpm\" (UID: \"835bf15c-547d-4bb3-89bd-767b9c7e8f94\") " pod="openstack/barbican-worker-679666fcbc-r7dpm" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.902370 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x48mv\" (UniqueName: \"kubernetes.io/projected/835bf15c-547d-4bb3-89bd-767b9c7e8f94-kube-api-access-x48mv\") pod \"barbican-worker-679666fcbc-r7dpm\" (UID: \"835bf15c-547d-4bb3-89bd-767b9c7e8f94\") " pod="openstack/barbican-worker-679666fcbc-r7dpm" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.902397 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-ovsdbserver-nb\") pod \"dnsmasq-dns-84555f7879-rcsn2\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.911387 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67805ea2-2459-4c00-9b18-54d3e60d1281-config-data-custom\") pod \"barbican-keystone-listener-855f768b54-b52dt\" (UID: \"67805ea2-2459-4c00-9b18-54d3e60d1281\") " pod="openstack/barbican-keystone-listener-855f768b54-b52dt" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.912235 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67805ea2-2459-4c00-9b18-54d3e60d1281-logs\") pod \"barbican-keystone-listener-855f768b54-b52dt\" (UID: \"67805ea2-2459-4c00-9b18-54d3e60d1281\") " pod="openstack/barbican-keystone-listener-855f768b54-b52dt" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.912741 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/835bf15c-547d-4bb3-89bd-767b9c7e8f94-combined-ca-bundle\") pod \"barbican-worker-679666fcbc-r7dpm\" (UID: \"835bf15c-547d-4bb3-89bd-767b9c7e8f94\") " pod="openstack/barbican-worker-679666fcbc-r7dpm" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.912952 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/835bf15c-547d-4bb3-89bd-767b9c7e8f94-logs\") pod \"barbican-worker-679666fcbc-r7dpm\" (UID: \"835bf15c-547d-4bb3-89bd-767b9c7e8f94\") " pod="openstack/barbican-worker-679666fcbc-r7dpm" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.913511 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67805ea2-2459-4c00-9b18-54d3e60d1281-combined-ca-bundle\") pod \"barbican-keystone-listener-855f768b54-b52dt\" (UID: \"67805ea2-2459-4c00-9b18-54d3e60d1281\") " pod="openstack/barbican-keystone-listener-855f768b54-b52dt" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.915918 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67805ea2-2459-4c00-9b18-54d3e60d1281-config-data\") pod \"barbican-keystone-listener-855f768b54-b52dt\" (UID: \"67805ea2-2459-4c00-9b18-54d3e60d1281\") " pod="openstack/barbican-keystone-listener-855f768b54-b52dt" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.916211 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/835bf15c-547d-4bb3-89bd-767b9c7e8f94-config-data\") pod \"barbican-worker-679666fcbc-r7dpm\" (UID: \"835bf15c-547d-4bb3-89bd-767b9c7e8f94\") " pod="openstack/barbican-worker-679666fcbc-r7dpm" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.932049 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-b77bff76-4gtq2"] Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.932268 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/835bf15c-547d-4bb3-89bd-767b9c7e8f94-config-data-custom\") pod \"barbican-worker-679666fcbc-r7dpm\" (UID: \"835bf15c-547d-4bb3-89bd-767b9c7e8f94\") " pod="openstack/barbican-worker-679666fcbc-r7dpm" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.933534 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.939814 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-b77bff76-4gtq2"] Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.941794 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.946066 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x48mv\" (UniqueName: \"kubernetes.io/projected/835bf15c-547d-4bb3-89bd-767b9c7e8f94-kube-api-access-x48mv\") pod \"barbican-worker-679666fcbc-r7dpm\" (UID: \"835bf15c-547d-4bb3-89bd-767b9c7e8f94\") " pod="openstack/barbican-worker-679666fcbc-r7dpm" Dec 01 16:09:53 crc kubenswrapper[4810]: I1201 16:09:53.946853 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7nmr\" (UniqueName: \"kubernetes.io/projected/67805ea2-2459-4c00-9b18-54d3e60d1281-kube-api-access-f7nmr\") pod \"barbican-keystone-listener-855f768b54-b52dt\" (UID: \"67805ea2-2459-4c00-9b18-54d3e60d1281\") " pod="openstack/barbican-keystone-listener-855f768b54-b52dt" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.003134 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-config\") pod \"dnsmasq-dns-84555f7879-rcsn2\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.003209 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa152182-eb48-4241-9d80-73f2c5fc5310-combined-ca-bundle\") pod \"barbican-api-b77bff76-4gtq2\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.003233 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa152182-eb48-4241-9d80-73f2c5fc5310-config-data-custom\") pod \"barbican-api-b77bff76-4gtq2\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.003279 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvf5f\" (UniqueName: \"kubernetes.io/projected/fa152182-eb48-4241-9d80-73f2c5fc5310-kube-api-access-gvf5f\") pod \"barbican-api-b77bff76-4gtq2\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.003319 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtl7j\" (UniqueName: \"kubernetes.io/projected/029bd0e5-0d9e-4734-a15c-abae3e06df04-kube-api-access-qtl7j\") pod \"dnsmasq-dns-84555f7879-rcsn2\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.003355 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-ovsdbserver-sb\") pod \"dnsmasq-dns-84555f7879-rcsn2\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.003381 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa152182-eb48-4241-9d80-73f2c5fc5310-logs\") pod \"barbican-api-b77bff76-4gtq2\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.003403 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa152182-eb48-4241-9d80-73f2c5fc5310-config-data\") pod \"barbican-api-b77bff76-4gtq2\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.003453 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-dns-svc\") pod \"dnsmasq-dns-84555f7879-rcsn2\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.003510 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-ovsdbserver-nb\") pod \"dnsmasq-dns-84555f7879-rcsn2\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.004339 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-ovsdbserver-sb\") pod \"dnsmasq-dns-84555f7879-rcsn2\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.004701 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-dns-svc\") pod \"dnsmasq-dns-84555f7879-rcsn2\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.004785 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-ovsdbserver-nb\") pod \"dnsmasq-dns-84555f7879-rcsn2\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.005858 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-config\") pod \"dnsmasq-dns-84555f7879-rcsn2\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.019062 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtl7j\" (UniqueName: \"kubernetes.io/projected/029bd0e5-0d9e-4734-a15c-abae3e06df04-kube-api-access-qtl7j\") pod \"dnsmasq-dns-84555f7879-rcsn2\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.050110 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-855f768b54-b52dt" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.069715 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-679666fcbc-r7dpm" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.105598 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa152182-eb48-4241-9d80-73f2c5fc5310-combined-ca-bundle\") pod \"barbican-api-b77bff76-4gtq2\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.105644 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa152182-eb48-4241-9d80-73f2c5fc5310-config-data-custom\") pod \"barbican-api-b77bff76-4gtq2\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.105693 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvf5f\" (UniqueName: \"kubernetes.io/projected/fa152182-eb48-4241-9d80-73f2c5fc5310-kube-api-access-gvf5f\") pod \"barbican-api-b77bff76-4gtq2\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.105740 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa152182-eb48-4241-9d80-73f2c5fc5310-logs\") pod \"barbican-api-b77bff76-4gtq2\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.105770 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa152182-eb48-4241-9d80-73f2c5fc5310-config-data\") pod \"barbican-api-b77bff76-4gtq2\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.109310 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa152182-eb48-4241-9d80-73f2c5fc5310-combined-ca-bundle\") pod \"barbican-api-b77bff76-4gtq2\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.109654 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa152182-eb48-4241-9d80-73f2c5fc5310-logs\") pod \"barbican-api-b77bff76-4gtq2\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.113442 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa152182-eb48-4241-9d80-73f2c5fc5310-config-data-custom\") pod \"barbican-api-b77bff76-4gtq2\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.117124 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa152182-eb48-4241-9d80-73f2c5fc5310-config-data\") pod \"barbican-api-b77bff76-4gtq2\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.124423 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvf5f\" (UniqueName: \"kubernetes.io/projected/fa152182-eb48-4241-9d80-73f2c5fc5310-kube-api-access-gvf5f\") pod \"barbican-api-b77bff76-4gtq2\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.164846 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.316357 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.573940 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-855f768b54-b52dt"] Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.673244 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-679666fcbc-r7dpm"] Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.683384 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84555f7879-rcsn2"] Dec 01 16:09:54 crc kubenswrapper[4810]: I1201 16:09:54.848156 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-b77bff76-4gtq2"] Dec 01 16:09:55 crc kubenswrapper[4810]: I1201 16:09:55.498796 4810 generic.go:334] "Generic (PLEG): container finished" podID="029bd0e5-0d9e-4734-a15c-abae3e06df04" containerID="fd717eb5e7d444f3f5b8f3205ff674410d870a14f32256415fbe7def4c969cb4" exitCode=0 Dec 01 16:09:55 crc kubenswrapper[4810]: I1201 16:09:55.499033 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84555f7879-rcsn2" event={"ID":"029bd0e5-0d9e-4734-a15c-abae3e06df04","Type":"ContainerDied","Data":"fd717eb5e7d444f3f5b8f3205ff674410d870a14f32256415fbe7def4c969cb4"} Dec 01 16:09:55 crc kubenswrapper[4810]: I1201 16:09:55.499238 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84555f7879-rcsn2" event={"ID":"029bd0e5-0d9e-4734-a15c-abae3e06df04","Type":"ContainerStarted","Data":"40c8d9e9d78b27fea97a8ce51ce522b91a958e0731a034bdab45ae6f799750b2"} Dec 01 16:09:55 crc kubenswrapper[4810]: I1201 16:09:55.501494 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-855f768b54-b52dt" event={"ID":"67805ea2-2459-4c00-9b18-54d3e60d1281","Type":"ContainerStarted","Data":"1b6e809863d8e0d74081c778492c64bf7710a5eaba127b2ae0d37720ed4b3dec"} Dec 01 16:09:55 crc kubenswrapper[4810]: I1201 16:09:55.501519 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-855f768b54-b52dt" event={"ID":"67805ea2-2459-4c00-9b18-54d3e60d1281","Type":"ContainerStarted","Data":"2572e1d3b5f68d9dbc7fbc5e11381e0da723e3fdb97c6bc0a8803ff8117c2000"} Dec 01 16:09:55 crc kubenswrapper[4810]: I1201 16:09:55.501529 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-855f768b54-b52dt" event={"ID":"67805ea2-2459-4c00-9b18-54d3e60d1281","Type":"ContainerStarted","Data":"d645e23f6d8ce88092befbf615be016f1df4b829656c19365a7aea13ab4a2dd7"} Dec 01 16:09:55 crc kubenswrapper[4810]: I1201 16:09:55.506560 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b77bff76-4gtq2" event={"ID":"fa152182-eb48-4241-9d80-73f2c5fc5310","Type":"ContainerStarted","Data":"bbda8149a7db5b2157b310da3506a0525a2e2aed87b3b5d64c5c8d7e2d3f9757"} Dec 01 16:09:55 crc kubenswrapper[4810]: I1201 16:09:55.506610 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b77bff76-4gtq2" event={"ID":"fa152182-eb48-4241-9d80-73f2c5fc5310","Type":"ContainerStarted","Data":"5f1d8eaa3d64ad44d9bd25eb7824637f056506dd2b88b09e4774d4cf0605d56f"} Dec 01 16:09:55 crc kubenswrapper[4810]: I1201 16:09:55.506623 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b77bff76-4gtq2" event={"ID":"fa152182-eb48-4241-9d80-73f2c5fc5310","Type":"ContainerStarted","Data":"12377c88a020aa18216c857587015bdbe410bf6dc428afdaf1a01abed658f692"} Dec 01 16:09:55 crc kubenswrapper[4810]: I1201 16:09:55.506650 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:55 crc kubenswrapper[4810]: I1201 16:09:55.506670 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:09:55 crc kubenswrapper[4810]: I1201 16:09:55.508127 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-679666fcbc-r7dpm" event={"ID":"835bf15c-547d-4bb3-89bd-767b9c7e8f94","Type":"ContainerStarted","Data":"d739ffabed851ee90b02ac54791723c73b80c764ffaa06a19c718a5b9dc8a8ca"} Dec 01 16:09:55 crc kubenswrapper[4810]: I1201 16:09:55.508151 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-679666fcbc-r7dpm" event={"ID":"835bf15c-547d-4bb3-89bd-767b9c7e8f94","Type":"ContainerStarted","Data":"7fde830639063263c0dc89155ea9cf19ed3dd80710e2d0b8108389b490101423"} Dec 01 16:09:55 crc kubenswrapper[4810]: I1201 16:09:55.508161 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-679666fcbc-r7dpm" event={"ID":"835bf15c-547d-4bb3-89bd-767b9c7e8f94","Type":"ContainerStarted","Data":"d7b3161d289926224957e23532978a0eff740329269f2e1b452db80d8e0b5bc0"} Dec 01 16:09:55 crc kubenswrapper[4810]: I1201 16:09:55.563049 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-b77bff76-4gtq2" podStartSLOduration=2.563016528 podStartE2EDuration="2.563016528s" podCreationTimestamp="2025-12-01 16:09:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:09:55.561521037 +0000 UTC m=+5761.325030660" watchObservedRunningTime="2025-12-01 16:09:55.563016528 +0000 UTC m=+5761.326526131" Dec 01 16:09:55 crc kubenswrapper[4810]: I1201 16:09:55.587696 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-855f768b54-b52dt" podStartSLOduration=2.587670487 podStartE2EDuration="2.587670487s" podCreationTimestamp="2025-12-01 16:09:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:09:55.581940481 +0000 UTC m=+5761.345450104" watchObservedRunningTime="2025-12-01 16:09:55.587670487 +0000 UTC m=+5761.351180090" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.219996 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-679666fcbc-r7dpm" podStartSLOduration=3.2199778 podStartE2EDuration="3.2199778s" podCreationTimestamp="2025-12-01 16:09:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:09:55.628414354 +0000 UTC m=+5761.391923957" watchObservedRunningTime="2025-12-01 16:09:56.2199778 +0000 UTC m=+5761.983487403" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.222965 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5fdf9c97c6-8mwkg"] Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.224395 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: W1201 16:09:56.226506 4810 reflector.go:561] object-"openstack"/"cert-barbican-public-svc": failed to list *v1.Secret: secrets "cert-barbican-public-svc" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Dec 01 16:09:56 crc kubenswrapper[4810]: E1201 16:09:56.226605 4810 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"cert-barbican-public-svc\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cert-barbican-public-svc\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.227157 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.243445 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5fdf9c97c6-8mwkg"] Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.350018 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b61ad6b-ee0b-4236-883c-139c4a499987-internal-tls-certs\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.350448 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b61ad6b-ee0b-4236-883c-139c4a499987-config-data-custom\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.350500 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b61ad6b-ee0b-4236-883c-139c4a499987-public-tls-certs\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.350530 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b61ad6b-ee0b-4236-883c-139c4a499987-combined-ca-bundle\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.350571 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b61ad6b-ee0b-4236-883c-139c4a499987-config-data\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.350593 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b61ad6b-ee0b-4236-883c-139c4a499987-logs\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.350612 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn99r\" (UniqueName: \"kubernetes.io/projected/9b61ad6b-ee0b-4236-883c-139c4a499987-kube-api-access-hn99r\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.452027 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b61ad6b-ee0b-4236-883c-139c4a499987-internal-tls-certs\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.452104 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b61ad6b-ee0b-4236-883c-139c4a499987-config-data-custom\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.452124 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b61ad6b-ee0b-4236-883c-139c4a499987-public-tls-certs\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.452353 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b61ad6b-ee0b-4236-883c-139c4a499987-combined-ca-bundle\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.452399 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b61ad6b-ee0b-4236-883c-139c4a499987-config-data\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.452424 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b61ad6b-ee0b-4236-883c-139c4a499987-logs\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.452440 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn99r\" (UniqueName: \"kubernetes.io/projected/9b61ad6b-ee0b-4236-883c-139c4a499987-kube-api-access-hn99r\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.453505 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b61ad6b-ee0b-4236-883c-139c4a499987-logs\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.460087 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b61ad6b-ee0b-4236-883c-139c4a499987-combined-ca-bundle\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.460559 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b61ad6b-ee0b-4236-883c-139c4a499987-config-data-custom\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.460999 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b61ad6b-ee0b-4236-883c-139c4a499987-internal-tls-certs\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.474791 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b61ad6b-ee0b-4236-883c-139c4a499987-config-data\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.486191 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn99r\" (UniqueName: \"kubernetes.io/projected/9b61ad6b-ee0b-4236-883c-139c4a499987-kube-api-access-hn99r\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.524604 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84555f7879-rcsn2" event={"ID":"029bd0e5-0d9e-4734-a15c-abae3e06df04","Type":"ContainerStarted","Data":"ec8fce0ff6a4639f090f98f71f8bf824cdbb75960b862dcba28175c8688023a3"} Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.525595 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:09:56 crc kubenswrapper[4810]: I1201 16:09:56.562718 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84555f7879-rcsn2" podStartSLOduration=3.562697584 podStartE2EDuration="3.562697584s" podCreationTimestamp="2025-12-01 16:09:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:09:56.556343581 +0000 UTC m=+5762.319853184" watchObservedRunningTime="2025-12-01 16:09:56.562697584 +0000 UTC m=+5762.326207187" Dec 01 16:09:57 crc kubenswrapper[4810]: I1201 16:09:57.451619 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 01 16:09:57 crc kubenswrapper[4810]: I1201 16:09:57.456303 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b61ad6b-ee0b-4236-883c-139c4a499987-public-tls-certs\") pod \"barbican-api-5fdf9c97c6-8mwkg\" (UID: \"9b61ad6b-ee0b-4236-883c-139c4a499987\") " pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:57 crc kubenswrapper[4810]: I1201 16:09:57.745832 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:58 crc kubenswrapper[4810]: I1201 16:09:58.166746 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5fdf9c97c6-8mwkg"] Dec 01 16:09:58 crc kubenswrapper[4810]: I1201 16:09:58.544523 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fdf9c97c6-8mwkg" event={"ID":"9b61ad6b-ee0b-4236-883c-139c4a499987","Type":"ContainerStarted","Data":"476101f2fbbc4a9d745202d14bc6a3274053380f9d1ecbf5a2173c8bf1f8b02c"} Dec 01 16:09:58 crc kubenswrapper[4810]: I1201 16:09:58.544567 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fdf9c97c6-8mwkg" event={"ID":"9b61ad6b-ee0b-4236-883c-139c4a499987","Type":"ContainerStarted","Data":"628ffacad74a99683155b2b24c7c6f3a98b55aec6a18e51c2bcc6f6d2e7ee969"} Dec 01 16:09:58 crc kubenswrapper[4810]: I1201 16:09:58.544580 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fdf9c97c6-8mwkg" event={"ID":"9b61ad6b-ee0b-4236-883c-139c4a499987","Type":"ContainerStarted","Data":"31f547326529a066b1c91200037e3187cb9a9760dd73dc0b5cf9a493e10c9bcc"} Dec 01 16:09:58 crc kubenswrapper[4810]: I1201 16:09:58.545076 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:58 crc kubenswrapper[4810]: I1201 16:09:58.545106 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:09:58 crc kubenswrapper[4810]: I1201 16:09:58.572748 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5fdf9c97c6-8mwkg" podStartSLOduration=2.572722115 podStartE2EDuration="2.572722115s" podCreationTimestamp="2025-12-01 16:09:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:09:58.567069832 +0000 UTC m=+5764.330579435" watchObservedRunningTime="2025-12-01 16:09:58.572722115 +0000 UTC m=+5764.336231728" Dec 01 16:10:01 crc kubenswrapper[4810]: I1201 16:10:01.755730 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-b77bff76-4gtq2" podUID="fa152182-eb48-4241-9d80-73f2c5fc5310" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 16:10:02 crc kubenswrapper[4810]: I1201 16:10:02.972719 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:10:02 crc kubenswrapper[4810]: I1201 16:10:02.973062 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:10:04 crc kubenswrapper[4810]: I1201 16:10:04.169344 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:10:04 crc kubenswrapper[4810]: I1201 16:10:04.241863 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64578c6499-mvcxj"] Dec 01 16:10:04 crc kubenswrapper[4810]: I1201 16:10:04.242120 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64578c6499-mvcxj" podUID="a2b19a52-ce05-49a9-bebd-b3e8a3c9363e" containerName="dnsmasq-dns" containerID="cri-o://e888a8bce60f9a4907d307433c7ad9b53378dd0ff8bb280328d9ef5bc7d41566" gracePeriod=10 Dec 01 16:10:04 crc kubenswrapper[4810]: I1201 16:10:04.739338 4810 generic.go:334] "Generic (PLEG): container finished" podID="a2b19a52-ce05-49a9-bebd-b3e8a3c9363e" containerID="e888a8bce60f9a4907d307433c7ad9b53378dd0ff8bb280328d9ef5bc7d41566" exitCode=0 Dec 01 16:10:04 crc kubenswrapper[4810]: I1201 16:10:04.739402 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64578c6499-mvcxj" event={"ID":"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e","Type":"ContainerDied","Data":"e888a8bce60f9a4907d307433c7ad9b53378dd0ff8bb280328d9ef5bc7d41566"} Dec 01 16:10:04 crc kubenswrapper[4810]: I1201 16:10:04.739775 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64578c6499-mvcxj" event={"ID":"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e","Type":"ContainerDied","Data":"6976574d391688ab9b2088891b29e0513bbb2ec2f539b4b7a0c65535b685c87f"} Dec 01 16:10:04 crc kubenswrapper[4810]: I1201 16:10:04.739794 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6976574d391688ab9b2088891b29e0513bbb2ec2f539b4b7a0c65535b685c87f" Dec 01 16:10:04 crc kubenswrapper[4810]: I1201 16:10:04.829726 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:10:04 crc kubenswrapper[4810]: I1201 16:10:04.979920 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-dns-svc\") pod \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " Dec 01 16:10:04 crc kubenswrapper[4810]: I1201 16:10:04.980412 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-config\") pod \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " Dec 01 16:10:04 crc kubenswrapper[4810]: I1201 16:10:04.980456 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r69qd\" (UniqueName: \"kubernetes.io/projected/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-kube-api-access-r69qd\") pod \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " Dec 01 16:10:04 crc kubenswrapper[4810]: I1201 16:10:04.980604 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-ovsdbserver-nb\") pod \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " Dec 01 16:10:04 crc kubenswrapper[4810]: I1201 16:10:04.980647 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-ovsdbserver-sb\") pod \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\" (UID: \"a2b19a52-ce05-49a9-bebd-b3e8a3c9363e\") " Dec 01 16:10:05 crc kubenswrapper[4810]: I1201 16:10:05.015810 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-kube-api-access-r69qd" (OuterVolumeSpecName: "kube-api-access-r69qd") pod "a2b19a52-ce05-49a9-bebd-b3e8a3c9363e" (UID: "a2b19a52-ce05-49a9-bebd-b3e8a3c9363e"). InnerVolumeSpecName "kube-api-access-r69qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:10:05 crc kubenswrapper[4810]: I1201 16:10:05.050512 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a2b19a52-ce05-49a9-bebd-b3e8a3c9363e" (UID: "a2b19a52-ce05-49a9-bebd-b3e8a3c9363e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:10:05 crc kubenswrapper[4810]: I1201 16:10:05.058938 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-config" (OuterVolumeSpecName: "config") pod "a2b19a52-ce05-49a9-bebd-b3e8a3c9363e" (UID: "a2b19a52-ce05-49a9-bebd-b3e8a3c9363e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:10:05 crc kubenswrapper[4810]: I1201 16:10:05.067022 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a2b19a52-ce05-49a9-bebd-b3e8a3c9363e" (UID: "a2b19a52-ce05-49a9-bebd-b3e8a3c9363e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:10:05 crc kubenswrapper[4810]: I1201 16:10:05.083107 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:05 crc kubenswrapper[4810]: I1201 16:10:05.083185 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r69qd\" (UniqueName: \"kubernetes.io/projected/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-kube-api-access-r69qd\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:05 crc kubenswrapper[4810]: I1201 16:10:05.083208 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:05 crc kubenswrapper[4810]: I1201 16:10:05.083225 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:05 crc kubenswrapper[4810]: I1201 16:10:05.094536 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a2b19a52-ce05-49a9-bebd-b3e8a3c9363e" (UID: "a2b19a52-ce05-49a9-bebd-b3e8a3c9363e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:10:05 crc kubenswrapper[4810]: I1201 16:10:05.184775 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:05 crc kubenswrapper[4810]: I1201 16:10:05.747181 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64578c6499-mvcxj" Dec 01 16:10:05 crc kubenswrapper[4810]: I1201 16:10:05.783845 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64578c6499-mvcxj"] Dec 01 16:10:05 crc kubenswrapper[4810]: I1201 16:10:05.793069 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64578c6499-mvcxj"] Dec 01 16:10:05 crc kubenswrapper[4810]: I1201 16:10:05.824583 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:10:06 crc kubenswrapper[4810]: I1201 16:10:06.043590 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:10:06 crc kubenswrapper[4810]: I1201 16:10:06.507788 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2b19a52-ce05-49a9-bebd-b3e8a3c9363e" path="/var/lib/kubelet/pods/a2b19a52-ce05-49a9-bebd-b3e8a3c9363e/volumes" Dec 01 16:10:09 crc kubenswrapper[4810]: I1201 16:10:09.132841 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:10:09 crc kubenswrapper[4810]: I1201 16:10:09.154645 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5fdf9c97c6-8mwkg" Dec 01 16:10:09 crc kubenswrapper[4810]: I1201 16:10:09.223408 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-b77bff76-4gtq2"] Dec 01 16:10:09 crc kubenswrapper[4810]: I1201 16:10:09.223729 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-b77bff76-4gtq2" podUID="fa152182-eb48-4241-9d80-73f2c5fc5310" containerName="barbican-api-log" containerID="cri-o://5f1d8eaa3d64ad44d9bd25eb7824637f056506dd2b88b09e4774d4cf0605d56f" gracePeriod=30 Dec 01 16:10:09 crc kubenswrapper[4810]: I1201 16:10:09.223809 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-b77bff76-4gtq2" podUID="fa152182-eb48-4241-9d80-73f2c5fc5310" containerName="barbican-api" containerID="cri-o://bbda8149a7db5b2157b310da3506a0525a2e2aed87b3b5d64c5c8d7e2d3f9757" gracePeriod=30 Dec 01 16:10:09 crc kubenswrapper[4810]: I1201 16:10:09.782597 4810 generic.go:334] "Generic (PLEG): container finished" podID="fa152182-eb48-4241-9d80-73f2c5fc5310" containerID="5f1d8eaa3d64ad44d9bd25eb7824637f056506dd2b88b09e4774d4cf0605d56f" exitCode=143 Dec 01 16:10:09 crc kubenswrapper[4810]: I1201 16:10:09.782816 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b77bff76-4gtq2" event={"ID":"fa152182-eb48-4241-9d80-73f2c5fc5310","Type":"ContainerDied","Data":"5f1d8eaa3d64ad44d9bd25eb7824637f056506dd2b88b09e4774d4cf0605d56f"} Dec 01 16:10:12 crc kubenswrapper[4810]: I1201 16:10:12.384573 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-b77bff76-4gtq2" podUID="fa152182-eb48-4241-9d80-73f2c5fc5310" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.1.32:9311/healthcheck\": read tcp 10.217.0.2:48608->10.217.1.32:9311: read: connection reset by peer" Dec 01 16:10:12 crc kubenswrapper[4810]: I1201 16:10:12.384545 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-b77bff76-4gtq2" podUID="fa152182-eb48-4241-9d80-73f2c5fc5310" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.1.32:9311/healthcheck\": read tcp 10.217.0.2:48596->10.217.1.32:9311: read: connection reset by peer" Dec 01 16:10:12 crc kubenswrapper[4810]: I1201 16:10:12.807105 4810 generic.go:334] "Generic (PLEG): container finished" podID="fa152182-eb48-4241-9d80-73f2c5fc5310" containerID="bbda8149a7db5b2157b310da3506a0525a2e2aed87b3b5d64c5c8d7e2d3f9757" exitCode=0 Dec 01 16:10:12 crc kubenswrapper[4810]: I1201 16:10:12.807332 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b77bff76-4gtq2" event={"ID":"fa152182-eb48-4241-9d80-73f2c5fc5310","Type":"ContainerDied","Data":"bbda8149a7db5b2157b310da3506a0525a2e2aed87b3b5d64c5c8d7e2d3f9757"} Dec 01 16:10:12 crc kubenswrapper[4810]: I1201 16:10:12.807357 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b77bff76-4gtq2" event={"ID":"fa152182-eb48-4241-9d80-73f2c5fc5310","Type":"ContainerDied","Data":"12377c88a020aa18216c857587015bdbe410bf6dc428afdaf1a01abed658f692"} Dec 01 16:10:12 crc kubenswrapper[4810]: I1201 16:10:12.807368 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12377c88a020aa18216c857587015bdbe410bf6dc428afdaf1a01abed658f692" Dec 01 16:10:12 crc kubenswrapper[4810]: I1201 16:10:12.820347 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:10:12 crc kubenswrapper[4810]: I1201 16:10:12.935639 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa152182-eb48-4241-9d80-73f2c5fc5310-logs\") pod \"fa152182-eb48-4241-9d80-73f2c5fc5310\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " Dec 01 16:10:12 crc kubenswrapper[4810]: I1201 16:10:12.935762 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa152182-eb48-4241-9d80-73f2c5fc5310-combined-ca-bundle\") pod \"fa152182-eb48-4241-9d80-73f2c5fc5310\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " Dec 01 16:10:12 crc kubenswrapper[4810]: I1201 16:10:12.935798 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa152182-eb48-4241-9d80-73f2c5fc5310-config-data\") pod \"fa152182-eb48-4241-9d80-73f2c5fc5310\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " Dec 01 16:10:12 crc kubenswrapper[4810]: I1201 16:10:12.935871 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvf5f\" (UniqueName: \"kubernetes.io/projected/fa152182-eb48-4241-9d80-73f2c5fc5310-kube-api-access-gvf5f\") pod \"fa152182-eb48-4241-9d80-73f2c5fc5310\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " Dec 01 16:10:12 crc kubenswrapper[4810]: I1201 16:10:12.935912 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa152182-eb48-4241-9d80-73f2c5fc5310-config-data-custom\") pod \"fa152182-eb48-4241-9d80-73f2c5fc5310\" (UID: \"fa152182-eb48-4241-9d80-73f2c5fc5310\") " Dec 01 16:10:12 crc kubenswrapper[4810]: I1201 16:10:12.937376 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa152182-eb48-4241-9d80-73f2c5fc5310-logs" (OuterVolumeSpecName: "logs") pod "fa152182-eb48-4241-9d80-73f2c5fc5310" (UID: "fa152182-eb48-4241-9d80-73f2c5fc5310"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:10:12 crc kubenswrapper[4810]: I1201 16:10:12.943054 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa152182-eb48-4241-9d80-73f2c5fc5310-kube-api-access-gvf5f" (OuterVolumeSpecName: "kube-api-access-gvf5f") pod "fa152182-eb48-4241-9d80-73f2c5fc5310" (UID: "fa152182-eb48-4241-9d80-73f2c5fc5310"). InnerVolumeSpecName "kube-api-access-gvf5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:10:12 crc kubenswrapper[4810]: I1201 16:10:12.945434 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa152182-eb48-4241-9d80-73f2c5fc5310-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fa152182-eb48-4241-9d80-73f2c5fc5310" (UID: "fa152182-eb48-4241-9d80-73f2c5fc5310"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:10:12 crc kubenswrapper[4810]: I1201 16:10:12.973962 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa152182-eb48-4241-9d80-73f2c5fc5310-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fa152182-eb48-4241-9d80-73f2c5fc5310" (UID: "fa152182-eb48-4241-9d80-73f2c5fc5310"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:10:12 crc kubenswrapper[4810]: I1201 16:10:12.991095 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa152182-eb48-4241-9d80-73f2c5fc5310-config-data" (OuterVolumeSpecName: "config-data") pod "fa152182-eb48-4241-9d80-73f2c5fc5310" (UID: "fa152182-eb48-4241-9d80-73f2c5fc5310"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:10:13 crc kubenswrapper[4810]: I1201 16:10:13.038068 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa152182-eb48-4241-9d80-73f2c5fc5310-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:13 crc kubenswrapper[4810]: I1201 16:10:13.038112 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa152182-eb48-4241-9d80-73f2c5fc5310-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:13 crc kubenswrapper[4810]: I1201 16:10:13.038159 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa152182-eb48-4241-9d80-73f2c5fc5310-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:13 crc kubenswrapper[4810]: I1201 16:10:13.038172 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvf5f\" (UniqueName: \"kubernetes.io/projected/fa152182-eb48-4241-9d80-73f2c5fc5310-kube-api-access-gvf5f\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:13 crc kubenswrapper[4810]: I1201 16:10:13.038185 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa152182-eb48-4241-9d80-73f2c5fc5310-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:13 crc kubenswrapper[4810]: I1201 16:10:13.815266 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-b77bff76-4gtq2" Dec 01 16:10:13 crc kubenswrapper[4810]: I1201 16:10:13.862179 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-b77bff76-4gtq2"] Dec 01 16:10:13 crc kubenswrapper[4810]: I1201 16:10:13.869534 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-b77bff76-4gtq2"] Dec 01 16:10:14 crc kubenswrapper[4810]: I1201 16:10:14.503875 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa152182-eb48-4241-9d80-73f2c5fc5310" path="/var/lib/kubelet/pods/fa152182-eb48-4241-9d80-73f2c5fc5310/volumes" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.344565 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-2611-account-create-update-b2jpl"] Dec 01 16:10:16 crc kubenswrapper[4810]: E1201 16:10:16.345171 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa152182-eb48-4241-9d80-73f2c5fc5310" containerName="barbican-api-log" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.345183 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa152182-eb48-4241-9d80-73f2c5fc5310" containerName="barbican-api-log" Dec 01 16:10:16 crc kubenswrapper[4810]: E1201 16:10:16.345191 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b19a52-ce05-49a9-bebd-b3e8a3c9363e" containerName="dnsmasq-dns" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.345198 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b19a52-ce05-49a9-bebd-b3e8a3c9363e" containerName="dnsmasq-dns" Dec 01 16:10:16 crc kubenswrapper[4810]: E1201 16:10:16.345228 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b19a52-ce05-49a9-bebd-b3e8a3c9363e" containerName="init" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.345234 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b19a52-ce05-49a9-bebd-b3e8a3c9363e" containerName="init" Dec 01 16:10:16 crc kubenswrapper[4810]: E1201 16:10:16.345244 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa152182-eb48-4241-9d80-73f2c5fc5310" containerName="barbican-api" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.345250 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa152182-eb48-4241-9d80-73f2c5fc5310" containerName="barbican-api" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.345397 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa152182-eb48-4241-9d80-73f2c5fc5310" containerName="barbican-api" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.345405 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa152182-eb48-4241-9d80-73f2c5fc5310" containerName="barbican-api-log" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.345415 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2b19a52-ce05-49a9-bebd-b3e8a3c9363e" containerName="dnsmasq-dns" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.345987 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2611-account-create-update-b2jpl" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.350116 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.358830 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-s6472"] Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.359922 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s6472" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.375900 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2611-account-create-update-b2jpl"] Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.387973 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-s6472"] Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.497225 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h47kd\" (UniqueName: \"kubernetes.io/projected/cbd26113-dc91-4686-8f81-93eea1d5eedc-kube-api-access-h47kd\") pod \"neutron-db-create-s6472\" (UID: \"cbd26113-dc91-4686-8f81-93eea1d5eedc\") " pod="openstack/neutron-db-create-s6472" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.497311 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbd26113-dc91-4686-8f81-93eea1d5eedc-operator-scripts\") pod \"neutron-db-create-s6472\" (UID: \"cbd26113-dc91-4686-8f81-93eea1d5eedc\") " pod="openstack/neutron-db-create-s6472" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.497372 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn77r\" (UniqueName: \"kubernetes.io/projected/87906600-d845-4859-9875-4137a77092f9-kube-api-access-xn77r\") pod \"neutron-2611-account-create-update-b2jpl\" (UID: \"87906600-d845-4859-9875-4137a77092f9\") " pod="openstack/neutron-2611-account-create-update-b2jpl" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.497410 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87906600-d845-4859-9875-4137a77092f9-operator-scripts\") pod \"neutron-2611-account-create-update-b2jpl\" (UID: \"87906600-d845-4859-9875-4137a77092f9\") " pod="openstack/neutron-2611-account-create-update-b2jpl" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.598977 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h47kd\" (UniqueName: \"kubernetes.io/projected/cbd26113-dc91-4686-8f81-93eea1d5eedc-kube-api-access-h47kd\") pod \"neutron-db-create-s6472\" (UID: \"cbd26113-dc91-4686-8f81-93eea1d5eedc\") " pod="openstack/neutron-db-create-s6472" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.599058 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbd26113-dc91-4686-8f81-93eea1d5eedc-operator-scripts\") pod \"neutron-db-create-s6472\" (UID: \"cbd26113-dc91-4686-8f81-93eea1d5eedc\") " pod="openstack/neutron-db-create-s6472" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.599135 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn77r\" (UniqueName: \"kubernetes.io/projected/87906600-d845-4859-9875-4137a77092f9-kube-api-access-xn77r\") pod \"neutron-2611-account-create-update-b2jpl\" (UID: \"87906600-d845-4859-9875-4137a77092f9\") " pod="openstack/neutron-2611-account-create-update-b2jpl" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.599177 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87906600-d845-4859-9875-4137a77092f9-operator-scripts\") pod \"neutron-2611-account-create-update-b2jpl\" (UID: \"87906600-d845-4859-9875-4137a77092f9\") " pod="openstack/neutron-2611-account-create-update-b2jpl" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.600011 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87906600-d845-4859-9875-4137a77092f9-operator-scripts\") pod \"neutron-2611-account-create-update-b2jpl\" (UID: \"87906600-d845-4859-9875-4137a77092f9\") " pod="openstack/neutron-2611-account-create-update-b2jpl" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.600196 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbd26113-dc91-4686-8f81-93eea1d5eedc-operator-scripts\") pod \"neutron-db-create-s6472\" (UID: \"cbd26113-dc91-4686-8f81-93eea1d5eedc\") " pod="openstack/neutron-db-create-s6472" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.617632 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h47kd\" (UniqueName: \"kubernetes.io/projected/cbd26113-dc91-4686-8f81-93eea1d5eedc-kube-api-access-h47kd\") pod \"neutron-db-create-s6472\" (UID: \"cbd26113-dc91-4686-8f81-93eea1d5eedc\") " pod="openstack/neutron-db-create-s6472" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.618061 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn77r\" (UniqueName: \"kubernetes.io/projected/87906600-d845-4859-9875-4137a77092f9-kube-api-access-xn77r\") pod \"neutron-2611-account-create-update-b2jpl\" (UID: \"87906600-d845-4859-9875-4137a77092f9\") " pod="openstack/neutron-2611-account-create-update-b2jpl" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.672572 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2611-account-create-update-b2jpl" Dec 01 16:10:16 crc kubenswrapper[4810]: I1201 16:10:16.686795 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s6472" Dec 01 16:10:17 crc kubenswrapper[4810]: I1201 16:10:17.199311 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2611-account-create-update-b2jpl"] Dec 01 16:10:17 crc kubenswrapper[4810]: I1201 16:10:17.275536 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-s6472"] Dec 01 16:10:17 crc kubenswrapper[4810]: I1201 16:10:17.850299 4810 generic.go:334] "Generic (PLEG): container finished" podID="cbd26113-dc91-4686-8f81-93eea1d5eedc" containerID="bfbbd76e2e527ce521228e1438ead5d83f6f8dced1636912a78a58ded9f2b6d6" exitCode=0 Dec 01 16:10:17 crc kubenswrapper[4810]: I1201 16:10:17.850702 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-s6472" event={"ID":"cbd26113-dc91-4686-8f81-93eea1d5eedc","Type":"ContainerDied","Data":"bfbbd76e2e527ce521228e1438ead5d83f6f8dced1636912a78a58ded9f2b6d6"} Dec 01 16:10:17 crc kubenswrapper[4810]: I1201 16:10:17.850835 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-s6472" event={"ID":"cbd26113-dc91-4686-8f81-93eea1d5eedc","Type":"ContainerStarted","Data":"57020e8c4672b1429e87be4e48205268657d9eb6ce2fae7209f1b7c13bdfd829"} Dec 01 16:10:17 crc kubenswrapper[4810]: I1201 16:10:17.852817 4810 generic.go:334] "Generic (PLEG): container finished" podID="87906600-d845-4859-9875-4137a77092f9" containerID="488324fcd51e26039702d3aa0feaa773895d50aef3a27513f8a21b002eb7d114" exitCode=0 Dec 01 16:10:17 crc kubenswrapper[4810]: I1201 16:10:17.852852 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2611-account-create-update-b2jpl" event={"ID":"87906600-d845-4859-9875-4137a77092f9","Type":"ContainerDied","Data":"488324fcd51e26039702d3aa0feaa773895d50aef3a27513f8a21b002eb7d114"} Dec 01 16:10:17 crc kubenswrapper[4810]: I1201 16:10:17.852878 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2611-account-create-update-b2jpl" event={"ID":"87906600-d845-4859-9875-4137a77092f9","Type":"ContainerStarted","Data":"6eeb1335b3a75c1a09a798557557c1ac8c005ba94787d9e3230bbb181c8373df"} Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.284272 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2611-account-create-update-b2jpl" Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.291093 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s6472" Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.444164 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbd26113-dc91-4686-8f81-93eea1d5eedc-operator-scripts\") pod \"cbd26113-dc91-4686-8f81-93eea1d5eedc\" (UID: \"cbd26113-dc91-4686-8f81-93eea1d5eedc\") " Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.444696 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h47kd\" (UniqueName: \"kubernetes.io/projected/cbd26113-dc91-4686-8f81-93eea1d5eedc-kube-api-access-h47kd\") pod \"cbd26113-dc91-4686-8f81-93eea1d5eedc\" (UID: \"cbd26113-dc91-4686-8f81-93eea1d5eedc\") " Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.444789 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn77r\" (UniqueName: \"kubernetes.io/projected/87906600-d845-4859-9875-4137a77092f9-kube-api-access-xn77r\") pod \"87906600-d845-4859-9875-4137a77092f9\" (UID: \"87906600-d845-4859-9875-4137a77092f9\") " Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.444920 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87906600-d845-4859-9875-4137a77092f9-operator-scripts\") pod \"87906600-d845-4859-9875-4137a77092f9\" (UID: \"87906600-d845-4859-9875-4137a77092f9\") " Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.444924 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbd26113-dc91-4686-8f81-93eea1d5eedc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cbd26113-dc91-4686-8f81-93eea1d5eedc" (UID: "cbd26113-dc91-4686-8f81-93eea1d5eedc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.445337 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87906600-d845-4859-9875-4137a77092f9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "87906600-d845-4859-9875-4137a77092f9" (UID: "87906600-d845-4859-9875-4137a77092f9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.445361 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbd26113-dc91-4686-8f81-93eea1d5eedc-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.452280 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbd26113-dc91-4686-8f81-93eea1d5eedc-kube-api-access-h47kd" (OuterVolumeSpecName: "kube-api-access-h47kd") pod "cbd26113-dc91-4686-8f81-93eea1d5eedc" (UID: "cbd26113-dc91-4686-8f81-93eea1d5eedc"). InnerVolumeSpecName "kube-api-access-h47kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.452862 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87906600-d845-4859-9875-4137a77092f9-kube-api-access-xn77r" (OuterVolumeSpecName: "kube-api-access-xn77r") pod "87906600-d845-4859-9875-4137a77092f9" (UID: "87906600-d845-4859-9875-4137a77092f9"). InnerVolumeSpecName "kube-api-access-xn77r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.546950 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn77r\" (UniqueName: \"kubernetes.io/projected/87906600-d845-4859-9875-4137a77092f9-kube-api-access-xn77r\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.547001 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87906600-d845-4859-9875-4137a77092f9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.547013 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h47kd\" (UniqueName: \"kubernetes.io/projected/cbd26113-dc91-4686-8f81-93eea1d5eedc-kube-api-access-h47kd\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.869320 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2611-account-create-update-b2jpl" event={"ID":"87906600-d845-4859-9875-4137a77092f9","Type":"ContainerDied","Data":"6eeb1335b3a75c1a09a798557557c1ac8c005ba94787d9e3230bbb181c8373df"} Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.869370 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6eeb1335b3a75c1a09a798557557c1ac8c005ba94787d9e3230bbb181c8373df" Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.869426 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2611-account-create-update-b2jpl" Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.871859 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-s6472" event={"ID":"cbd26113-dc91-4686-8f81-93eea1d5eedc","Type":"ContainerDied","Data":"57020e8c4672b1429e87be4e48205268657d9eb6ce2fae7209f1b7c13bdfd829"} Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.871905 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57020e8c4672b1429e87be4e48205268657d9eb6ce2fae7209f1b7c13bdfd829" Dec 01 16:10:19 crc kubenswrapper[4810]: I1201 16:10:19.871926 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-s6472" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.638627 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-djlnt"] Dec 01 16:10:21 crc kubenswrapper[4810]: E1201 16:10:21.639190 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87906600-d845-4859-9875-4137a77092f9" containerName="mariadb-account-create-update" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.639205 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="87906600-d845-4859-9875-4137a77092f9" containerName="mariadb-account-create-update" Dec 01 16:10:21 crc kubenswrapper[4810]: E1201 16:10:21.639226 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbd26113-dc91-4686-8f81-93eea1d5eedc" containerName="mariadb-database-create" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.639236 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbd26113-dc91-4686-8f81-93eea1d5eedc" containerName="mariadb-database-create" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.639396 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="87906600-d845-4859-9875-4137a77092f9" containerName="mariadb-account-create-update" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.639413 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbd26113-dc91-4686-8f81-93eea1d5eedc" containerName="mariadb-database-create" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.640016 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-djlnt" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.642291 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.642534 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-l8ttf" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.656376 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-djlnt"] Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.656634 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.790382 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caaddf08-f0f1-40d1-ba3a-88c187a1f08b-combined-ca-bundle\") pod \"neutron-db-sync-djlnt\" (UID: \"caaddf08-f0f1-40d1-ba3a-88c187a1f08b\") " pod="openstack/neutron-db-sync-djlnt" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.790433 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pjfn\" (UniqueName: \"kubernetes.io/projected/caaddf08-f0f1-40d1-ba3a-88c187a1f08b-kube-api-access-6pjfn\") pod \"neutron-db-sync-djlnt\" (UID: \"caaddf08-f0f1-40d1-ba3a-88c187a1f08b\") " pod="openstack/neutron-db-sync-djlnt" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.790558 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/caaddf08-f0f1-40d1-ba3a-88c187a1f08b-config\") pod \"neutron-db-sync-djlnt\" (UID: \"caaddf08-f0f1-40d1-ba3a-88c187a1f08b\") " pod="openstack/neutron-db-sync-djlnt" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.891875 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/caaddf08-f0f1-40d1-ba3a-88c187a1f08b-config\") pod \"neutron-db-sync-djlnt\" (UID: \"caaddf08-f0f1-40d1-ba3a-88c187a1f08b\") " pod="openstack/neutron-db-sync-djlnt" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.891956 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caaddf08-f0f1-40d1-ba3a-88c187a1f08b-combined-ca-bundle\") pod \"neutron-db-sync-djlnt\" (UID: \"caaddf08-f0f1-40d1-ba3a-88c187a1f08b\") " pod="openstack/neutron-db-sync-djlnt" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.891983 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pjfn\" (UniqueName: \"kubernetes.io/projected/caaddf08-f0f1-40d1-ba3a-88c187a1f08b-kube-api-access-6pjfn\") pod \"neutron-db-sync-djlnt\" (UID: \"caaddf08-f0f1-40d1-ba3a-88c187a1f08b\") " pod="openstack/neutron-db-sync-djlnt" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.897986 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caaddf08-f0f1-40d1-ba3a-88c187a1f08b-combined-ca-bundle\") pod \"neutron-db-sync-djlnt\" (UID: \"caaddf08-f0f1-40d1-ba3a-88c187a1f08b\") " pod="openstack/neutron-db-sync-djlnt" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.898058 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/caaddf08-f0f1-40d1-ba3a-88c187a1f08b-config\") pod \"neutron-db-sync-djlnt\" (UID: \"caaddf08-f0f1-40d1-ba3a-88c187a1f08b\") " pod="openstack/neutron-db-sync-djlnt" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.917363 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pjfn\" (UniqueName: \"kubernetes.io/projected/caaddf08-f0f1-40d1-ba3a-88c187a1f08b-kube-api-access-6pjfn\") pod \"neutron-db-sync-djlnt\" (UID: \"caaddf08-f0f1-40d1-ba3a-88c187a1f08b\") " pod="openstack/neutron-db-sync-djlnt" Dec 01 16:10:21 crc kubenswrapper[4810]: I1201 16:10:21.956690 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-djlnt" Dec 01 16:10:22 crc kubenswrapper[4810]: I1201 16:10:22.412162 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-djlnt"] Dec 01 16:10:22 crc kubenswrapper[4810]: W1201 16:10:22.422908 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcaaddf08_f0f1_40d1_ba3a_88c187a1f08b.slice/crio-3c44bf36d724b5cad59fda9517e4ee54c91099b1317c816876ffd0b24339b8b3 WatchSource:0}: Error finding container 3c44bf36d724b5cad59fda9517e4ee54c91099b1317c816876ffd0b24339b8b3: Status 404 returned error can't find the container with id 3c44bf36d724b5cad59fda9517e4ee54c91099b1317c816876ffd0b24339b8b3 Dec 01 16:10:22 crc kubenswrapper[4810]: I1201 16:10:22.902643 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-djlnt" event={"ID":"caaddf08-f0f1-40d1-ba3a-88c187a1f08b","Type":"ContainerStarted","Data":"e7c3e1ede2fceeee5ec3674a8a1085fdf71cdb92211e67d47e9bcfeffbd1c695"} Dec 01 16:10:22 crc kubenswrapper[4810]: I1201 16:10:22.903126 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-djlnt" event={"ID":"caaddf08-f0f1-40d1-ba3a-88c187a1f08b","Type":"ContainerStarted","Data":"3c44bf36d724b5cad59fda9517e4ee54c91099b1317c816876ffd0b24339b8b3"} Dec 01 16:10:22 crc kubenswrapper[4810]: I1201 16:10:22.927631 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-djlnt" podStartSLOduration=1.927606934 podStartE2EDuration="1.927606934s" podCreationTimestamp="2025-12-01 16:10:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:10:22.926002641 +0000 UTC m=+5788.689512254" watchObservedRunningTime="2025-12-01 16:10:22.927606934 +0000 UTC m=+5788.691116537" Dec 01 16:10:26 crc kubenswrapper[4810]: I1201 16:10:26.934157 4810 generic.go:334] "Generic (PLEG): container finished" podID="caaddf08-f0f1-40d1-ba3a-88c187a1f08b" containerID="e7c3e1ede2fceeee5ec3674a8a1085fdf71cdb92211e67d47e9bcfeffbd1c695" exitCode=0 Dec 01 16:10:26 crc kubenswrapper[4810]: I1201 16:10:26.934236 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-djlnt" event={"ID":"caaddf08-f0f1-40d1-ba3a-88c187a1f08b","Type":"ContainerDied","Data":"e7c3e1ede2fceeee5ec3674a8a1085fdf71cdb92211e67d47e9bcfeffbd1c695"} Dec 01 16:10:28 crc kubenswrapper[4810]: I1201 16:10:28.259638 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-djlnt" Dec 01 16:10:28 crc kubenswrapper[4810]: I1201 16:10:28.432974 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pjfn\" (UniqueName: \"kubernetes.io/projected/caaddf08-f0f1-40d1-ba3a-88c187a1f08b-kube-api-access-6pjfn\") pod \"caaddf08-f0f1-40d1-ba3a-88c187a1f08b\" (UID: \"caaddf08-f0f1-40d1-ba3a-88c187a1f08b\") " Dec 01 16:10:28 crc kubenswrapper[4810]: I1201 16:10:28.433290 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caaddf08-f0f1-40d1-ba3a-88c187a1f08b-combined-ca-bundle\") pod \"caaddf08-f0f1-40d1-ba3a-88c187a1f08b\" (UID: \"caaddf08-f0f1-40d1-ba3a-88c187a1f08b\") " Dec 01 16:10:28 crc kubenswrapper[4810]: I1201 16:10:28.433521 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/caaddf08-f0f1-40d1-ba3a-88c187a1f08b-config\") pod \"caaddf08-f0f1-40d1-ba3a-88c187a1f08b\" (UID: \"caaddf08-f0f1-40d1-ba3a-88c187a1f08b\") " Dec 01 16:10:28 crc kubenswrapper[4810]: I1201 16:10:28.439798 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caaddf08-f0f1-40d1-ba3a-88c187a1f08b-kube-api-access-6pjfn" (OuterVolumeSpecName: "kube-api-access-6pjfn") pod "caaddf08-f0f1-40d1-ba3a-88c187a1f08b" (UID: "caaddf08-f0f1-40d1-ba3a-88c187a1f08b"). InnerVolumeSpecName "kube-api-access-6pjfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:10:28 crc kubenswrapper[4810]: I1201 16:10:28.457604 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caaddf08-f0f1-40d1-ba3a-88c187a1f08b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "caaddf08-f0f1-40d1-ba3a-88c187a1f08b" (UID: "caaddf08-f0f1-40d1-ba3a-88c187a1f08b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:10:28 crc kubenswrapper[4810]: I1201 16:10:28.458987 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caaddf08-f0f1-40d1-ba3a-88c187a1f08b-config" (OuterVolumeSpecName: "config") pod "caaddf08-f0f1-40d1-ba3a-88c187a1f08b" (UID: "caaddf08-f0f1-40d1-ba3a-88c187a1f08b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:10:28 crc kubenswrapper[4810]: I1201 16:10:28.535339 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/caaddf08-f0f1-40d1-ba3a-88c187a1f08b-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:28 crc kubenswrapper[4810]: I1201 16:10:28.535391 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pjfn\" (UniqueName: \"kubernetes.io/projected/caaddf08-f0f1-40d1-ba3a-88c187a1f08b-kube-api-access-6pjfn\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:28 crc kubenswrapper[4810]: I1201 16:10:28.535402 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caaddf08-f0f1-40d1-ba3a-88c187a1f08b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:28 crc kubenswrapper[4810]: I1201 16:10:28.958278 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-djlnt" event={"ID":"caaddf08-f0f1-40d1-ba3a-88c187a1f08b","Type":"ContainerDied","Data":"3c44bf36d724b5cad59fda9517e4ee54c91099b1317c816876ffd0b24339b8b3"} Dec 01 16:10:28 crc kubenswrapper[4810]: I1201 16:10:28.958354 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c44bf36d724b5cad59fda9517e4ee54c91099b1317c816876ffd0b24339b8b3" Dec 01 16:10:28 crc kubenswrapper[4810]: I1201 16:10:28.958885 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-djlnt" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.114185 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59596df985-rf9x5"] Dec 01 16:10:29 crc kubenswrapper[4810]: E1201 16:10:29.116257 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caaddf08-f0f1-40d1-ba3a-88c187a1f08b" containerName="neutron-db-sync" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.116330 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="caaddf08-f0f1-40d1-ba3a-88c187a1f08b" containerName="neutron-db-sync" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.116575 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="caaddf08-f0f1-40d1-ba3a-88c187a1f08b" containerName="neutron-db-sync" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.117588 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.145139 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59596df985-rf9x5"] Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.241801 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-54b67b4b84-rgd8r"] Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.243723 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.246818 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.246886 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.246928 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.247081 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-l8ttf" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.249503 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-config\") pod \"dnsmasq-dns-59596df985-rf9x5\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.249554 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-ovsdbserver-sb\") pod \"dnsmasq-dns-59596df985-rf9x5\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.249625 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jltr\" (UniqueName: \"kubernetes.io/projected/3b855aff-4523-4dba-8c45-341e7c9807bc-kube-api-access-6jltr\") pod \"dnsmasq-dns-59596df985-rf9x5\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.249738 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-ovsdbserver-nb\") pod \"dnsmasq-dns-59596df985-rf9x5\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.250152 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-dns-svc\") pod \"dnsmasq-dns-59596df985-rf9x5\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.259934 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-54b67b4b84-rgd8r"] Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.352927 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-ovndb-tls-certs\") pod \"neutron-54b67b4b84-rgd8r\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.353000 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-ovsdbserver-nb\") pod \"dnsmasq-dns-59596df985-rf9x5\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.353063 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-config\") pod \"neutron-54b67b4b84-rgd8r\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.353130 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfh9v\" (UniqueName: \"kubernetes.io/projected/36e37e75-4923-4444-8ebc-a40204eb43d8-kube-api-access-hfh9v\") pod \"neutron-54b67b4b84-rgd8r\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.353183 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-dns-svc\") pod \"dnsmasq-dns-59596df985-rf9x5\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.353261 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-httpd-config\") pod \"neutron-54b67b4b84-rgd8r\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.353307 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-config\") pod \"dnsmasq-dns-59596df985-rf9x5\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.353331 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-combined-ca-bundle\") pod \"neutron-54b67b4b84-rgd8r\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.353361 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-ovsdbserver-sb\") pod \"dnsmasq-dns-59596df985-rf9x5\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.353387 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jltr\" (UniqueName: \"kubernetes.io/projected/3b855aff-4523-4dba-8c45-341e7c9807bc-kube-api-access-6jltr\") pod \"dnsmasq-dns-59596df985-rf9x5\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.354084 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-ovsdbserver-nb\") pod \"dnsmasq-dns-59596df985-rf9x5\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.354497 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-ovsdbserver-sb\") pod \"dnsmasq-dns-59596df985-rf9x5\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.354688 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-config\") pod \"dnsmasq-dns-59596df985-rf9x5\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.354775 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-dns-svc\") pod \"dnsmasq-dns-59596df985-rf9x5\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.371682 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jltr\" (UniqueName: \"kubernetes.io/projected/3b855aff-4523-4dba-8c45-341e7c9807bc-kube-api-access-6jltr\") pod \"dnsmasq-dns-59596df985-rf9x5\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.437284 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.454852 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-config\") pod \"neutron-54b67b4b84-rgd8r\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.455425 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfh9v\" (UniqueName: \"kubernetes.io/projected/36e37e75-4923-4444-8ebc-a40204eb43d8-kube-api-access-hfh9v\") pod \"neutron-54b67b4b84-rgd8r\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.455563 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-httpd-config\") pod \"neutron-54b67b4b84-rgd8r\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.455610 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-combined-ca-bundle\") pod \"neutron-54b67b4b84-rgd8r\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.455656 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-ovndb-tls-certs\") pod \"neutron-54b67b4b84-rgd8r\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.459675 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-ovndb-tls-certs\") pod \"neutron-54b67b4b84-rgd8r\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.461877 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-config\") pod \"neutron-54b67b4b84-rgd8r\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.463880 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-httpd-config\") pod \"neutron-54b67b4b84-rgd8r\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.466139 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-combined-ca-bundle\") pod \"neutron-54b67b4b84-rgd8r\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.476430 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfh9v\" (UniqueName: \"kubernetes.io/projected/36e37e75-4923-4444-8ebc-a40204eb43d8-kube-api-access-hfh9v\") pod \"neutron-54b67b4b84-rgd8r\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.559834 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:10:29 crc kubenswrapper[4810]: I1201 16:10:29.978364 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59596df985-rf9x5"] Dec 01 16:10:30 crc kubenswrapper[4810]: I1201 16:10:30.247667 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-54b67b4b84-rgd8r"] Dec 01 16:10:30 crc kubenswrapper[4810]: I1201 16:10:30.979339 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54b67b4b84-rgd8r" event={"ID":"36e37e75-4923-4444-8ebc-a40204eb43d8","Type":"ContainerStarted","Data":"53782854fb166923313ae7d496dfb979c8e9cae6e6f1b6ddb888f0d0f4c88e52"} Dec 01 16:10:30 crc kubenswrapper[4810]: I1201 16:10:30.979685 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54b67b4b84-rgd8r" event={"ID":"36e37e75-4923-4444-8ebc-a40204eb43d8","Type":"ContainerStarted","Data":"0e9ef3a90011b18fa53c09a0ee9bee8027c12f958de5a5e85201296faa549d17"} Dec 01 16:10:30 crc kubenswrapper[4810]: I1201 16:10:30.979698 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54b67b4b84-rgd8r" event={"ID":"36e37e75-4923-4444-8ebc-a40204eb43d8","Type":"ContainerStarted","Data":"e7dd23fffb0c3141fe71a1e2451046ad325d798e7af87e1e9b8bd0b07efe0ec5"} Dec 01 16:10:30 crc kubenswrapper[4810]: I1201 16:10:30.980879 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:10:30 crc kubenswrapper[4810]: I1201 16:10:30.984309 4810 generic.go:334] "Generic (PLEG): container finished" podID="3b855aff-4523-4dba-8c45-341e7c9807bc" containerID="12b03cbcefc24ecd02a6041471e8b3cf5452f647fd4d5d5bf8253ec98dc1401d" exitCode=0 Dec 01 16:10:30 crc kubenswrapper[4810]: I1201 16:10:30.984346 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59596df985-rf9x5" event={"ID":"3b855aff-4523-4dba-8c45-341e7c9807bc","Type":"ContainerDied","Data":"12b03cbcefc24ecd02a6041471e8b3cf5452f647fd4d5d5bf8253ec98dc1401d"} Dec 01 16:10:30 crc kubenswrapper[4810]: I1201 16:10:30.984371 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59596df985-rf9x5" event={"ID":"3b855aff-4523-4dba-8c45-341e7c9807bc","Type":"ContainerStarted","Data":"e4eb764fb9a76a7c4422900028d0a1b3092852d86e40764f21606ff155597072"} Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:30.999789 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-54b67b4b84-rgd8r" podStartSLOduration=1.999771824 podStartE2EDuration="1.999771824s" podCreationTimestamp="2025-12-01 16:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:10:30.993659478 +0000 UTC m=+5796.757169081" watchObservedRunningTime="2025-12-01 16:10:30.999771824 +0000 UTC m=+5796.763281417" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.229605 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-9776bf6cc-vk57d"] Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.231810 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.236984 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.237280 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.246689 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9776bf6cc-vk57d"] Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.295553 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fc53f93-4c92-44c4-998b-8be33a38d18f-internal-tls-certs\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.295622 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fc53f93-4c92-44c4-998b-8be33a38d18f-ovndb-tls-certs\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.295662 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fc53f93-4c92-44c4-998b-8be33a38d18f-combined-ca-bundle\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.295697 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3fc53f93-4c92-44c4-998b-8be33a38d18f-config\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.295775 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3fc53f93-4c92-44c4-998b-8be33a38d18f-httpd-config\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.295799 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fc53f93-4c92-44c4-998b-8be33a38d18f-public-tls-certs\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.295851 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxgr8\" (UniqueName: \"kubernetes.io/projected/3fc53f93-4c92-44c4-998b-8be33a38d18f-kube-api-access-pxgr8\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.403769 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3fc53f93-4c92-44c4-998b-8be33a38d18f-config\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.403824 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3fc53f93-4c92-44c4-998b-8be33a38d18f-httpd-config\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.403846 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fc53f93-4c92-44c4-998b-8be33a38d18f-public-tls-certs\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.403884 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxgr8\" (UniqueName: \"kubernetes.io/projected/3fc53f93-4c92-44c4-998b-8be33a38d18f-kube-api-access-pxgr8\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.403987 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fc53f93-4c92-44c4-998b-8be33a38d18f-internal-tls-certs\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.404011 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fc53f93-4c92-44c4-998b-8be33a38d18f-ovndb-tls-certs\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.404032 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fc53f93-4c92-44c4-998b-8be33a38d18f-combined-ca-bundle\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.423383 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3fc53f93-4c92-44c4-998b-8be33a38d18f-config\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.424341 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fc53f93-4c92-44c4-998b-8be33a38d18f-combined-ca-bundle\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.424858 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fc53f93-4c92-44c4-998b-8be33a38d18f-public-tls-certs\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.428240 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fc53f93-4c92-44c4-998b-8be33a38d18f-internal-tls-certs\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.428441 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3fc53f93-4c92-44c4-998b-8be33a38d18f-httpd-config\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.433052 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fc53f93-4c92-44c4-998b-8be33a38d18f-ovndb-tls-certs\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.457276 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxgr8\" (UniqueName: \"kubernetes.io/projected/3fc53f93-4c92-44c4-998b-8be33a38d18f-kube-api-access-pxgr8\") pod \"neutron-9776bf6cc-vk57d\" (UID: \"3fc53f93-4c92-44c4-998b-8be33a38d18f\") " pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:31 crc kubenswrapper[4810]: I1201 16:10:31.555032 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:32 crc kubenswrapper[4810]: I1201 16:10:32.002701 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59596df985-rf9x5" event={"ID":"3b855aff-4523-4dba-8c45-341e7c9807bc","Type":"ContainerStarted","Data":"e30ac6813835f238af5ecd4c2ae2a64754ebabc796d14bb5c6d179b184bdfc67"} Dec 01 16:10:32 crc kubenswrapper[4810]: I1201 16:10:32.026590 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59596df985-rf9x5" podStartSLOduration=3.026574186 podStartE2EDuration="3.026574186s" podCreationTimestamp="2025-12-01 16:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:10:32.025725764 +0000 UTC m=+5797.789235377" watchObservedRunningTime="2025-12-01 16:10:32.026574186 +0000 UTC m=+5797.790083789" Dec 01 16:10:32 crc kubenswrapper[4810]: I1201 16:10:32.097808 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9776bf6cc-vk57d"] Dec 01 16:10:32 crc kubenswrapper[4810]: W1201 16:10:32.099428 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3fc53f93_4c92_44c4_998b_8be33a38d18f.slice/crio-589d82e88009d0b8eac526cc88fdf0c916f1daf3ef6bd63fe96d1b6322b52ac4 WatchSource:0}: Error finding container 589d82e88009d0b8eac526cc88fdf0c916f1daf3ef6bd63fe96d1b6322b52ac4: Status 404 returned error can't find the container with id 589d82e88009d0b8eac526cc88fdf0c916f1daf3ef6bd63fe96d1b6322b52ac4 Dec 01 16:10:32 crc kubenswrapper[4810]: I1201 16:10:32.972233 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:10:32 crc kubenswrapper[4810]: I1201 16:10:32.972549 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:10:32 crc kubenswrapper[4810]: I1201 16:10:32.972583 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 16:10:32 crc kubenswrapper[4810]: I1201 16:10:32.973220 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a05157f175de8551ab2001ecba2c36ed1be28837c6a63ea1ad99ac3229a74f07"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:10:32 crc kubenswrapper[4810]: I1201 16:10:32.973280 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://a05157f175de8551ab2001ecba2c36ed1be28837c6a63ea1ad99ac3229a74f07" gracePeriod=600 Dec 01 16:10:33 crc kubenswrapper[4810]: I1201 16:10:33.012662 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9776bf6cc-vk57d" event={"ID":"3fc53f93-4c92-44c4-998b-8be33a38d18f","Type":"ContainerStarted","Data":"152a805059ea2d8c1b4ee81fc26d3b7557e35bd03648ad5249365662c6e473c4"} Dec 01 16:10:33 crc kubenswrapper[4810]: I1201 16:10:33.012698 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9776bf6cc-vk57d" event={"ID":"3fc53f93-4c92-44c4-998b-8be33a38d18f","Type":"ContainerStarted","Data":"cc2a24a5fadbee40427da2b97a0e0a00a77b699b118362470bfe62021ab11f2e"} Dec 01 16:10:33 crc kubenswrapper[4810]: I1201 16:10:33.012708 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9776bf6cc-vk57d" event={"ID":"3fc53f93-4c92-44c4-998b-8be33a38d18f","Type":"ContainerStarted","Data":"589d82e88009d0b8eac526cc88fdf0c916f1daf3ef6bd63fe96d1b6322b52ac4"} Dec 01 16:10:33 crc kubenswrapper[4810]: I1201 16:10:33.012721 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:33 crc kubenswrapper[4810]: I1201 16:10:33.013164 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:10:33 crc kubenswrapper[4810]: I1201 16:10:33.030312 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-9776bf6cc-vk57d" podStartSLOduration=2.030294262 podStartE2EDuration="2.030294262s" podCreationTimestamp="2025-12-01 16:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:10:33.027566848 +0000 UTC m=+5798.791076451" watchObservedRunningTime="2025-12-01 16:10:33.030294262 +0000 UTC m=+5798.793803865" Dec 01 16:10:34 crc kubenswrapper[4810]: I1201 16:10:34.024987 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="a05157f175de8551ab2001ecba2c36ed1be28837c6a63ea1ad99ac3229a74f07" exitCode=0 Dec 01 16:10:34 crc kubenswrapper[4810]: I1201 16:10:34.025046 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"a05157f175de8551ab2001ecba2c36ed1be28837c6a63ea1ad99ac3229a74f07"} Dec 01 16:10:34 crc kubenswrapper[4810]: I1201 16:10:34.025529 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d"} Dec 01 16:10:34 crc kubenswrapper[4810]: I1201 16:10:34.025577 4810 scope.go:117] "RemoveContainer" containerID="5723179c03108d98d7585da5472bfeafa08e46f1607bf65b9d864d81db1f7706" Dec 01 16:10:39 crc kubenswrapper[4810]: I1201 16:10:39.438691 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:10:39 crc kubenswrapper[4810]: I1201 16:10:39.500834 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84555f7879-rcsn2"] Dec 01 16:10:39 crc kubenswrapper[4810]: I1201 16:10:39.501083 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84555f7879-rcsn2" podUID="029bd0e5-0d9e-4734-a15c-abae3e06df04" containerName="dnsmasq-dns" containerID="cri-o://ec8fce0ff6a4639f090f98f71f8bf824cdbb75960b862dcba28175c8688023a3" gracePeriod=10 Dec 01 16:10:39 crc kubenswrapper[4810]: I1201 16:10:39.998989 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.125922 4810 generic.go:334] "Generic (PLEG): container finished" podID="029bd0e5-0d9e-4734-a15c-abae3e06df04" containerID="ec8fce0ff6a4639f090f98f71f8bf824cdbb75960b862dcba28175c8688023a3" exitCode=0 Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.125974 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84555f7879-rcsn2" event={"ID":"029bd0e5-0d9e-4734-a15c-abae3e06df04","Type":"ContainerDied","Data":"ec8fce0ff6a4639f090f98f71f8bf824cdbb75960b862dcba28175c8688023a3"} Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.126004 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84555f7879-rcsn2" event={"ID":"029bd0e5-0d9e-4734-a15c-abae3e06df04","Type":"ContainerDied","Data":"40c8d9e9d78b27fea97a8ce51ce522b91a958e0731a034bdab45ae6f799750b2"} Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.126014 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84555f7879-rcsn2" Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.126024 4810 scope.go:117] "RemoveContainer" containerID="ec8fce0ff6a4639f090f98f71f8bf824cdbb75960b862dcba28175c8688023a3" Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.159519 4810 scope.go:117] "RemoveContainer" containerID="fd717eb5e7d444f3f5b8f3205ff674410d870a14f32256415fbe7def4c969cb4" Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.167248 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtl7j\" (UniqueName: \"kubernetes.io/projected/029bd0e5-0d9e-4734-a15c-abae3e06df04-kube-api-access-qtl7j\") pod \"029bd0e5-0d9e-4734-a15c-abae3e06df04\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.167391 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-config\") pod \"029bd0e5-0d9e-4734-a15c-abae3e06df04\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.167422 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-ovsdbserver-nb\") pod \"029bd0e5-0d9e-4734-a15c-abae3e06df04\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.167452 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-dns-svc\") pod \"029bd0e5-0d9e-4734-a15c-abae3e06df04\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.167489 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-ovsdbserver-sb\") pod \"029bd0e5-0d9e-4734-a15c-abae3e06df04\" (UID: \"029bd0e5-0d9e-4734-a15c-abae3e06df04\") " Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.183716 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/029bd0e5-0d9e-4734-a15c-abae3e06df04-kube-api-access-qtl7j" (OuterVolumeSpecName: "kube-api-access-qtl7j") pod "029bd0e5-0d9e-4734-a15c-abae3e06df04" (UID: "029bd0e5-0d9e-4734-a15c-abae3e06df04"). InnerVolumeSpecName "kube-api-access-qtl7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.202504 4810 scope.go:117] "RemoveContainer" containerID="ec8fce0ff6a4639f090f98f71f8bf824cdbb75960b862dcba28175c8688023a3" Dec 01 16:10:40 crc kubenswrapper[4810]: E1201 16:10:40.204961 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec8fce0ff6a4639f090f98f71f8bf824cdbb75960b862dcba28175c8688023a3\": container with ID starting with ec8fce0ff6a4639f090f98f71f8bf824cdbb75960b862dcba28175c8688023a3 not found: ID does not exist" containerID="ec8fce0ff6a4639f090f98f71f8bf824cdbb75960b862dcba28175c8688023a3" Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.205353 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec8fce0ff6a4639f090f98f71f8bf824cdbb75960b862dcba28175c8688023a3"} err="failed to get container status \"ec8fce0ff6a4639f090f98f71f8bf824cdbb75960b862dcba28175c8688023a3\": rpc error: code = NotFound desc = could not find container \"ec8fce0ff6a4639f090f98f71f8bf824cdbb75960b862dcba28175c8688023a3\": container with ID starting with ec8fce0ff6a4639f090f98f71f8bf824cdbb75960b862dcba28175c8688023a3 not found: ID does not exist" Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.205585 4810 scope.go:117] "RemoveContainer" containerID="fd717eb5e7d444f3f5b8f3205ff674410d870a14f32256415fbe7def4c969cb4" Dec 01 16:10:40 crc kubenswrapper[4810]: E1201 16:10:40.208934 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd717eb5e7d444f3f5b8f3205ff674410d870a14f32256415fbe7def4c969cb4\": container with ID starting with fd717eb5e7d444f3f5b8f3205ff674410d870a14f32256415fbe7def4c969cb4 not found: ID does not exist" containerID="fd717eb5e7d444f3f5b8f3205ff674410d870a14f32256415fbe7def4c969cb4" Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.209000 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd717eb5e7d444f3f5b8f3205ff674410d870a14f32256415fbe7def4c969cb4"} err="failed to get container status \"fd717eb5e7d444f3f5b8f3205ff674410d870a14f32256415fbe7def4c969cb4\": rpc error: code = NotFound desc = could not find container \"fd717eb5e7d444f3f5b8f3205ff674410d870a14f32256415fbe7def4c969cb4\": container with ID starting with fd717eb5e7d444f3f5b8f3205ff674410d870a14f32256415fbe7def4c969cb4 not found: ID does not exist" Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.231651 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "029bd0e5-0d9e-4734-a15c-abae3e06df04" (UID: "029bd0e5-0d9e-4734-a15c-abae3e06df04"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.231662 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-config" (OuterVolumeSpecName: "config") pod "029bd0e5-0d9e-4734-a15c-abae3e06df04" (UID: "029bd0e5-0d9e-4734-a15c-abae3e06df04"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.233739 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "029bd0e5-0d9e-4734-a15c-abae3e06df04" (UID: "029bd0e5-0d9e-4734-a15c-abae3e06df04"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.243542 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "029bd0e5-0d9e-4734-a15c-abae3e06df04" (UID: "029bd0e5-0d9e-4734-a15c-abae3e06df04"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.270394 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.270431 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.270441 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.270451 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/029bd0e5-0d9e-4734-a15c-abae3e06df04-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.270461 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtl7j\" (UniqueName: \"kubernetes.io/projected/029bd0e5-0d9e-4734-a15c-abae3e06df04-kube-api-access-qtl7j\") on node \"crc\" DevicePath \"\"" Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.459446 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84555f7879-rcsn2"] Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.470511 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84555f7879-rcsn2"] Dec 01 16:10:40 crc kubenswrapper[4810]: I1201 16:10:40.502335 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="029bd0e5-0d9e-4734-a15c-abae3e06df04" path="/var/lib/kubelet/pods/029bd0e5-0d9e-4734-a15c-abae3e06df04/volumes" Dec 01 16:10:59 crc kubenswrapper[4810]: I1201 16:10:59.572026 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:11:01 crc kubenswrapper[4810]: I1201 16:11:01.567323 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-9776bf6cc-vk57d" Dec 01 16:11:01 crc kubenswrapper[4810]: I1201 16:11:01.642425 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-54b67b4b84-rgd8r"] Dec 01 16:11:01 crc kubenswrapper[4810]: I1201 16:11:01.642924 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-54b67b4b84-rgd8r" podUID="36e37e75-4923-4444-8ebc-a40204eb43d8" containerName="neutron-api" containerID="cri-o://0e9ef3a90011b18fa53c09a0ee9bee8027c12f958de5a5e85201296faa549d17" gracePeriod=30 Dec 01 16:11:01 crc kubenswrapper[4810]: I1201 16:11:01.643212 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-54b67b4b84-rgd8r" podUID="36e37e75-4923-4444-8ebc-a40204eb43d8" containerName="neutron-httpd" containerID="cri-o://53782854fb166923313ae7d496dfb979c8e9cae6e6f1b6ddb888f0d0f4c88e52" gracePeriod=30 Dec 01 16:11:02 crc kubenswrapper[4810]: I1201 16:11:02.330335 4810 generic.go:334] "Generic (PLEG): container finished" podID="36e37e75-4923-4444-8ebc-a40204eb43d8" containerID="53782854fb166923313ae7d496dfb979c8e9cae6e6f1b6ddb888f0d0f4c88e52" exitCode=0 Dec 01 16:11:02 crc kubenswrapper[4810]: I1201 16:11:02.330396 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54b67b4b84-rgd8r" event={"ID":"36e37e75-4923-4444-8ebc-a40204eb43d8","Type":"ContainerDied","Data":"53782854fb166923313ae7d496dfb979c8e9cae6e6f1b6ddb888f0d0f4c88e52"} Dec 01 16:11:05 crc kubenswrapper[4810]: I1201 16:11:05.989971 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.121379 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-ovndb-tls-certs\") pod \"36e37e75-4923-4444-8ebc-a40204eb43d8\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.121429 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-httpd-config\") pod \"36e37e75-4923-4444-8ebc-a40204eb43d8\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.121507 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfh9v\" (UniqueName: \"kubernetes.io/projected/36e37e75-4923-4444-8ebc-a40204eb43d8-kube-api-access-hfh9v\") pod \"36e37e75-4923-4444-8ebc-a40204eb43d8\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.121626 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-combined-ca-bundle\") pod \"36e37e75-4923-4444-8ebc-a40204eb43d8\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.121724 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-config\") pod \"36e37e75-4923-4444-8ebc-a40204eb43d8\" (UID: \"36e37e75-4923-4444-8ebc-a40204eb43d8\") " Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.127414 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "36e37e75-4923-4444-8ebc-a40204eb43d8" (UID: "36e37e75-4923-4444-8ebc-a40204eb43d8"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.130027 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36e37e75-4923-4444-8ebc-a40204eb43d8-kube-api-access-hfh9v" (OuterVolumeSpecName: "kube-api-access-hfh9v") pod "36e37e75-4923-4444-8ebc-a40204eb43d8" (UID: "36e37e75-4923-4444-8ebc-a40204eb43d8"). InnerVolumeSpecName "kube-api-access-hfh9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.169782 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-config" (OuterVolumeSpecName: "config") pod "36e37e75-4923-4444-8ebc-a40204eb43d8" (UID: "36e37e75-4923-4444-8ebc-a40204eb43d8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.169874 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36e37e75-4923-4444-8ebc-a40204eb43d8" (UID: "36e37e75-4923-4444-8ebc-a40204eb43d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.193305 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "36e37e75-4923-4444-8ebc-a40204eb43d8" (UID: "36e37e75-4923-4444-8ebc-a40204eb43d8"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.223364 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.223392 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.223401 4810 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.223409 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/36e37e75-4923-4444-8ebc-a40204eb43d8-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.223418 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfh9v\" (UniqueName: \"kubernetes.io/projected/36e37e75-4923-4444-8ebc-a40204eb43d8-kube-api-access-hfh9v\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.377725 4810 generic.go:334] "Generic (PLEG): container finished" podID="36e37e75-4923-4444-8ebc-a40204eb43d8" containerID="0e9ef3a90011b18fa53c09a0ee9bee8027c12f958de5a5e85201296faa549d17" exitCode=0 Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.377796 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54b67b4b84-rgd8r" event={"ID":"36e37e75-4923-4444-8ebc-a40204eb43d8","Type":"ContainerDied","Data":"0e9ef3a90011b18fa53c09a0ee9bee8027c12f958de5a5e85201296faa549d17"} Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.377819 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54b67b4b84-rgd8r" Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.377837 4810 scope.go:117] "RemoveContainer" containerID="53782854fb166923313ae7d496dfb979c8e9cae6e6f1b6ddb888f0d0f4c88e52" Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.377824 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54b67b4b84-rgd8r" event={"ID":"36e37e75-4923-4444-8ebc-a40204eb43d8","Type":"ContainerDied","Data":"e7dd23fffb0c3141fe71a1e2451046ad325d798e7af87e1e9b8bd0b07efe0ec5"} Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.399069 4810 scope.go:117] "RemoveContainer" containerID="0e9ef3a90011b18fa53c09a0ee9bee8027c12f958de5a5e85201296faa549d17" Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.416341 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-54b67b4b84-rgd8r"] Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.422771 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-54b67b4b84-rgd8r"] Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.424600 4810 scope.go:117] "RemoveContainer" containerID="53782854fb166923313ae7d496dfb979c8e9cae6e6f1b6ddb888f0d0f4c88e52" Dec 01 16:11:06 crc kubenswrapper[4810]: E1201 16:11:06.425191 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53782854fb166923313ae7d496dfb979c8e9cae6e6f1b6ddb888f0d0f4c88e52\": container with ID starting with 53782854fb166923313ae7d496dfb979c8e9cae6e6f1b6ddb888f0d0f4c88e52 not found: ID does not exist" containerID="53782854fb166923313ae7d496dfb979c8e9cae6e6f1b6ddb888f0d0f4c88e52" Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.425237 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53782854fb166923313ae7d496dfb979c8e9cae6e6f1b6ddb888f0d0f4c88e52"} err="failed to get container status \"53782854fb166923313ae7d496dfb979c8e9cae6e6f1b6ddb888f0d0f4c88e52\": rpc error: code = NotFound desc = could not find container \"53782854fb166923313ae7d496dfb979c8e9cae6e6f1b6ddb888f0d0f4c88e52\": container with ID starting with 53782854fb166923313ae7d496dfb979c8e9cae6e6f1b6ddb888f0d0f4c88e52 not found: ID does not exist" Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.425270 4810 scope.go:117] "RemoveContainer" containerID="0e9ef3a90011b18fa53c09a0ee9bee8027c12f958de5a5e85201296faa549d17" Dec 01 16:11:06 crc kubenswrapper[4810]: E1201 16:11:06.425690 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e9ef3a90011b18fa53c09a0ee9bee8027c12f958de5a5e85201296faa549d17\": container with ID starting with 0e9ef3a90011b18fa53c09a0ee9bee8027c12f958de5a5e85201296faa549d17 not found: ID does not exist" containerID="0e9ef3a90011b18fa53c09a0ee9bee8027c12f958de5a5e85201296faa549d17" Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.425720 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e9ef3a90011b18fa53c09a0ee9bee8027c12f958de5a5e85201296faa549d17"} err="failed to get container status \"0e9ef3a90011b18fa53c09a0ee9bee8027c12f958de5a5e85201296faa549d17\": rpc error: code = NotFound desc = could not find container \"0e9ef3a90011b18fa53c09a0ee9bee8027c12f958de5a5e85201296faa549d17\": container with ID starting with 0e9ef3a90011b18fa53c09a0ee9bee8027c12f958de5a5e85201296faa549d17 not found: ID does not exist" Dec 01 16:11:06 crc kubenswrapper[4810]: I1201 16:11:06.503451 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36e37e75-4923-4444-8ebc-a40204eb43d8" path="/var/lib/kubelet/pods/36e37e75-4923-4444-8ebc-a40204eb43d8/volumes" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.521717 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-zcb6q"] Dec 01 16:11:11 crc kubenswrapper[4810]: E1201 16:11:11.522932 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36e37e75-4923-4444-8ebc-a40204eb43d8" containerName="neutron-api" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.522955 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="36e37e75-4923-4444-8ebc-a40204eb43d8" containerName="neutron-api" Dec 01 16:11:11 crc kubenswrapper[4810]: E1201 16:11:11.522980 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="029bd0e5-0d9e-4734-a15c-abae3e06df04" containerName="dnsmasq-dns" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.522986 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="029bd0e5-0d9e-4734-a15c-abae3e06df04" containerName="dnsmasq-dns" Dec 01 16:11:11 crc kubenswrapper[4810]: E1201 16:11:11.523005 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="029bd0e5-0d9e-4734-a15c-abae3e06df04" containerName="init" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.523013 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="029bd0e5-0d9e-4734-a15c-abae3e06df04" containerName="init" Dec 01 16:11:11 crc kubenswrapper[4810]: E1201 16:11:11.523031 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36e37e75-4923-4444-8ebc-a40204eb43d8" containerName="neutron-httpd" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.523039 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="36e37e75-4923-4444-8ebc-a40204eb43d8" containerName="neutron-httpd" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.523205 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="36e37e75-4923-4444-8ebc-a40204eb43d8" containerName="neutron-httpd" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.523220 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="36e37e75-4923-4444-8ebc-a40204eb43d8" containerName="neutron-api" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.523227 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="029bd0e5-0d9e-4734-a15c-abae3e06df04" containerName="dnsmasq-dns" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.523935 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.537963 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.538373 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.538615 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.539102 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.541366 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-5464k" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.566532 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-zcb6q"] Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.621139 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-etc-swift\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.621361 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-swiftconf\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.621874 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-dispersionconf\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.621895 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-combined-ca-bundle\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.622017 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwgnp\" (UniqueName: \"kubernetes.io/projected/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-kube-api-access-rwgnp\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.622042 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-scripts\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.622063 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-ring-data-devices\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.692983 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-667758f697-spt2d"] Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.694623 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.712940 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-667758f697-spt2d"] Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.722868 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-scripts\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.722915 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-ring-data-devices\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.722964 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-etc-swift\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.723007 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-swiftconf\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.723053 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-dispersionconf\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.723069 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-combined-ca-bundle\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.723115 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwgnp\" (UniqueName: \"kubernetes.io/projected/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-kube-api-access-rwgnp\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.723753 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-etc-swift\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.724348 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-scripts\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.724791 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-ring-data-devices\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.730064 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-dispersionconf\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.730683 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-swiftconf\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.759115 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-combined-ca-bundle\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.760994 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwgnp\" (UniqueName: \"kubernetes.io/projected/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-kube-api-access-rwgnp\") pod \"swift-ring-rebalance-zcb6q\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.824684 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-ovsdbserver-sb\") pod \"dnsmasq-dns-667758f697-spt2d\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.824784 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-config\") pod \"dnsmasq-dns-667758f697-spt2d\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.824826 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-ovsdbserver-nb\") pod \"dnsmasq-dns-667758f697-spt2d\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.824887 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6n4j\" (UniqueName: \"kubernetes.io/projected/51ae399d-0f9d-4644-8c93-d3e3a0446e47-kube-api-access-g6n4j\") pod \"dnsmasq-dns-667758f697-spt2d\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.824958 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-dns-svc\") pod \"dnsmasq-dns-667758f697-spt2d\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.896351 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.926411 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-ovsdbserver-nb\") pod \"dnsmasq-dns-667758f697-spt2d\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.927278 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6n4j\" (UniqueName: \"kubernetes.io/projected/51ae399d-0f9d-4644-8c93-d3e3a0446e47-kube-api-access-g6n4j\") pod \"dnsmasq-dns-667758f697-spt2d\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.927358 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-dns-svc\") pod \"dnsmasq-dns-667758f697-spt2d\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.927364 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-ovsdbserver-nb\") pod \"dnsmasq-dns-667758f697-spt2d\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.927494 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-ovsdbserver-sb\") pod \"dnsmasq-dns-667758f697-spt2d\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.927606 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-config\") pod \"dnsmasq-dns-667758f697-spt2d\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.928071 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-dns-svc\") pod \"dnsmasq-dns-667758f697-spt2d\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.928182 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-config\") pod \"dnsmasq-dns-667758f697-spt2d\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.928688 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-ovsdbserver-sb\") pod \"dnsmasq-dns-667758f697-spt2d\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:11 crc kubenswrapper[4810]: I1201 16:11:11.952281 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6n4j\" (UniqueName: \"kubernetes.io/projected/51ae399d-0f9d-4644-8c93-d3e3a0446e47-kube-api-access-g6n4j\") pod \"dnsmasq-dns-667758f697-spt2d\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:12 crc kubenswrapper[4810]: I1201 16:11:12.022828 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:12 crc kubenswrapper[4810]: I1201 16:11:12.531117 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-667758f697-spt2d"] Dec 01 16:11:12 crc kubenswrapper[4810]: I1201 16:11:12.575090 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-zcb6q"] Dec 01 16:11:12 crc kubenswrapper[4810]: W1201 16:11:12.575698 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58f5e1dc_2c1b_4bee_b02f_c584327d7cbc.slice/crio-2d36f1a03e67865e2934a1c445cde4d115957f0345b7965590589f272f5fd9e7 WatchSource:0}: Error finding container 2d36f1a03e67865e2934a1c445cde4d115957f0345b7965590589f272f5fd9e7: Status 404 returned error can't find the container with id 2d36f1a03e67865e2934a1c445cde4d115957f0345b7965590589f272f5fd9e7 Dec 01 16:11:13 crc kubenswrapper[4810]: I1201 16:11:13.434181 4810 generic.go:334] "Generic (PLEG): container finished" podID="51ae399d-0f9d-4644-8c93-d3e3a0446e47" containerID="e12e791eb5079652c6fb5db332814d8bffc049a0a255bc677e536bfc4641e702" exitCode=0 Dec 01 16:11:13 crc kubenswrapper[4810]: I1201 16:11:13.434357 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-667758f697-spt2d" event={"ID":"51ae399d-0f9d-4644-8c93-d3e3a0446e47","Type":"ContainerDied","Data":"e12e791eb5079652c6fb5db332814d8bffc049a0a255bc677e536bfc4641e702"} Dec 01 16:11:13 crc kubenswrapper[4810]: I1201 16:11:13.434542 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-667758f697-spt2d" event={"ID":"51ae399d-0f9d-4644-8c93-d3e3a0446e47","Type":"ContainerStarted","Data":"5a160d68b049723c7d16369f6e31516beac91ba4c7482c043d9fe8ab5edc2740"} Dec 01 16:11:13 crc kubenswrapper[4810]: I1201 16:11:13.436971 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zcb6q" event={"ID":"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc","Type":"ContainerStarted","Data":"bdfbfcce0958fa9998fe9b61fa53ec6756c93ac40498acc9a0397631abb51364"} Dec 01 16:11:13 crc kubenswrapper[4810]: I1201 16:11:13.437019 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zcb6q" event={"ID":"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc","Type":"ContainerStarted","Data":"2d36f1a03e67865e2934a1c445cde4d115957f0345b7965590589f272f5fd9e7"} Dec 01 16:11:13 crc kubenswrapper[4810]: I1201 16:11:13.487112 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-zcb6q" podStartSLOduration=2.487091877 podStartE2EDuration="2.487091877s" podCreationTimestamp="2025-12-01 16:11:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:11:13.484208749 +0000 UTC m=+5839.247718352" watchObservedRunningTime="2025-12-01 16:11:13.487091877 +0000 UTC m=+5839.250601480" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.051844 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7f6569d978-7v5jj"] Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.053897 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.056651 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.070391 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7f6569d978-7v5jj"] Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.169969 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-etc-swift\") pod \"swift-proxy-7f6569d978-7v5jj\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.170027 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-combined-ca-bundle\") pod \"swift-proxy-7f6569d978-7v5jj\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.170083 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlltf\" (UniqueName: \"kubernetes.io/projected/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-kube-api-access-vlltf\") pod \"swift-proxy-7f6569d978-7v5jj\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.170123 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-log-httpd\") pod \"swift-proxy-7f6569d978-7v5jj\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.170168 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-run-httpd\") pod \"swift-proxy-7f6569d978-7v5jj\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.170191 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-config-data\") pod \"swift-proxy-7f6569d978-7v5jj\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.272150 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-combined-ca-bundle\") pod \"swift-proxy-7f6569d978-7v5jj\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.272265 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlltf\" (UniqueName: \"kubernetes.io/projected/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-kube-api-access-vlltf\") pod \"swift-proxy-7f6569d978-7v5jj\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.272372 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-log-httpd\") pod \"swift-proxy-7f6569d978-7v5jj\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.272447 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-run-httpd\") pod \"swift-proxy-7f6569d978-7v5jj\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.272497 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-config-data\") pod \"swift-proxy-7f6569d978-7v5jj\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.272544 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-etc-swift\") pod \"swift-proxy-7f6569d978-7v5jj\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.273352 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-log-httpd\") pod \"swift-proxy-7f6569d978-7v5jj\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.273369 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-run-httpd\") pod \"swift-proxy-7f6569d978-7v5jj\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.277883 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-combined-ca-bundle\") pod \"swift-proxy-7f6569d978-7v5jj\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.278256 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-etc-swift\") pod \"swift-proxy-7f6569d978-7v5jj\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.278793 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-config-data\") pod \"swift-proxy-7f6569d978-7v5jj\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.299377 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlltf\" (UniqueName: \"kubernetes.io/projected/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-kube-api-access-vlltf\") pod \"swift-proxy-7f6569d978-7v5jj\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.372422 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.445922 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-667758f697-spt2d" event={"ID":"51ae399d-0f9d-4644-8c93-d3e3a0446e47","Type":"ContainerStarted","Data":"c2bb3483fe18662ea5bae4c150bf2664a3f69c896f0b0654b9a4e39c381eb3c3"} Dec 01 16:11:14 crc kubenswrapper[4810]: I1201 16:11:14.476883 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-667758f697-spt2d" podStartSLOduration=3.476857304 podStartE2EDuration="3.476857304s" podCreationTimestamp="2025-12-01 16:11:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:11:14.464428936 +0000 UTC m=+5840.227938549" watchObservedRunningTime="2025-12-01 16:11:14.476857304 +0000 UTC m=+5840.240366907" Dec 01 16:11:15 crc kubenswrapper[4810]: I1201 16:11:15.125059 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7f6569d978-7v5jj"] Dec 01 16:11:15 crc kubenswrapper[4810]: I1201 16:11:15.454833 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7f6569d978-7v5jj" event={"ID":"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54","Type":"ContainerStarted","Data":"c6be535f8192f6d7b973867cf0fd2e46fb98d3297a6cf8ec785e44aae4cba1e6"} Dec 01 16:11:15 crc kubenswrapper[4810]: I1201 16:11:15.454882 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7f6569d978-7v5jj" event={"ID":"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54","Type":"ContainerStarted","Data":"233e7d222bd0132507b30122b7f599a4550c92bb6a3a55b22e6a6a5a954a3383"} Dec 01 16:11:15 crc kubenswrapper[4810]: I1201 16:11:15.454941 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:16 crc kubenswrapper[4810]: I1201 16:11:16.463738 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7f6569d978-7v5jj" event={"ID":"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54","Type":"ContainerStarted","Data":"a9ccc903529390bc2ba5cdda6d6e4b0fb29e6f47291bb54baf00ec6c2c910696"} Dec 01 16:11:16 crc kubenswrapper[4810]: I1201 16:11:16.464076 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:16 crc kubenswrapper[4810]: I1201 16:11:16.489643 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7f6569d978-7v5jj" podStartSLOduration=2.48962665 podStartE2EDuration="2.48962665s" podCreationTimestamp="2025-12-01 16:11:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:11:16.483923025 +0000 UTC m=+5842.247432638" watchObservedRunningTime="2025-12-01 16:11:16.48962665 +0000 UTC m=+5842.253136253" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.200553 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-79646c549d-jxbn6"] Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.202669 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.206902 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.209265 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.225980 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-79646c549d-jxbn6"] Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.324068 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9c91a11-9c09-49e7-b228-b37eddc05cee-combined-ca-bundle\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.324116 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9c91a11-9c09-49e7-b228-b37eddc05cee-internal-tls-certs\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.324137 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dhjf\" (UniqueName: \"kubernetes.io/projected/e9c91a11-9c09-49e7-b228-b37eddc05cee-kube-api-access-8dhjf\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.324308 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9c91a11-9c09-49e7-b228-b37eddc05cee-run-httpd\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.324365 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9c91a11-9c09-49e7-b228-b37eddc05cee-config-data\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.324457 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9c91a11-9c09-49e7-b228-b37eddc05cee-log-httpd\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.324887 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9c91a11-9c09-49e7-b228-b37eddc05cee-public-tls-certs\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.324922 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e9c91a11-9c09-49e7-b228-b37eddc05cee-etc-swift\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.426904 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9c91a11-9c09-49e7-b228-b37eddc05cee-log-httpd\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.427060 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9c91a11-9c09-49e7-b228-b37eddc05cee-public-tls-certs\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.427081 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e9c91a11-9c09-49e7-b228-b37eddc05cee-etc-swift\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.427111 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9c91a11-9c09-49e7-b228-b37eddc05cee-combined-ca-bundle\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.427130 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9c91a11-9c09-49e7-b228-b37eddc05cee-internal-tls-certs\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.427147 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dhjf\" (UniqueName: \"kubernetes.io/projected/e9c91a11-9c09-49e7-b228-b37eddc05cee-kube-api-access-8dhjf\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.427169 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9c91a11-9c09-49e7-b228-b37eddc05cee-run-httpd\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.427186 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9c91a11-9c09-49e7-b228-b37eddc05cee-config-data\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.427458 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9c91a11-9c09-49e7-b228-b37eddc05cee-log-httpd\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.427913 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e9c91a11-9c09-49e7-b228-b37eddc05cee-run-httpd\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.433039 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9c91a11-9c09-49e7-b228-b37eddc05cee-internal-tls-certs\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.433290 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9c91a11-9c09-49e7-b228-b37eddc05cee-config-data\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.433851 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9c91a11-9c09-49e7-b228-b37eddc05cee-public-tls-certs\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.435687 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e9c91a11-9c09-49e7-b228-b37eddc05cee-etc-swift\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.442078 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9c91a11-9c09-49e7-b228-b37eddc05cee-combined-ca-bundle\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.453314 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dhjf\" (UniqueName: \"kubernetes.io/projected/e9c91a11-9c09-49e7-b228-b37eddc05cee-kube-api-access-8dhjf\") pod \"swift-proxy-79646c549d-jxbn6\" (UID: \"e9c91a11-9c09-49e7-b228-b37eddc05cee\") " pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.477830 4810 generic.go:334] "Generic (PLEG): container finished" podID="58f5e1dc-2c1b-4bee-b02f-c584327d7cbc" containerID="bdfbfcce0958fa9998fe9b61fa53ec6756c93ac40498acc9a0397631abb51364" exitCode=0 Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.477900 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zcb6q" event={"ID":"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc","Type":"ContainerDied","Data":"bdfbfcce0958fa9998fe9b61fa53ec6756c93ac40498acc9a0397631abb51364"} Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.478206 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:17 crc kubenswrapper[4810]: I1201 16:11:17.521921 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:18 crc kubenswrapper[4810]: I1201 16:11:18.149806 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-79646c549d-jxbn6"] Dec 01 16:11:18 crc kubenswrapper[4810]: I1201 16:11:18.516986 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-79646c549d-jxbn6" event={"ID":"e9c91a11-9c09-49e7-b228-b37eddc05cee","Type":"ContainerStarted","Data":"2ffc7152d2d0e36224d2f2bb44615fdf2d6b4511939e828f33da228893f5eb26"} Dec 01 16:11:18 crc kubenswrapper[4810]: I1201 16:11:18.517030 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-79646c549d-jxbn6" event={"ID":"e9c91a11-9c09-49e7-b228-b37eddc05cee","Type":"ContainerStarted","Data":"edeb302a13647b7d104ba2cbfb1ba4222d1d7a69badb40a8154982cdd077e473"} Dec 01 16:11:18 crc kubenswrapper[4810]: I1201 16:11:18.840673 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:18 crc kubenswrapper[4810]: I1201 16:11:18.952148 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-combined-ca-bundle\") pod \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " Dec 01 16:11:18 crc kubenswrapper[4810]: I1201 16:11:18.952225 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-scripts\") pod \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " Dec 01 16:11:18 crc kubenswrapper[4810]: I1201 16:11:18.952297 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-dispersionconf\") pod \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " Dec 01 16:11:18 crc kubenswrapper[4810]: I1201 16:11:18.952367 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwgnp\" (UniqueName: \"kubernetes.io/projected/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-kube-api-access-rwgnp\") pod \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " Dec 01 16:11:18 crc kubenswrapper[4810]: I1201 16:11:18.952405 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-etc-swift\") pod \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " Dec 01 16:11:18 crc kubenswrapper[4810]: I1201 16:11:18.952509 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-ring-data-devices\") pod \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " Dec 01 16:11:18 crc kubenswrapper[4810]: I1201 16:11:18.952557 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-swiftconf\") pod \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\" (UID: \"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc\") " Dec 01 16:11:18 crc kubenswrapper[4810]: I1201 16:11:18.953297 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "58f5e1dc-2c1b-4bee-b02f-c584327d7cbc" (UID: "58f5e1dc-2c1b-4bee-b02f-c584327d7cbc"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:11:18 crc kubenswrapper[4810]: I1201 16:11:18.953580 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "58f5e1dc-2c1b-4bee-b02f-c584327d7cbc" (UID: "58f5e1dc-2c1b-4bee-b02f-c584327d7cbc"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:11:18 crc kubenswrapper[4810]: I1201 16:11:18.957256 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-kube-api-access-rwgnp" (OuterVolumeSpecName: "kube-api-access-rwgnp") pod "58f5e1dc-2c1b-4bee-b02f-c584327d7cbc" (UID: "58f5e1dc-2c1b-4bee-b02f-c584327d7cbc"). InnerVolumeSpecName "kube-api-access-rwgnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:11:18 crc kubenswrapper[4810]: I1201 16:11:18.959713 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "58f5e1dc-2c1b-4bee-b02f-c584327d7cbc" (UID: "58f5e1dc-2c1b-4bee-b02f-c584327d7cbc"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:11:18 crc kubenswrapper[4810]: I1201 16:11:18.979724 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58f5e1dc-2c1b-4bee-b02f-c584327d7cbc" (UID: "58f5e1dc-2c1b-4bee-b02f-c584327d7cbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:11:18 crc kubenswrapper[4810]: I1201 16:11:18.982216 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "58f5e1dc-2c1b-4bee-b02f-c584327d7cbc" (UID: "58f5e1dc-2c1b-4bee-b02f-c584327d7cbc"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:11:18 crc kubenswrapper[4810]: I1201 16:11:18.982598 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-scripts" (OuterVolumeSpecName: "scripts") pod "58f5e1dc-2c1b-4bee-b02f-c584327d7cbc" (UID: "58f5e1dc-2c1b-4bee-b02f-c584327d7cbc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:11:19 crc kubenswrapper[4810]: I1201 16:11:19.054714 4810 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:19 crc kubenswrapper[4810]: I1201 16:11:19.054747 4810 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:19 crc kubenswrapper[4810]: I1201 16:11:19.054756 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:19 crc kubenswrapper[4810]: I1201 16:11:19.054764 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:19 crc kubenswrapper[4810]: I1201 16:11:19.054772 4810 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:19 crc kubenswrapper[4810]: I1201 16:11:19.054781 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwgnp\" (UniqueName: \"kubernetes.io/projected/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-kube-api-access-rwgnp\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:19 crc kubenswrapper[4810]: I1201 16:11:19.054791 4810 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/58f5e1dc-2c1b-4bee-b02f-c584327d7cbc-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:19 crc kubenswrapper[4810]: I1201 16:11:19.501594 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-zcb6q" event={"ID":"58f5e1dc-2c1b-4bee-b02f-c584327d7cbc","Type":"ContainerDied","Data":"2d36f1a03e67865e2934a1c445cde4d115957f0345b7965590589f272f5fd9e7"} Dec 01 16:11:19 crc kubenswrapper[4810]: I1201 16:11:19.501873 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d36f1a03e67865e2934a1c445cde4d115957f0345b7965590589f272f5fd9e7" Dec 01 16:11:19 crc kubenswrapper[4810]: I1201 16:11:19.501942 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-zcb6q" Dec 01 16:11:19 crc kubenswrapper[4810]: I1201 16:11:19.509988 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-79646c549d-jxbn6" event={"ID":"e9c91a11-9c09-49e7-b228-b37eddc05cee","Type":"ContainerStarted","Data":"0e3e4cddeeb872fa4a6b09337b5084a27dc1d1071d30a5fd2b3330a94247a116"} Dec 01 16:11:19 crc kubenswrapper[4810]: I1201 16:11:19.511488 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:19 crc kubenswrapper[4810]: I1201 16:11:19.511535 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:19 crc kubenswrapper[4810]: I1201 16:11:19.561435 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-79646c549d-jxbn6" podStartSLOduration=2.561412785 podStartE2EDuration="2.561412785s" podCreationTimestamp="2025-12-01 16:11:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:11:19.544982579 +0000 UTC m=+5845.308492172" watchObservedRunningTime="2025-12-01 16:11:19.561412785 +0000 UTC m=+5845.324922388" Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.025418 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.087643 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59596df985-rf9x5"] Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.087877 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59596df985-rf9x5" podUID="3b855aff-4523-4dba-8c45-341e7c9807bc" containerName="dnsmasq-dns" containerID="cri-o://e30ac6813835f238af5ecd4c2ae2a64754ebabc796d14bb5c6d179b184bdfc67" gracePeriod=10 Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.546498 4810 generic.go:334] "Generic (PLEG): container finished" podID="3b855aff-4523-4dba-8c45-341e7c9807bc" containerID="e30ac6813835f238af5ecd4c2ae2a64754ebabc796d14bb5c6d179b184bdfc67" exitCode=0 Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.546909 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59596df985-rf9x5" event={"ID":"3b855aff-4523-4dba-8c45-341e7c9807bc","Type":"ContainerDied","Data":"e30ac6813835f238af5ecd4c2ae2a64754ebabc796d14bb5c6d179b184bdfc67"} Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.546942 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59596df985-rf9x5" event={"ID":"3b855aff-4523-4dba-8c45-341e7c9807bc","Type":"ContainerDied","Data":"e4eb764fb9a76a7c4422900028d0a1b3092852d86e40764f21606ff155597072"} Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.546962 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4eb764fb9a76a7c4422900028d0a1b3092852d86e40764f21606ff155597072" Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.594961 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.633816 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-config\") pod \"3b855aff-4523-4dba-8c45-341e7c9807bc\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.634005 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jltr\" (UniqueName: \"kubernetes.io/projected/3b855aff-4523-4dba-8c45-341e7c9807bc-kube-api-access-6jltr\") pod \"3b855aff-4523-4dba-8c45-341e7c9807bc\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.634049 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-ovsdbserver-nb\") pod \"3b855aff-4523-4dba-8c45-341e7c9807bc\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.634074 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-ovsdbserver-sb\") pod \"3b855aff-4523-4dba-8c45-341e7c9807bc\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.634124 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-dns-svc\") pod \"3b855aff-4523-4dba-8c45-341e7c9807bc\" (UID: \"3b855aff-4523-4dba-8c45-341e7c9807bc\") " Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.642321 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b855aff-4523-4dba-8c45-341e7c9807bc-kube-api-access-6jltr" (OuterVolumeSpecName: "kube-api-access-6jltr") pod "3b855aff-4523-4dba-8c45-341e7c9807bc" (UID: "3b855aff-4523-4dba-8c45-341e7c9807bc"). InnerVolumeSpecName "kube-api-access-6jltr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.685704 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3b855aff-4523-4dba-8c45-341e7c9807bc" (UID: "3b855aff-4523-4dba-8c45-341e7c9807bc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.707609 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3b855aff-4523-4dba-8c45-341e7c9807bc" (UID: "3b855aff-4523-4dba-8c45-341e7c9807bc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.733033 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-config" (OuterVolumeSpecName: "config") pod "3b855aff-4523-4dba-8c45-341e7c9807bc" (UID: "3b855aff-4523-4dba-8c45-341e7c9807bc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.737337 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jltr\" (UniqueName: \"kubernetes.io/projected/3b855aff-4523-4dba-8c45-341e7c9807bc-kube-api-access-6jltr\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.737392 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.737411 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.737422 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.740549 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3b855aff-4523-4dba-8c45-341e7c9807bc" (UID: "3b855aff-4523-4dba-8c45-341e7c9807bc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:11:22 crc kubenswrapper[4810]: I1201 16:11:22.839439 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b855aff-4523-4dba-8c45-341e7c9807bc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:23 crc kubenswrapper[4810]: I1201 16:11:23.554003 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59596df985-rf9x5" Dec 01 16:11:23 crc kubenswrapper[4810]: I1201 16:11:23.586848 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59596df985-rf9x5"] Dec 01 16:11:23 crc kubenswrapper[4810]: I1201 16:11:23.602451 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59596df985-rf9x5"] Dec 01 16:11:24 crc kubenswrapper[4810]: I1201 16:11:24.375589 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:24 crc kubenswrapper[4810]: I1201 16:11:24.376714 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:24 crc kubenswrapper[4810]: I1201 16:11:24.503033 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b855aff-4523-4dba-8c45-341e7c9807bc" path="/var/lib/kubelet/pods/3b855aff-4523-4dba-8c45-341e7c9807bc/volumes" Dec 01 16:11:27 crc kubenswrapper[4810]: I1201 16:11:27.529040 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:27 crc kubenswrapper[4810]: I1201 16:11:27.541384 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-79646c549d-jxbn6" Dec 01 16:11:27 crc kubenswrapper[4810]: I1201 16:11:27.627918 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-7f6569d978-7v5jj"] Dec 01 16:11:27 crc kubenswrapper[4810]: I1201 16:11:27.628292 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-7f6569d978-7v5jj" podUID="dd2571e7-5504-4e2f-897b-bcbe5d8c6f54" containerName="proxy-server" containerID="cri-o://a9ccc903529390bc2ba5cdda6d6e4b0fb29e6f47291bb54baf00ec6c2c910696" gracePeriod=30 Dec 01 16:11:27 crc kubenswrapper[4810]: I1201 16:11:27.628250 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-7f6569d978-7v5jj" podUID="dd2571e7-5504-4e2f-897b-bcbe5d8c6f54" containerName="proxy-httpd" containerID="cri-o://c6be535f8192f6d7b973867cf0fd2e46fb98d3297a6cf8ec785e44aae4cba1e6" gracePeriod=30 Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.600729 4810 generic.go:334] "Generic (PLEG): container finished" podID="dd2571e7-5504-4e2f-897b-bcbe5d8c6f54" containerID="a9ccc903529390bc2ba5cdda6d6e4b0fb29e6f47291bb54baf00ec6c2c910696" exitCode=0 Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.601071 4810 generic.go:334] "Generic (PLEG): container finished" podID="dd2571e7-5504-4e2f-897b-bcbe5d8c6f54" containerID="c6be535f8192f6d7b973867cf0fd2e46fb98d3297a6cf8ec785e44aae4cba1e6" exitCode=0 Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.600960 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7f6569d978-7v5jj" event={"ID":"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54","Type":"ContainerDied","Data":"a9ccc903529390bc2ba5cdda6d6e4b0fb29e6f47291bb54baf00ec6c2c910696"} Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.601119 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7f6569d978-7v5jj" event={"ID":"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54","Type":"ContainerDied","Data":"c6be535f8192f6d7b973867cf0fd2e46fb98d3297a6cf8ec785e44aae4cba1e6"} Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.601140 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7f6569d978-7v5jj" event={"ID":"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54","Type":"ContainerDied","Data":"233e7d222bd0132507b30122b7f599a4550c92bb6a3a55b22e6a6a5a954a3383"} Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.601153 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="233e7d222bd0132507b30122b7f599a4550c92bb6a3a55b22e6a6a5a954a3383" Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.660540 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.745076 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlltf\" (UniqueName: \"kubernetes.io/projected/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-kube-api-access-vlltf\") pod \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.745151 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-etc-swift\") pod \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.745251 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-config-data\") pod \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.745303 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-log-httpd\") pod \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.745338 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-run-httpd\") pod \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.745406 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-combined-ca-bundle\") pod \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\" (UID: \"dd2571e7-5504-4e2f-897b-bcbe5d8c6f54\") " Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.746389 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "dd2571e7-5504-4e2f-897b-bcbe5d8c6f54" (UID: "dd2571e7-5504-4e2f-897b-bcbe5d8c6f54"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.746531 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "dd2571e7-5504-4e2f-897b-bcbe5d8c6f54" (UID: "dd2571e7-5504-4e2f-897b-bcbe5d8c6f54"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.751112 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "dd2571e7-5504-4e2f-897b-bcbe5d8c6f54" (UID: "dd2571e7-5504-4e2f-897b-bcbe5d8c6f54"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.754382 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-kube-api-access-vlltf" (OuterVolumeSpecName: "kube-api-access-vlltf") pod "dd2571e7-5504-4e2f-897b-bcbe5d8c6f54" (UID: "dd2571e7-5504-4e2f-897b-bcbe5d8c6f54"). InnerVolumeSpecName "kube-api-access-vlltf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.800099 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-config-data" (OuterVolumeSpecName: "config-data") pod "dd2571e7-5504-4e2f-897b-bcbe5d8c6f54" (UID: "dd2571e7-5504-4e2f-897b-bcbe5d8c6f54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.812289 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd2571e7-5504-4e2f-897b-bcbe5d8c6f54" (UID: "dd2571e7-5504-4e2f-897b-bcbe5d8c6f54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.847393 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.847430 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.847439 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.847449 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.847460 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlltf\" (UniqueName: \"kubernetes.io/projected/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-kube-api-access-vlltf\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:28 crc kubenswrapper[4810]: I1201 16:11:28.847481 4810 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 01 16:11:29 crc kubenswrapper[4810]: I1201 16:11:29.608993 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7f6569d978-7v5jj" Dec 01 16:11:29 crc kubenswrapper[4810]: I1201 16:11:29.645912 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-7f6569d978-7v5jj"] Dec 01 16:11:29 crc kubenswrapper[4810]: I1201 16:11:29.654552 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-7f6569d978-7v5jj"] Dec 01 16:11:30 crc kubenswrapper[4810]: I1201 16:11:30.502143 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd2571e7-5504-4e2f-897b-bcbe5d8c6f54" path="/var/lib/kubelet/pods/dd2571e7-5504-4e2f-897b-bcbe5d8c6f54/volumes" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.335416 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-4cz4n"] Dec 01 16:11:59 crc kubenswrapper[4810]: E1201 16:11:59.336387 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b855aff-4523-4dba-8c45-341e7c9807bc" containerName="init" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.336402 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b855aff-4523-4dba-8c45-341e7c9807bc" containerName="init" Dec 01 16:11:59 crc kubenswrapper[4810]: E1201 16:11:59.336415 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b855aff-4523-4dba-8c45-341e7c9807bc" containerName="dnsmasq-dns" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.336422 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b855aff-4523-4dba-8c45-341e7c9807bc" containerName="dnsmasq-dns" Dec 01 16:11:59 crc kubenswrapper[4810]: E1201 16:11:59.336429 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd2571e7-5504-4e2f-897b-bcbe5d8c6f54" containerName="proxy-httpd" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.336436 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd2571e7-5504-4e2f-897b-bcbe5d8c6f54" containerName="proxy-httpd" Dec 01 16:11:59 crc kubenswrapper[4810]: E1201 16:11:59.336445 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd2571e7-5504-4e2f-897b-bcbe5d8c6f54" containerName="proxy-server" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.336453 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd2571e7-5504-4e2f-897b-bcbe5d8c6f54" containerName="proxy-server" Dec 01 16:11:59 crc kubenswrapper[4810]: E1201 16:11:59.336503 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58f5e1dc-2c1b-4bee-b02f-c584327d7cbc" containerName="swift-ring-rebalance" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.336510 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="58f5e1dc-2c1b-4bee-b02f-c584327d7cbc" containerName="swift-ring-rebalance" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.336672 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd2571e7-5504-4e2f-897b-bcbe5d8c6f54" containerName="proxy-httpd" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.336690 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b855aff-4523-4dba-8c45-341e7c9807bc" containerName="dnsmasq-dns" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.336708 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd2571e7-5504-4e2f-897b-bcbe5d8c6f54" containerName="proxy-server" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.336720 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="58f5e1dc-2c1b-4bee-b02f-c584327d7cbc" containerName="swift-ring-rebalance" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.337350 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4cz4n" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.344212 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-4cz4n"] Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.380005 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-c3d8-account-create-update-zw6k8"] Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.382559 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c3d8-account-create-update-zw6k8" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.385572 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.396102 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-c3d8-account-create-update-zw6k8"] Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.462150 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddpcb\" (UniqueName: \"kubernetes.io/projected/9aaa2816-bc3b-44e6-98a0-f16da83a36f8-kube-api-access-ddpcb\") pod \"cinder-db-create-4cz4n\" (UID: \"9aaa2816-bc3b-44e6-98a0-f16da83a36f8\") " pod="openstack/cinder-db-create-4cz4n" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.462219 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9aaa2816-bc3b-44e6-98a0-f16da83a36f8-operator-scripts\") pod \"cinder-db-create-4cz4n\" (UID: \"9aaa2816-bc3b-44e6-98a0-f16da83a36f8\") " pod="openstack/cinder-db-create-4cz4n" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.564275 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32c0cc0a-d9d0-4874-9e5c-bd5890560011-operator-scripts\") pod \"cinder-c3d8-account-create-update-zw6k8\" (UID: \"32c0cc0a-d9d0-4874-9e5c-bd5890560011\") " pod="openstack/cinder-c3d8-account-create-update-zw6k8" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.564364 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mksdc\" (UniqueName: \"kubernetes.io/projected/32c0cc0a-d9d0-4874-9e5c-bd5890560011-kube-api-access-mksdc\") pod \"cinder-c3d8-account-create-update-zw6k8\" (UID: \"32c0cc0a-d9d0-4874-9e5c-bd5890560011\") " pod="openstack/cinder-c3d8-account-create-update-zw6k8" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.564407 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddpcb\" (UniqueName: \"kubernetes.io/projected/9aaa2816-bc3b-44e6-98a0-f16da83a36f8-kube-api-access-ddpcb\") pod \"cinder-db-create-4cz4n\" (UID: \"9aaa2816-bc3b-44e6-98a0-f16da83a36f8\") " pod="openstack/cinder-db-create-4cz4n" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.564445 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9aaa2816-bc3b-44e6-98a0-f16da83a36f8-operator-scripts\") pod \"cinder-db-create-4cz4n\" (UID: \"9aaa2816-bc3b-44e6-98a0-f16da83a36f8\") " pod="openstack/cinder-db-create-4cz4n" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.565177 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9aaa2816-bc3b-44e6-98a0-f16da83a36f8-operator-scripts\") pod \"cinder-db-create-4cz4n\" (UID: \"9aaa2816-bc3b-44e6-98a0-f16da83a36f8\") " pod="openstack/cinder-db-create-4cz4n" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.589118 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddpcb\" (UniqueName: \"kubernetes.io/projected/9aaa2816-bc3b-44e6-98a0-f16da83a36f8-kube-api-access-ddpcb\") pod \"cinder-db-create-4cz4n\" (UID: \"9aaa2816-bc3b-44e6-98a0-f16da83a36f8\") " pod="openstack/cinder-db-create-4cz4n" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.666399 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mksdc\" (UniqueName: \"kubernetes.io/projected/32c0cc0a-d9d0-4874-9e5c-bd5890560011-kube-api-access-mksdc\") pod \"cinder-c3d8-account-create-update-zw6k8\" (UID: \"32c0cc0a-d9d0-4874-9e5c-bd5890560011\") " pod="openstack/cinder-c3d8-account-create-update-zw6k8" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.666589 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32c0cc0a-d9d0-4874-9e5c-bd5890560011-operator-scripts\") pod \"cinder-c3d8-account-create-update-zw6k8\" (UID: \"32c0cc0a-d9d0-4874-9e5c-bd5890560011\") " pod="openstack/cinder-c3d8-account-create-update-zw6k8" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.667384 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32c0cc0a-d9d0-4874-9e5c-bd5890560011-operator-scripts\") pod \"cinder-c3d8-account-create-update-zw6k8\" (UID: \"32c0cc0a-d9d0-4874-9e5c-bd5890560011\") " pod="openstack/cinder-c3d8-account-create-update-zw6k8" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.667899 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4cz4n" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.688176 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mksdc\" (UniqueName: \"kubernetes.io/projected/32c0cc0a-d9d0-4874-9e5c-bd5890560011-kube-api-access-mksdc\") pod \"cinder-c3d8-account-create-update-zw6k8\" (UID: \"32c0cc0a-d9d0-4874-9e5c-bd5890560011\") " pod="openstack/cinder-c3d8-account-create-update-zw6k8" Dec 01 16:11:59 crc kubenswrapper[4810]: I1201 16:11:59.707916 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c3d8-account-create-update-zw6k8" Dec 01 16:12:00 crc kubenswrapper[4810]: I1201 16:12:00.123339 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-4cz4n"] Dec 01 16:12:00 crc kubenswrapper[4810]: W1201 16:12:00.129827 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9aaa2816_bc3b_44e6_98a0_f16da83a36f8.slice/crio-498f6c6f0b9c9ce58c502d2a0ea892dbe9c61d087adcfb828c407febe41e0d82 WatchSource:0}: Error finding container 498f6c6f0b9c9ce58c502d2a0ea892dbe9c61d087adcfb828c407febe41e0d82: Status 404 returned error can't find the container with id 498f6c6f0b9c9ce58c502d2a0ea892dbe9c61d087adcfb828c407febe41e0d82 Dec 01 16:12:00 crc kubenswrapper[4810]: I1201 16:12:00.188650 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-c3d8-account-create-update-zw6k8"] Dec 01 16:12:00 crc kubenswrapper[4810]: W1201 16:12:00.204718 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32c0cc0a_d9d0_4874_9e5c_bd5890560011.slice/crio-ff6ceae8900ed53c15958bda89039df3e589f4da6ef10c2fd3e80a5f96a93f74 WatchSource:0}: Error finding container ff6ceae8900ed53c15958bda89039df3e589f4da6ef10c2fd3e80a5f96a93f74: Status 404 returned error can't find the container with id ff6ceae8900ed53c15958bda89039df3e589f4da6ef10c2fd3e80a5f96a93f74 Dec 01 16:12:00 crc kubenswrapper[4810]: I1201 16:12:00.892212 4810 generic.go:334] "Generic (PLEG): container finished" podID="9aaa2816-bc3b-44e6-98a0-f16da83a36f8" containerID="6f6b96a27c5e6f72fafb31afd0e93a344d4f49c098de4539403d56ca5eafcf66" exitCode=0 Dec 01 16:12:00 crc kubenswrapper[4810]: I1201 16:12:00.892560 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-4cz4n" event={"ID":"9aaa2816-bc3b-44e6-98a0-f16da83a36f8","Type":"ContainerDied","Data":"6f6b96a27c5e6f72fafb31afd0e93a344d4f49c098de4539403d56ca5eafcf66"} Dec 01 16:12:00 crc kubenswrapper[4810]: I1201 16:12:00.892590 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-4cz4n" event={"ID":"9aaa2816-bc3b-44e6-98a0-f16da83a36f8","Type":"ContainerStarted","Data":"498f6c6f0b9c9ce58c502d2a0ea892dbe9c61d087adcfb828c407febe41e0d82"} Dec 01 16:12:00 crc kubenswrapper[4810]: I1201 16:12:00.894271 4810 generic.go:334] "Generic (PLEG): container finished" podID="32c0cc0a-d9d0-4874-9e5c-bd5890560011" containerID="946060ab18c96a9ed5cc43bf6050a5344505fca618bd8e710c6b6ca16190677b" exitCode=0 Dec 01 16:12:00 crc kubenswrapper[4810]: I1201 16:12:00.894298 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c3d8-account-create-update-zw6k8" event={"ID":"32c0cc0a-d9d0-4874-9e5c-bd5890560011","Type":"ContainerDied","Data":"946060ab18c96a9ed5cc43bf6050a5344505fca618bd8e710c6b6ca16190677b"} Dec 01 16:12:00 crc kubenswrapper[4810]: I1201 16:12:00.894312 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c3d8-account-create-update-zw6k8" event={"ID":"32c0cc0a-d9d0-4874-9e5c-bd5890560011","Type":"ContainerStarted","Data":"ff6ceae8900ed53c15958bda89039df3e589f4da6ef10c2fd3e80a5f96a93f74"} Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.299674 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4cz4n" Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.306361 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c3d8-account-create-update-zw6k8" Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.409985 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32c0cc0a-d9d0-4874-9e5c-bd5890560011-operator-scripts\") pod \"32c0cc0a-d9d0-4874-9e5c-bd5890560011\" (UID: \"32c0cc0a-d9d0-4874-9e5c-bd5890560011\") " Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.410101 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mksdc\" (UniqueName: \"kubernetes.io/projected/32c0cc0a-d9d0-4874-9e5c-bd5890560011-kube-api-access-mksdc\") pod \"32c0cc0a-d9d0-4874-9e5c-bd5890560011\" (UID: \"32c0cc0a-d9d0-4874-9e5c-bd5890560011\") " Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.410191 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddpcb\" (UniqueName: \"kubernetes.io/projected/9aaa2816-bc3b-44e6-98a0-f16da83a36f8-kube-api-access-ddpcb\") pod \"9aaa2816-bc3b-44e6-98a0-f16da83a36f8\" (UID: \"9aaa2816-bc3b-44e6-98a0-f16da83a36f8\") " Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.410211 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9aaa2816-bc3b-44e6-98a0-f16da83a36f8-operator-scripts\") pod \"9aaa2816-bc3b-44e6-98a0-f16da83a36f8\" (UID: \"9aaa2816-bc3b-44e6-98a0-f16da83a36f8\") " Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.410549 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32c0cc0a-d9d0-4874-9e5c-bd5890560011-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "32c0cc0a-d9d0-4874-9e5c-bd5890560011" (UID: "32c0cc0a-d9d0-4874-9e5c-bd5890560011"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.410856 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aaa2816-bc3b-44e6-98a0-f16da83a36f8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9aaa2816-bc3b-44e6-98a0-f16da83a36f8" (UID: "9aaa2816-bc3b-44e6-98a0-f16da83a36f8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.415310 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32c0cc0a-d9d0-4874-9e5c-bd5890560011-kube-api-access-mksdc" (OuterVolumeSpecName: "kube-api-access-mksdc") pod "32c0cc0a-d9d0-4874-9e5c-bd5890560011" (UID: "32c0cc0a-d9d0-4874-9e5c-bd5890560011"). InnerVolumeSpecName "kube-api-access-mksdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.416017 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aaa2816-bc3b-44e6-98a0-f16da83a36f8-kube-api-access-ddpcb" (OuterVolumeSpecName: "kube-api-access-ddpcb") pod "9aaa2816-bc3b-44e6-98a0-f16da83a36f8" (UID: "9aaa2816-bc3b-44e6-98a0-f16da83a36f8"). InnerVolumeSpecName "kube-api-access-ddpcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.513005 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32c0cc0a-d9d0-4874-9e5c-bd5890560011-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.513079 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mksdc\" (UniqueName: \"kubernetes.io/projected/32c0cc0a-d9d0-4874-9e5c-bd5890560011-kube-api-access-mksdc\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.513115 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddpcb\" (UniqueName: \"kubernetes.io/projected/9aaa2816-bc3b-44e6-98a0-f16da83a36f8-kube-api-access-ddpcb\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.513135 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9aaa2816-bc3b-44e6-98a0-f16da83a36f8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.909795 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-4cz4n" event={"ID":"9aaa2816-bc3b-44e6-98a0-f16da83a36f8","Type":"ContainerDied","Data":"498f6c6f0b9c9ce58c502d2a0ea892dbe9c61d087adcfb828c407febe41e0d82"} Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.909840 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="498f6c6f0b9c9ce58c502d2a0ea892dbe9c61d087adcfb828c407febe41e0d82" Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.909818 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4cz4n" Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.911452 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c3d8-account-create-update-zw6k8" event={"ID":"32c0cc0a-d9d0-4874-9e5c-bd5890560011","Type":"ContainerDied","Data":"ff6ceae8900ed53c15958bda89039df3e589f4da6ef10c2fd3e80a5f96a93f74"} Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.911634 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff6ceae8900ed53c15958bda89039df3e589f4da6ef10c2fd3e80a5f96a93f74" Dec 01 16:12:02 crc kubenswrapper[4810]: I1201 16:12:02.911560 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c3d8-account-create-update-zw6k8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.344058 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9j77z"] Dec 01 16:12:04 crc kubenswrapper[4810]: E1201 16:12:04.344926 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aaa2816-bc3b-44e6-98a0-f16da83a36f8" containerName="mariadb-database-create" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.344947 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aaa2816-bc3b-44e6-98a0-f16da83a36f8" containerName="mariadb-database-create" Dec 01 16:12:04 crc kubenswrapper[4810]: E1201 16:12:04.344962 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32c0cc0a-d9d0-4874-9e5c-bd5890560011" containerName="mariadb-account-create-update" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.344971 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="32c0cc0a-d9d0-4874-9e5c-bd5890560011" containerName="mariadb-account-create-update" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.345197 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="32c0cc0a-d9d0-4874-9e5c-bd5890560011" containerName="mariadb-account-create-update" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.345230 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aaa2816-bc3b-44e6-98a0-f16da83a36f8" containerName="mariadb-database-create" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.347570 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9j77z" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.359560 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9j77z"] Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.444540 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6076eb68-acf2-4264-a1ef-549cb5137a50-catalog-content\") pod \"redhat-marketplace-9j77z\" (UID: \"6076eb68-acf2-4264-a1ef-549cb5137a50\") " pod="openshift-marketplace/redhat-marketplace-9j77z" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.444606 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htvbp\" (UniqueName: \"kubernetes.io/projected/6076eb68-acf2-4264-a1ef-549cb5137a50-kube-api-access-htvbp\") pod \"redhat-marketplace-9j77z\" (UID: \"6076eb68-acf2-4264-a1ef-549cb5137a50\") " pod="openshift-marketplace/redhat-marketplace-9j77z" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.444644 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6076eb68-acf2-4264-a1ef-549cb5137a50-utilities\") pod \"redhat-marketplace-9j77z\" (UID: \"6076eb68-acf2-4264-a1ef-549cb5137a50\") " pod="openshift-marketplace/redhat-marketplace-9j77z" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.546608 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6076eb68-acf2-4264-a1ef-549cb5137a50-catalog-content\") pod \"redhat-marketplace-9j77z\" (UID: \"6076eb68-acf2-4264-a1ef-549cb5137a50\") " pod="openshift-marketplace/redhat-marketplace-9j77z" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.546673 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htvbp\" (UniqueName: \"kubernetes.io/projected/6076eb68-acf2-4264-a1ef-549cb5137a50-kube-api-access-htvbp\") pod \"redhat-marketplace-9j77z\" (UID: \"6076eb68-acf2-4264-a1ef-549cb5137a50\") " pod="openshift-marketplace/redhat-marketplace-9j77z" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.546716 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6076eb68-acf2-4264-a1ef-549cb5137a50-utilities\") pod \"redhat-marketplace-9j77z\" (UID: \"6076eb68-acf2-4264-a1ef-549cb5137a50\") " pod="openshift-marketplace/redhat-marketplace-9j77z" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.547304 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6076eb68-acf2-4264-a1ef-549cb5137a50-catalog-content\") pod \"redhat-marketplace-9j77z\" (UID: \"6076eb68-acf2-4264-a1ef-549cb5137a50\") " pod="openshift-marketplace/redhat-marketplace-9j77z" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.547318 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6076eb68-acf2-4264-a1ef-549cb5137a50-utilities\") pod \"redhat-marketplace-9j77z\" (UID: \"6076eb68-acf2-4264-a1ef-549cb5137a50\") " pod="openshift-marketplace/redhat-marketplace-9j77z" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.574310 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htvbp\" (UniqueName: \"kubernetes.io/projected/6076eb68-acf2-4264-a1ef-549cb5137a50-kube-api-access-htvbp\") pod \"redhat-marketplace-9j77z\" (UID: \"6076eb68-acf2-4264-a1ef-549cb5137a50\") " pod="openshift-marketplace/redhat-marketplace-9j77z" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.632915 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-45dk8"] Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.634002 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.637003 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-dvwkv" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.637287 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.641304 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.646906 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-45dk8"] Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.669603 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9j77z" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.748996 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-config-data\") pod \"cinder-db-sync-45dk8\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.749048 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-combined-ca-bundle\") pod \"cinder-db-sync-45dk8\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.749072 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eac202a1-bd74-43cf-841c-9b63602c1c24-etc-machine-id\") pod \"cinder-db-sync-45dk8\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.749096 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-scripts\") pod \"cinder-db-sync-45dk8\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.749139 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqshf\" (UniqueName: \"kubernetes.io/projected/eac202a1-bd74-43cf-841c-9b63602c1c24-kube-api-access-xqshf\") pod \"cinder-db-sync-45dk8\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.749167 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-db-sync-config-data\") pod \"cinder-db-sync-45dk8\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.851892 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-db-sync-config-data\") pod \"cinder-db-sync-45dk8\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.852320 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-config-data\") pod \"cinder-db-sync-45dk8\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.852392 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-combined-ca-bundle\") pod \"cinder-db-sync-45dk8\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.852432 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eac202a1-bd74-43cf-841c-9b63602c1c24-etc-machine-id\") pod \"cinder-db-sync-45dk8\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.852493 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-scripts\") pod \"cinder-db-sync-45dk8\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.852606 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqshf\" (UniqueName: \"kubernetes.io/projected/eac202a1-bd74-43cf-841c-9b63602c1c24-kube-api-access-xqshf\") pod \"cinder-db-sync-45dk8\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.852971 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eac202a1-bd74-43cf-841c-9b63602c1c24-etc-machine-id\") pod \"cinder-db-sync-45dk8\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.857288 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-scripts\") pod \"cinder-db-sync-45dk8\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.859534 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-db-sync-config-data\") pod \"cinder-db-sync-45dk8\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.869221 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-config-data\") pod \"cinder-db-sync-45dk8\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.871898 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-combined-ca-bundle\") pod \"cinder-db-sync-45dk8\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.872742 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqshf\" (UniqueName: \"kubernetes.io/projected/eac202a1-bd74-43cf-841c-9b63602c1c24-kube-api-access-xqshf\") pod \"cinder-db-sync-45dk8\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:04 crc kubenswrapper[4810]: I1201 16:12:04.952113 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:05 crc kubenswrapper[4810]: I1201 16:12:05.138074 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9j77z"] Dec 01 16:12:05 crc kubenswrapper[4810]: W1201 16:12:05.145504 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6076eb68_acf2_4264_a1ef_549cb5137a50.slice/crio-7952f4fbf6002079b3d6f0a4f98bc99bdad575d973e6fd4ff1a1f0a84eaa459f WatchSource:0}: Error finding container 7952f4fbf6002079b3d6f0a4f98bc99bdad575d973e6fd4ff1a1f0a84eaa459f: Status 404 returned error can't find the container with id 7952f4fbf6002079b3d6f0a4f98bc99bdad575d973e6fd4ff1a1f0a84eaa459f Dec 01 16:12:05 crc kubenswrapper[4810]: I1201 16:12:05.374970 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-45dk8"] Dec 01 16:12:05 crc kubenswrapper[4810]: I1201 16:12:05.944642 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-45dk8" event={"ID":"eac202a1-bd74-43cf-841c-9b63602c1c24","Type":"ContainerStarted","Data":"f98456462717d9ca893cd8ec8685e17043a3d1a8de5ca47ee03ad3253cdab906"} Dec 01 16:12:05 crc kubenswrapper[4810]: I1201 16:12:05.944690 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-45dk8" event={"ID":"eac202a1-bd74-43cf-841c-9b63602c1c24","Type":"ContainerStarted","Data":"969a1155009342456c12875dfeaf1c78722452d1d56478d19908b0b38ea33bd9"} Dec 01 16:12:05 crc kubenswrapper[4810]: I1201 16:12:05.946552 4810 generic.go:334] "Generic (PLEG): container finished" podID="6076eb68-acf2-4264-a1ef-549cb5137a50" containerID="f173e9847dbbb2ac329dfa10e173b73ff6c361cbc699af7c2f5eb9df81325b19" exitCode=0 Dec 01 16:12:05 crc kubenswrapper[4810]: I1201 16:12:05.946609 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9j77z" event={"ID":"6076eb68-acf2-4264-a1ef-549cb5137a50","Type":"ContainerDied","Data":"f173e9847dbbb2ac329dfa10e173b73ff6c361cbc699af7c2f5eb9df81325b19"} Dec 01 16:12:05 crc kubenswrapper[4810]: I1201 16:12:05.946642 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9j77z" event={"ID":"6076eb68-acf2-4264-a1ef-549cb5137a50","Type":"ContainerStarted","Data":"7952f4fbf6002079b3d6f0a4f98bc99bdad575d973e6fd4ff1a1f0a84eaa459f"} Dec 01 16:12:05 crc kubenswrapper[4810]: I1201 16:12:05.950676 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:12:05 crc kubenswrapper[4810]: I1201 16:12:05.968786 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-45dk8" podStartSLOduration=1.968767726 podStartE2EDuration="1.968767726s" podCreationTimestamp="2025-12-01 16:12:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:12:05.963368959 +0000 UTC m=+5891.726878582" watchObservedRunningTime="2025-12-01 16:12:05.968767726 +0000 UTC m=+5891.732277329" Dec 01 16:12:06 crc kubenswrapper[4810]: I1201 16:12:06.958355 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9j77z" event={"ID":"6076eb68-acf2-4264-a1ef-549cb5137a50","Type":"ContainerStarted","Data":"039d67f6e3c92d7f1442a5c034338d3bef2346f41cd5000c341e836960163c50"} Dec 01 16:12:07 crc kubenswrapper[4810]: I1201 16:12:07.976368 4810 generic.go:334] "Generic (PLEG): container finished" podID="6076eb68-acf2-4264-a1ef-549cb5137a50" containerID="039d67f6e3c92d7f1442a5c034338d3bef2346f41cd5000c341e836960163c50" exitCode=0 Dec 01 16:12:07 crc kubenswrapper[4810]: I1201 16:12:07.976449 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9j77z" event={"ID":"6076eb68-acf2-4264-a1ef-549cb5137a50","Type":"ContainerDied","Data":"039d67f6e3c92d7f1442a5c034338d3bef2346f41cd5000c341e836960163c50"} Dec 01 16:12:08 crc kubenswrapper[4810]: I1201 16:12:08.985713 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9j77z" event={"ID":"6076eb68-acf2-4264-a1ef-549cb5137a50","Type":"ContainerStarted","Data":"712394811bb983e233135b2770367451ed6736d169c1f785eb059008d8c7b021"} Dec 01 16:12:08 crc kubenswrapper[4810]: I1201 16:12:08.987832 4810 generic.go:334] "Generic (PLEG): container finished" podID="eac202a1-bd74-43cf-841c-9b63602c1c24" containerID="f98456462717d9ca893cd8ec8685e17043a3d1a8de5ca47ee03ad3253cdab906" exitCode=0 Dec 01 16:12:08 crc kubenswrapper[4810]: I1201 16:12:08.987867 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-45dk8" event={"ID":"eac202a1-bd74-43cf-841c-9b63602c1c24","Type":"ContainerDied","Data":"f98456462717d9ca893cd8ec8685e17043a3d1a8de5ca47ee03ad3253cdab906"} Dec 01 16:12:09 crc kubenswrapper[4810]: I1201 16:12:09.009302 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9j77z" podStartSLOduration=2.514725443 podStartE2EDuration="5.009283182s" podCreationTimestamp="2025-12-01 16:12:04 +0000 UTC" firstStartedPulling="2025-12-01 16:12:05.950325435 +0000 UTC m=+5891.713835058" lastFinishedPulling="2025-12-01 16:12:08.444883184 +0000 UTC m=+5894.208392797" observedRunningTime="2025-12-01 16:12:09.00407637 +0000 UTC m=+5894.767585983" watchObservedRunningTime="2025-12-01 16:12:09.009283182 +0000 UTC m=+5894.772792805" Dec 01 16:12:09 crc kubenswrapper[4810]: I1201 16:12:09.557844 4810 scope.go:117] "RemoveContainer" containerID="3f95994d6c1e8562e5f518ad159a1d053739f1f742b61c865a4bab6d18909ca7" Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.281242 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.446442 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-config-data\") pod \"eac202a1-bd74-43cf-841c-9b63602c1c24\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.446593 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-db-sync-config-data\") pod \"eac202a1-bd74-43cf-841c-9b63602c1c24\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.446635 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqshf\" (UniqueName: \"kubernetes.io/projected/eac202a1-bd74-43cf-841c-9b63602c1c24-kube-api-access-xqshf\") pod \"eac202a1-bd74-43cf-841c-9b63602c1c24\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.446712 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-scripts\") pod \"eac202a1-bd74-43cf-841c-9b63602c1c24\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.446806 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eac202a1-bd74-43cf-841c-9b63602c1c24-etc-machine-id\") pod \"eac202a1-bd74-43cf-841c-9b63602c1c24\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.446832 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-combined-ca-bundle\") pod \"eac202a1-bd74-43cf-841c-9b63602c1c24\" (UID: \"eac202a1-bd74-43cf-841c-9b63602c1c24\") " Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.446886 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eac202a1-bd74-43cf-841c-9b63602c1c24-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "eac202a1-bd74-43cf-841c-9b63602c1c24" (UID: "eac202a1-bd74-43cf-841c-9b63602c1c24"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.447173 4810 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eac202a1-bd74-43cf-841c-9b63602c1c24-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.452100 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-scripts" (OuterVolumeSpecName: "scripts") pod "eac202a1-bd74-43cf-841c-9b63602c1c24" (UID: "eac202a1-bd74-43cf-841c-9b63602c1c24"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.452384 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "eac202a1-bd74-43cf-841c-9b63602c1c24" (UID: "eac202a1-bd74-43cf-841c-9b63602c1c24"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.453112 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eac202a1-bd74-43cf-841c-9b63602c1c24-kube-api-access-xqshf" (OuterVolumeSpecName: "kube-api-access-xqshf") pod "eac202a1-bd74-43cf-841c-9b63602c1c24" (UID: "eac202a1-bd74-43cf-841c-9b63602c1c24"). InnerVolumeSpecName "kube-api-access-xqshf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.471570 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eac202a1-bd74-43cf-841c-9b63602c1c24" (UID: "eac202a1-bd74-43cf-841c-9b63602c1c24"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.522924 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-config-data" (OuterVolumeSpecName: "config-data") pod "eac202a1-bd74-43cf-841c-9b63602c1c24" (UID: "eac202a1-bd74-43cf-841c-9b63602c1c24"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.549314 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.549363 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.549376 4810 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.549385 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqshf\" (UniqueName: \"kubernetes.io/projected/eac202a1-bd74-43cf-841c-9b63602c1c24-kube-api-access-xqshf\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:10 crc kubenswrapper[4810]: I1201 16:12:10.549395 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eac202a1-bd74-43cf-841c-9b63602c1c24-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.004389 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-45dk8" event={"ID":"eac202a1-bd74-43cf-841c-9b63602c1c24","Type":"ContainerDied","Data":"969a1155009342456c12875dfeaf1c78722452d1d56478d19908b0b38ea33bd9"} Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.004443 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-45dk8" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.004455 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="969a1155009342456c12875dfeaf1c78722452d1d56478d19908b0b38ea33bd9" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.335939 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-655d5b56f-qjvr4"] Dec 01 16:12:11 crc kubenswrapper[4810]: E1201 16:12:11.339509 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eac202a1-bd74-43cf-841c-9b63602c1c24" containerName="cinder-db-sync" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.339687 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="eac202a1-bd74-43cf-841c-9b63602c1c24" containerName="cinder-db-sync" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.339967 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="eac202a1-bd74-43cf-841c-9b63602c1c24" containerName="cinder-db-sync" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.340946 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.361683 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-655d5b56f-qjvr4"] Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.477195 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-ovsdbserver-sb\") pod \"dnsmasq-dns-655d5b56f-qjvr4\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.477516 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-dns-svc\") pod \"dnsmasq-dns-655d5b56f-qjvr4\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.477551 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6xcw\" (UniqueName: \"kubernetes.io/projected/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-kube-api-access-r6xcw\") pod \"dnsmasq-dns-655d5b56f-qjvr4\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.477706 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-ovsdbserver-nb\") pod \"dnsmasq-dns-655d5b56f-qjvr4\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.477756 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-config\") pod \"dnsmasq-dns-655d5b56f-qjvr4\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.510283 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.516147 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.520555 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.521294 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-dvwkv" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.521326 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.521329 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.525902 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.579668 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-config\") pod \"dnsmasq-dns-655d5b56f-qjvr4\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.579764 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-ovsdbserver-sb\") pod \"dnsmasq-dns-655d5b56f-qjvr4\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.579812 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-dns-svc\") pod \"dnsmasq-dns-655d5b56f-qjvr4\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.579831 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6xcw\" (UniqueName: \"kubernetes.io/projected/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-kube-api-access-r6xcw\") pod \"dnsmasq-dns-655d5b56f-qjvr4\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.579949 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-ovsdbserver-nb\") pod \"dnsmasq-dns-655d5b56f-qjvr4\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.580888 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-ovsdbserver-nb\") pod \"dnsmasq-dns-655d5b56f-qjvr4\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.581625 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-config\") pod \"dnsmasq-dns-655d5b56f-qjvr4\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.582116 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-dns-svc\") pod \"dnsmasq-dns-655d5b56f-qjvr4\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.583592 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-ovsdbserver-sb\") pod \"dnsmasq-dns-655d5b56f-qjvr4\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.604324 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6xcw\" (UniqueName: \"kubernetes.io/projected/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-kube-api-access-r6xcw\") pod \"dnsmasq-dns-655d5b56f-qjvr4\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.660806 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.681021 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b48b617-0ac4-4501-948c-3170743e02d9-logs\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.681066 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk4r9\" (UniqueName: \"kubernetes.io/projected/9b48b617-0ac4-4501-948c-3170743e02d9-kube-api-access-hk4r9\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.681112 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.681162 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-scripts\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.681187 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b48b617-0ac4-4501-948c-3170743e02d9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.681286 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-config-data-custom\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.681315 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-config-data\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.782651 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.782702 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-scripts\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.782734 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b48b617-0ac4-4501-948c-3170743e02d9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.782795 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-config-data-custom\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.782828 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-config-data\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.782901 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b48b617-0ac4-4501-948c-3170743e02d9-logs\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.782924 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk4r9\" (UniqueName: \"kubernetes.io/projected/9b48b617-0ac4-4501-948c-3170743e02d9-kube-api-access-hk4r9\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.786316 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b48b617-0ac4-4501-948c-3170743e02d9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.786790 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b48b617-0ac4-4501-948c-3170743e02d9-logs\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.789922 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.791023 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-config-data\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.793622 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-config-data-custom\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.806378 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-scripts\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.807026 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk4r9\" (UniqueName: \"kubernetes.io/projected/9b48b617-0ac4-4501-948c-3170743e02d9-kube-api-access-hk4r9\") pod \"cinder-api-0\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " pod="openstack/cinder-api-0" Dec 01 16:12:11 crc kubenswrapper[4810]: I1201 16:12:11.832048 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 16:12:12 crc kubenswrapper[4810]: I1201 16:12:12.139107 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-655d5b56f-qjvr4"] Dec 01 16:12:12 crc kubenswrapper[4810]: W1201 16:12:12.146786 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d5dc495_73c7_4eaa_b9be_61f8360ae8dc.slice/crio-18a9dfb7b3ef4c2d128cc7ecad4dcd48c04f858f7d19b1b08ddf92f85a348aed WatchSource:0}: Error finding container 18a9dfb7b3ef4c2d128cc7ecad4dcd48c04f858f7d19b1b08ddf92f85a348aed: Status 404 returned error can't find the container with id 18a9dfb7b3ef4c2d128cc7ecad4dcd48c04f858f7d19b1b08ddf92f85a348aed Dec 01 16:12:12 crc kubenswrapper[4810]: I1201 16:12:12.262623 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 16:12:13 crc kubenswrapper[4810]: I1201 16:12:13.057414 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9b48b617-0ac4-4501-948c-3170743e02d9","Type":"ContainerStarted","Data":"56b0ed08fe7a81b890c2925e53af51a1c7fb63b59433bc41c2664f7275e6d850"} Dec 01 16:12:13 crc kubenswrapper[4810]: I1201 16:12:13.057749 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9b48b617-0ac4-4501-948c-3170743e02d9","Type":"ContainerStarted","Data":"aa6f0d3db0101bf564747ae9606b80a50ec677ac22609197b50e202708f2e8a9"} Dec 01 16:12:13 crc kubenswrapper[4810]: I1201 16:12:13.063994 4810 generic.go:334] "Generic (PLEG): container finished" podID="0d5dc495-73c7-4eaa-b9be-61f8360ae8dc" containerID="f21ef54d9cfd076fae9a3da2004aa9e803aedc1288e566da226ece0ed47eae3a" exitCode=0 Dec 01 16:12:13 crc kubenswrapper[4810]: I1201 16:12:13.064042 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" event={"ID":"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc","Type":"ContainerDied","Data":"f21ef54d9cfd076fae9a3da2004aa9e803aedc1288e566da226ece0ed47eae3a"} Dec 01 16:12:13 crc kubenswrapper[4810]: I1201 16:12:13.064080 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" event={"ID":"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc","Type":"ContainerStarted","Data":"18a9dfb7b3ef4c2d128cc7ecad4dcd48c04f858f7d19b1b08ddf92f85a348aed"} Dec 01 16:12:14 crc kubenswrapper[4810]: I1201 16:12:14.073539 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9b48b617-0ac4-4501-948c-3170743e02d9","Type":"ContainerStarted","Data":"4a31b57b8a8ed21175d6881348b00a949c61884f599ad70341def7c64575376b"} Dec 01 16:12:14 crc kubenswrapper[4810]: I1201 16:12:14.073859 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 01 16:12:14 crc kubenswrapper[4810]: I1201 16:12:14.075313 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" event={"ID":"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc","Type":"ContainerStarted","Data":"4bfbdb437e9e428fbacca57469e1b01885410191d71e644ae4e97e13a8b2ceb0"} Dec 01 16:12:14 crc kubenswrapper[4810]: I1201 16:12:14.075434 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:14 crc kubenswrapper[4810]: I1201 16:12:14.094592 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.094568923 podStartE2EDuration="3.094568923s" podCreationTimestamp="2025-12-01 16:12:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:12:14.087864681 +0000 UTC m=+5899.851374304" watchObservedRunningTime="2025-12-01 16:12:14.094568923 +0000 UTC m=+5899.858078526" Dec 01 16:12:14 crc kubenswrapper[4810]: I1201 16:12:14.115550 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" podStartSLOduration=3.115530053 podStartE2EDuration="3.115530053s" podCreationTimestamp="2025-12-01 16:12:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:12:14.108073419 +0000 UTC m=+5899.871583022" watchObservedRunningTime="2025-12-01 16:12:14.115530053 +0000 UTC m=+5899.879039656" Dec 01 16:12:14 crc kubenswrapper[4810]: I1201 16:12:14.285697 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 16:12:14 crc kubenswrapper[4810]: I1201 16:12:14.670052 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9j77z" Dec 01 16:12:14 crc kubenswrapper[4810]: I1201 16:12:14.670885 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9j77z" Dec 01 16:12:14 crc kubenswrapper[4810]: I1201 16:12:14.713831 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9j77z" Dec 01 16:12:15 crc kubenswrapper[4810]: I1201 16:12:15.131629 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9j77z" Dec 01 16:12:15 crc kubenswrapper[4810]: I1201 16:12:15.180413 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9j77z"] Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.093511 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="9b48b617-0ac4-4501-948c-3170743e02d9" containerName="cinder-api-log" containerID="cri-o://56b0ed08fe7a81b890c2925e53af51a1c7fb63b59433bc41c2664f7275e6d850" gracePeriod=30 Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.093546 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="9b48b617-0ac4-4501-948c-3170743e02d9" containerName="cinder-api" containerID="cri-o://4a31b57b8a8ed21175d6881348b00a949c61884f599ad70341def7c64575376b" gracePeriod=30 Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.635339 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.779149 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b48b617-0ac4-4501-948c-3170743e02d9-logs\") pod \"9b48b617-0ac4-4501-948c-3170743e02d9\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.779497 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b48b617-0ac4-4501-948c-3170743e02d9-etc-machine-id\") pod \"9b48b617-0ac4-4501-948c-3170743e02d9\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.779512 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b48b617-0ac4-4501-948c-3170743e02d9-logs" (OuterVolumeSpecName: "logs") pod "9b48b617-0ac4-4501-948c-3170743e02d9" (UID: "9b48b617-0ac4-4501-948c-3170743e02d9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.779603 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-scripts\") pod \"9b48b617-0ac4-4501-948c-3170743e02d9\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.779622 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b48b617-0ac4-4501-948c-3170743e02d9-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9b48b617-0ac4-4501-948c-3170743e02d9" (UID: "9b48b617-0ac4-4501-948c-3170743e02d9"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.779629 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-combined-ca-bundle\") pod \"9b48b617-0ac4-4501-948c-3170743e02d9\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.779707 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-config-data\") pod \"9b48b617-0ac4-4501-948c-3170743e02d9\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.779784 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hk4r9\" (UniqueName: \"kubernetes.io/projected/9b48b617-0ac4-4501-948c-3170743e02d9-kube-api-access-hk4r9\") pod \"9b48b617-0ac4-4501-948c-3170743e02d9\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.779810 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-config-data-custom\") pod \"9b48b617-0ac4-4501-948c-3170743e02d9\" (UID: \"9b48b617-0ac4-4501-948c-3170743e02d9\") " Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.780316 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b48b617-0ac4-4501-948c-3170743e02d9-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.780352 4810 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9b48b617-0ac4-4501-948c-3170743e02d9-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.784700 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9b48b617-0ac4-4501-948c-3170743e02d9" (UID: "9b48b617-0ac4-4501-948c-3170743e02d9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.784951 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-scripts" (OuterVolumeSpecName: "scripts") pod "9b48b617-0ac4-4501-948c-3170743e02d9" (UID: "9b48b617-0ac4-4501-948c-3170743e02d9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.785603 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b48b617-0ac4-4501-948c-3170743e02d9-kube-api-access-hk4r9" (OuterVolumeSpecName: "kube-api-access-hk4r9") pod "9b48b617-0ac4-4501-948c-3170743e02d9" (UID: "9b48b617-0ac4-4501-948c-3170743e02d9"). InnerVolumeSpecName "kube-api-access-hk4r9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.810639 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9b48b617-0ac4-4501-948c-3170743e02d9" (UID: "9b48b617-0ac4-4501-948c-3170743e02d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.837779 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-config-data" (OuterVolumeSpecName: "config-data") pod "9b48b617-0ac4-4501-948c-3170743e02d9" (UID: "9b48b617-0ac4-4501-948c-3170743e02d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.882434 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hk4r9\" (UniqueName: \"kubernetes.io/projected/9b48b617-0ac4-4501-948c-3170743e02d9-kube-api-access-hk4r9\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.882686 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.882772 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.882889 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:16 crc kubenswrapper[4810]: I1201 16:12:16.882966 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b48b617-0ac4-4501-948c-3170743e02d9-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.102094 4810 generic.go:334] "Generic (PLEG): container finished" podID="9b48b617-0ac4-4501-948c-3170743e02d9" containerID="4a31b57b8a8ed21175d6881348b00a949c61884f599ad70341def7c64575376b" exitCode=0 Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.102141 4810 generic.go:334] "Generic (PLEG): container finished" podID="9b48b617-0ac4-4501-948c-3170743e02d9" containerID="56b0ed08fe7a81b890c2925e53af51a1c7fb63b59433bc41c2664f7275e6d850" exitCode=143 Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.102152 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9b48b617-0ac4-4501-948c-3170743e02d9","Type":"ContainerDied","Data":"4a31b57b8a8ed21175d6881348b00a949c61884f599ad70341def7c64575376b"} Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.102199 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9b48b617-0ac4-4501-948c-3170743e02d9","Type":"ContainerDied","Data":"56b0ed08fe7a81b890c2925e53af51a1c7fb63b59433bc41c2664f7275e6d850"} Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.102210 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9b48b617-0ac4-4501-948c-3170743e02d9","Type":"ContainerDied","Data":"aa6f0d3db0101bf564747ae9606b80a50ec677ac22609197b50e202708f2e8a9"} Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.102226 4810 scope.go:117] "RemoveContainer" containerID="4a31b57b8a8ed21175d6881348b00a949c61884f599ad70341def7c64575376b" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.102388 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9j77z" podUID="6076eb68-acf2-4264-a1ef-549cb5137a50" containerName="registry-server" containerID="cri-o://712394811bb983e233135b2770367451ed6736d169c1f785eb059008d8c7b021" gracePeriod=2 Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.102886 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.151589 4810 scope.go:117] "RemoveContainer" containerID="56b0ed08fe7a81b890c2925e53af51a1c7fb63b59433bc41c2664f7275e6d850" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.159858 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.169950 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.178556 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 01 16:12:17 crc kubenswrapper[4810]: E1201 16:12:17.180461 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b48b617-0ac4-4501-948c-3170743e02d9" containerName="cinder-api" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.180498 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b48b617-0ac4-4501-948c-3170743e02d9" containerName="cinder-api" Dec 01 16:12:17 crc kubenswrapper[4810]: E1201 16:12:17.180511 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b48b617-0ac4-4501-948c-3170743e02d9" containerName="cinder-api-log" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.180519 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b48b617-0ac4-4501-948c-3170743e02d9" containerName="cinder-api-log" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.180680 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b48b617-0ac4-4501-948c-3170743e02d9" containerName="cinder-api" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.180695 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b48b617-0ac4-4501-948c-3170743e02d9" containerName="cinder-api-log" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.182242 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.186910 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.186978 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.187174 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-dvwkv" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.187286 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.187329 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.187417 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.190254 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.227814 4810 scope.go:117] "RemoveContainer" containerID="4a31b57b8a8ed21175d6881348b00a949c61884f599ad70341def7c64575376b" Dec 01 16:12:17 crc kubenswrapper[4810]: E1201 16:12:17.228235 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a31b57b8a8ed21175d6881348b00a949c61884f599ad70341def7c64575376b\": container with ID starting with 4a31b57b8a8ed21175d6881348b00a949c61884f599ad70341def7c64575376b not found: ID does not exist" containerID="4a31b57b8a8ed21175d6881348b00a949c61884f599ad70341def7c64575376b" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.228289 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a31b57b8a8ed21175d6881348b00a949c61884f599ad70341def7c64575376b"} err="failed to get container status \"4a31b57b8a8ed21175d6881348b00a949c61884f599ad70341def7c64575376b\": rpc error: code = NotFound desc = could not find container \"4a31b57b8a8ed21175d6881348b00a949c61884f599ad70341def7c64575376b\": container with ID starting with 4a31b57b8a8ed21175d6881348b00a949c61884f599ad70341def7c64575376b not found: ID does not exist" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.228329 4810 scope.go:117] "RemoveContainer" containerID="56b0ed08fe7a81b890c2925e53af51a1c7fb63b59433bc41c2664f7275e6d850" Dec 01 16:12:17 crc kubenswrapper[4810]: E1201 16:12:17.228680 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56b0ed08fe7a81b890c2925e53af51a1c7fb63b59433bc41c2664f7275e6d850\": container with ID starting with 56b0ed08fe7a81b890c2925e53af51a1c7fb63b59433bc41c2664f7275e6d850 not found: ID does not exist" containerID="56b0ed08fe7a81b890c2925e53af51a1c7fb63b59433bc41c2664f7275e6d850" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.228713 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56b0ed08fe7a81b890c2925e53af51a1c7fb63b59433bc41c2664f7275e6d850"} err="failed to get container status \"56b0ed08fe7a81b890c2925e53af51a1c7fb63b59433bc41c2664f7275e6d850\": rpc error: code = NotFound desc = could not find container \"56b0ed08fe7a81b890c2925e53af51a1c7fb63b59433bc41c2664f7275e6d850\": container with ID starting with 56b0ed08fe7a81b890c2925e53af51a1c7fb63b59433bc41c2664f7275e6d850 not found: ID does not exist" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.228734 4810 scope.go:117] "RemoveContainer" containerID="4a31b57b8a8ed21175d6881348b00a949c61884f599ad70341def7c64575376b" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.229051 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a31b57b8a8ed21175d6881348b00a949c61884f599ad70341def7c64575376b"} err="failed to get container status \"4a31b57b8a8ed21175d6881348b00a949c61884f599ad70341def7c64575376b\": rpc error: code = NotFound desc = could not find container \"4a31b57b8a8ed21175d6881348b00a949c61884f599ad70341def7c64575376b\": container with ID starting with 4a31b57b8a8ed21175d6881348b00a949c61884f599ad70341def7c64575376b not found: ID does not exist" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.229098 4810 scope.go:117] "RemoveContainer" containerID="56b0ed08fe7a81b890c2925e53af51a1c7fb63b59433bc41c2664f7275e6d850" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.229383 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56b0ed08fe7a81b890c2925e53af51a1c7fb63b59433bc41c2664f7275e6d850"} err="failed to get container status \"56b0ed08fe7a81b890c2925e53af51a1c7fb63b59433bc41c2664f7275e6d850\": rpc error: code = NotFound desc = could not find container \"56b0ed08fe7a81b890c2925e53af51a1c7fb63b59433bc41c2664f7275e6d850\": container with ID starting with 56b0ed08fe7a81b890c2925e53af51a1c7fb63b59433bc41c2664f7275e6d850 not found: ID does not exist" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.290406 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6l8p\" (UniqueName: \"kubernetes.io/projected/0e609f03-2ec3-4488-aba2-7170f6b301e2-kube-api-access-r6l8p\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.290485 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-config-data-custom\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.290659 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-config-data\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.290746 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.290796 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-scripts\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.290945 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e609f03-2ec3-4488-aba2-7170f6b301e2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.291030 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.291147 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e609f03-2ec3-4488-aba2-7170f6b301e2-logs\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.291241 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-public-tls-certs\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.393262 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-config-data\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.393313 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.393336 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-scripts\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.394027 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e609f03-2ec3-4488-aba2-7170f6b301e2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.394076 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.394078 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e609f03-2ec3-4488-aba2-7170f6b301e2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.394143 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e609f03-2ec3-4488-aba2-7170f6b301e2-logs\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.394212 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-public-tls-certs\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.394276 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6l8p\" (UniqueName: \"kubernetes.io/projected/0e609f03-2ec3-4488-aba2-7170f6b301e2-kube-api-access-r6l8p\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.394319 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-config-data-custom\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.395088 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e609f03-2ec3-4488-aba2-7170f6b301e2-logs\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.398526 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-public-tls-certs\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.398531 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.398837 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-config-data\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.401975 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-scripts\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.402279 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.402401 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-config-data-custom\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.409563 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6l8p\" (UniqueName: \"kubernetes.io/projected/0e609f03-2ec3-4488-aba2-7170f6b301e2-kube-api-access-r6l8p\") pod \"cinder-api-0\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.535857 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9j77z" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.600931 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.705071 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htvbp\" (UniqueName: \"kubernetes.io/projected/6076eb68-acf2-4264-a1ef-549cb5137a50-kube-api-access-htvbp\") pod \"6076eb68-acf2-4264-a1ef-549cb5137a50\" (UID: \"6076eb68-acf2-4264-a1ef-549cb5137a50\") " Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.705273 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6076eb68-acf2-4264-a1ef-549cb5137a50-utilities\") pod \"6076eb68-acf2-4264-a1ef-549cb5137a50\" (UID: \"6076eb68-acf2-4264-a1ef-549cb5137a50\") " Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.705385 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6076eb68-acf2-4264-a1ef-549cb5137a50-catalog-content\") pod \"6076eb68-acf2-4264-a1ef-549cb5137a50\" (UID: \"6076eb68-acf2-4264-a1ef-549cb5137a50\") " Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.710568 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6076eb68-acf2-4264-a1ef-549cb5137a50-utilities" (OuterVolumeSpecName: "utilities") pod "6076eb68-acf2-4264-a1ef-549cb5137a50" (UID: "6076eb68-acf2-4264-a1ef-549cb5137a50"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.740125 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6076eb68-acf2-4264-a1ef-549cb5137a50-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6076eb68-acf2-4264-a1ef-549cb5137a50" (UID: "6076eb68-acf2-4264-a1ef-549cb5137a50"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.741931 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6076eb68-acf2-4264-a1ef-549cb5137a50-kube-api-access-htvbp" (OuterVolumeSpecName: "kube-api-access-htvbp") pod "6076eb68-acf2-4264-a1ef-549cb5137a50" (UID: "6076eb68-acf2-4264-a1ef-549cb5137a50"). InnerVolumeSpecName "kube-api-access-htvbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.807998 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htvbp\" (UniqueName: \"kubernetes.io/projected/6076eb68-acf2-4264-a1ef-549cb5137a50-kube-api-access-htvbp\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.808411 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6076eb68-acf2-4264-a1ef-549cb5137a50-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:17 crc kubenswrapper[4810]: I1201 16:12:17.808453 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6076eb68-acf2-4264-a1ef-549cb5137a50-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:18 crc kubenswrapper[4810]: W1201 16:12:18.070756 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e609f03_2ec3_4488_aba2_7170f6b301e2.slice/crio-3b4d40bcf29ef4dc56f4f305033af409f59a61aa629468624b874daaa55b4955 WatchSource:0}: Error finding container 3b4d40bcf29ef4dc56f4f305033af409f59a61aa629468624b874daaa55b4955: Status 404 returned error can't find the container with id 3b4d40bcf29ef4dc56f4f305033af409f59a61aa629468624b874daaa55b4955 Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.071714 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.116153 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0e609f03-2ec3-4488-aba2-7170f6b301e2","Type":"ContainerStarted","Data":"3b4d40bcf29ef4dc56f4f305033af409f59a61aa629468624b874daaa55b4955"} Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.120644 4810 generic.go:334] "Generic (PLEG): container finished" podID="6076eb68-acf2-4264-a1ef-549cb5137a50" containerID="712394811bb983e233135b2770367451ed6736d169c1f785eb059008d8c7b021" exitCode=0 Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.120704 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9j77z" event={"ID":"6076eb68-acf2-4264-a1ef-549cb5137a50","Type":"ContainerDied","Data":"712394811bb983e233135b2770367451ed6736d169c1f785eb059008d8c7b021"} Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.120742 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9j77z" event={"ID":"6076eb68-acf2-4264-a1ef-549cb5137a50","Type":"ContainerDied","Data":"7952f4fbf6002079b3d6f0a4f98bc99bdad575d973e6fd4ff1a1f0a84eaa459f"} Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.120753 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9j77z" Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.120768 4810 scope.go:117] "RemoveContainer" containerID="712394811bb983e233135b2770367451ed6736d169c1f785eb059008d8c7b021" Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.154620 4810 scope.go:117] "RemoveContainer" containerID="039d67f6e3c92d7f1442a5c034338d3bef2346f41cd5000c341e836960163c50" Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.156479 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9j77z"] Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.164109 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9j77z"] Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.188768 4810 scope.go:117] "RemoveContainer" containerID="f173e9847dbbb2ac329dfa10e173b73ff6c361cbc699af7c2f5eb9df81325b19" Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.214667 4810 scope.go:117] "RemoveContainer" containerID="712394811bb983e233135b2770367451ed6736d169c1f785eb059008d8c7b021" Dec 01 16:12:18 crc kubenswrapper[4810]: E1201 16:12:18.217242 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"712394811bb983e233135b2770367451ed6736d169c1f785eb059008d8c7b021\": container with ID starting with 712394811bb983e233135b2770367451ed6736d169c1f785eb059008d8c7b021 not found: ID does not exist" containerID="712394811bb983e233135b2770367451ed6736d169c1f785eb059008d8c7b021" Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.217304 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"712394811bb983e233135b2770367451ed6736d169c1f785eb059008d8c7b021"} err="failed to get container status \"712394811bb983e233135b2770367451ed6736d169c1f785eb059008d8c7b021\": rpc error: code = NotFound desc = could not find container \"712394811bb983e233135b2770367451ed6736d169c1f785eb059008d8c7b021\": container with ID starting with 712394811bb983e233135b2770367451ed6736d169c1f785eb059008d8c7b021 not found: ID does not exist" Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.217346 4810 scope.go:117] "RemoveContainer" containerID="039d67f6e3c92d7f1442a5c034338d3bef2346f41cd5000c341e836960163c50" Dec 01 16:12:18 crc kubenswrapper[4810]: E1201 16:12:18.218334 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"039d67f6e3c92d7f1442a5c034338d3bef2346f41cd5000c341e836960163c50\": container with ID starting with 039d67f6e3c92d7f1442a5c034338d3bef2346f41cd5000c341e836960163c50 not found: ID does not exist" containerID="039d67f6e3c92d7f1442a5c034338d3bef2346f41cd5000c341e836960163c50" Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.218374 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"039d67f6e3c92d7f1442a5c034338d3bef2346f41cd5000c341e836960163c50"} err="failed to get container status \"039d67f6e3c92d7f1442a5c034338d3bef2346f41cd5000c341e836960163c50\": rpc error: code = NotFound desc = could not find container \"039d67f6e3c92d7f1442a5c034338d3bef2346f41cd5000c341e836960163c50\": container with ID starting with 039d67f6e3c92d7f1442a5c034338d3bef2346f41cd5000c341e836960163c50 not found: ID does not exist" Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.218405 4810 scope.go:117] "RemoveContainer" containerID="f173e9847dbbb2ac329dfa10e173b73ff6c361cbc699af7c2f5eb9df81325b19" Dec 01 16:12:18 crc kubenswrapper[4810]: E1201 16:12:18.219693 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f173e9847dbbb2ac329dfa10e173b73ff6c361cbc699af7c2f5eb9df81325b19\": container with ID starting with f173e9847dbbb2ac329dfa10e173b73ff6c361cbc699af7c2f5eb9df81325b19 not found: ID does not exist" containerID="f173e9847dbbb2ac329dfa10e173b73ff6c361cbc699af7c2f5eb9df81325b19" Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.219746 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f173e9847dbbb2ac329dfa10e173b73ff6c361cbc699af7c2f5eb9df81325b19"} err="failed to get container status \"f173e9847dbbb2ac329dfa10e173b73ff6c361cbc699af7c2f5eb9df81325b19\": rpc error: code = NotFound desc = could not find container \"f173e9847dbbb2ac329dfa10e173b73ff6c361cbc699af7c2f5eb9df81325b19\": container with ID starting with f173e9847dbbb2ac329dfa10e173b73ff6c361cbc699af7c2f5eb9df81325b19 not found: ID does not exist" Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.504408 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6076eb68-acf2-4264-a1ef-549cb5137a50" path="/var/lib/kubelet/pods/6076eb68-acf2-4264-a1ef-549cb5137a50/volumes" Dec 01 16:12:18 crc kubenswrapper[4810]: I1201 16:12:18.505292 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b48b617-0ac4-4501-948c-3170743e02d9" path="/var/lib/kubelet/pods/9b48b617-0ac4-4501-948c-3170743e02d9/volumes" Dec 01 16:12:19 crc kubenswrapper[4810]: I1201 16:12:19.148178 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0e609f03-2ec3-4488-aba2-7170f6b301e2","Type":"ContainerStarted","Data":"632d12484e59c135f099203905b141945f3ea14b7cef809b43c12179e5d02f0d"} Dec 01 16:12:20 crc kubenswrapper[4810]: I1201 16:12:20.161120 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0e609f03-2ec3-4488-aba2-7170f6b301e2","Type":"ContainerStarted","Data":"5f22b7a6893cd7cc768e24d0ee86bcff916cc26bf3e5561d937f6d272f7fd993"} Dec 01 16:12:20 crc kubenswrapper[4810]: I1201 16:12:20.161487 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 01 16:12:20 crc kubenswrapper[4810]: I1201 16:12:20.183314 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.183292502 podStartE2EDuration="3.183292502s" podCreationTimestamp="2025-12-01 16:12:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:12:20.178972755 +0000 UTC m=+5905.942482378" watchObservedRunningTime="2025-12-01 16:12:20.183292502 +0000 UTC m=+5905.946802105" Dec 01 16:12:21 crc kubenswrapper[4810]: I1201 16:12:21.662633 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:12:21 crc kubenswrapper[4810]: I1201 16:12:21.718561 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-667758f697-spt2d"] Dec 01 16:12:21 crc kubenswrapper[4810]: I1201 16:12:21.718801 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-667758f697-spt2d" podUID="51ae399d-0f9d-4644-8c93-d3e3a0446e47" containerName="dnsmasq-dns" containerID="cri-o://c2bb3483fe18662ea5bae4c150bf2664a3f69c896f0b0654b9a4e39c381eb3c3" gracePeriod=10 Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.184400 4810 generic.go:334] "Generic (PLEG): container finished" podID="51ae399d-0f9d-4644-8c93-d3e3a0446e47" containerID="c2bb3483fe18662ea5bae4c150bf2664a3f69c896f0b0654b9a4e39c381eb3c3" exitCode=0 Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.184494 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-667758f697-spt2d" event={"ID":"51ae399d-0f9d-4644-8c93-d3e3a0446e47","Type":"ContainerDied","Data":"c2bb3483fe18662ea5bae4c150bf2664a3f69c896f0b0654b9a4e39c381eb3c3"} Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.184743 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-667758f697-spt2d" event={"ID":"51ae399d-0f9d-4644-8c93-d3e3a0446e47","Type":"ContainerDied","Data":"5a160d68b049723c7d16369f6e31516beac91ba4c7482c043d9fe8ab5edc2740"} Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.184760 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a160d68b049723c7d16369f6e31516beac91ba4c7482c043d9fe8ab5edc2740" Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.234026 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.394250 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-dns-svc\") pod \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.394328 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6n4j\" (UniqueName: \"kubernetes.io/projected/51ae399d-0f9d-4644-8c93-d3e3a0446e47-kube-api-access-g6n4j\") pod \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.394406 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-ovsdbserver-sb\") pod \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.394523 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-ovsdbserver-nb\") pod \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.394550 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-config\") pod \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\" (UID: \"51ae399d-0f9d-4644-8c93-d3e3a0446e47\") " Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.399829 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51ae399d-0f9d-4644-8c93-d3e3a0446e47-kube-api-access-g6n4j" (OuterVolumeSpecName: "kube-api-access-g6n4j") pod "51ae399d-0f9d-4644-8c93-d3e3a0446e47" (UID: "51ae399d-0f9d-4644-8c93-d3e3a0446e47"). InnerVolumeSpecName "kube-api-access-g6n4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.440163 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "51ae399d-0f9d-4644-8c93-d3e3a0446e47" (UID: "51ae399d-0f9d-4644-8c93-d3e3a0446e47"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.444002 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "51ae399d-0f9d-4644-8c93-d3e3a0446e47" (UID: "51ae399d-0f9d-4644-8c93-d3e3a0446e47"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.451347 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-config" (OuterVolumeSpecName: "config") pod "51ae399d-0f9d-4644-8c93-d3e3a0446e47" (UID: "51ae399d-0f9d-4644-8c93-d3e3a0446e47"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.457034 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "51ae399d-0f9d-4644-8c93-d3e3a0446e47" (UID: "51ae399d-0f9d-4644-8c93-d3e3a0446e47"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.496203 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.496233 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.496241 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.496251 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6n4j\" (UniqueName: \"kubernetes.io/projected/51ae399d-0f9d-4644-8c93-d3e3a0446e47-kube-api-access-g6n4j\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:22 crc kubenswrapper[4810]: I1201 16:12:22.496263 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51ae399d-0f9d-4644-8c93-d3e3a0446e47-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:23 crc kubenswrapper[4810]: I1201 16:12:23.192188 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-667758f697-spt2d" Dec 01 16:12:23 crc kubenswrapper[4810]: I1201 16:12:23.219575 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-667758f697-spt2d"] Dec 01 16:12:23 crc kubenswrapper[4810]: I1201 16:12:23.230294 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-667758f697-spt2d"] Dec 01 16:12:24 crc kubenswrapper[4810]: I1201 16:12:24.501319 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51ae399d-0f9d-4644-8c93-d3e3a0446e47" path="/var/lib/kubelet/pods/51ae399d-0f9d-4644-8c93-d3e3a0446e47/volumes" Dec 01 16:12:27 crc kubenswrapper[4810]: I1201 16:12:27.024937 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-667758f697-spt2d" podUID="51ae399d-0f9d-4644-8c93-d3e3a0446e47" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.41:5353: i/o timeout" Dec 01 16:12:29 crc kubenswrapper[4810]: I1201 16:12:29.631552 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 01 16:12:46 crc kubenswrapper[4810]: I1201 16:12:46.783848 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 16:12:46 crc kubenswrapper[4810]: E1201 16:12:46.784931 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51ae399d-0f9d-4644-8c93-d3e3a0446e47" containerName="dnsmasq-dns" Dec 01 16:12:46 crc kubenswrapper[4810]: I1201 16:12:46.784947 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="51ae399d-0f9d-4644-8c93-d3e3a0446e47" containerName="dnsmasq-dns" Dec 01 16:12:46 crc kubenswrapper[4810]: E1201 16:12:46.784960 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6076eb68-acf2-4264-a1ef-549cb5137a50" containerName="extract-utilities" Dec 01 16:12:46 crc kubenswrapper[4810]: I1201 16:12:46.784966 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6076eb68-acf2-4264-a1ef-549cb5137a50" containerName="extract-utilities" Dec 01 16:12:46 crc kubenswrapper[4810]: E1201 16:12:46.784980 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6076eb68-acf2-4264-a1ef-549cb5137a50" containerName="extract-content" Dec 01 16:12:46 crc kubenswrapper[4810]: I1201 16:12:46.784988 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6076eb68-acf2-4264-a1ef-549cb5137a50" containerName="extract-content" Dec 01 16:12:46 crc kubenswrapper[4810]: E1201 16:12:46.784997 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6076eb68-acf2-4264-a1ef-549cb5137a50" containerName="registry-server" Dec 01 16:12:46 crc kubenswrapper[4810]: I1201 16:12:46.785003 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6076eb68-acf2-4264-a1ef-549cb5137a50" containerName="registry-server" Dec 01 16:12:46 crc kubenswrapper[4810]: E1201 16:12:46.785023 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51ae399d-0f9d-4644-8c93-d3e3a0446e47" containerName="init" Dec 01 16:12:46 crc kubenswrapper[4810]: I1201 16:12:46.785029 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="51ae399d-0f9d-4644-8c93-d3e3a0446e47" containerName="init" Dec 01 16:12:46 crc kubenswrapper[4810]: I1201 16:12:46.785209 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6076eb68-acf2-4264-a1ef-549cb5137a50" containerName="registry-server" Dec 01 16:12:46 crc kubenswrapper[4810]: I1201 16:12:46.785226 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="51ae399d-0f9d-4644-8c93-d3e3a0446e47" containerName="dnsmasq-dns" Dec 01 16:12:46 crc kubenswrapper[4810]: I1201 16:12:46.786286 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 16:12:46 crc kubenswrapper[4810]: I1201 16:12:46.788751 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 01 16:12:46 crc kubenswrapper[4810]: I1201 16:12:46.797865 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 16:12:46 crc kubenswrapper[4810]: I1201 16:12:46.955554 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-config-data\") pod \"cinder-scheduler-0\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " pod="openstack/cinder-scheduler-0" Dec 01 16:12:46 crc kubenswrapper[4810]: I1201 16:12:46.955591 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " pod="openstack/cinder-scheduler-0" Dec 01 16:12:46 crc kubenswrapper[4810]: I1201 16:12:46.955627 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " pod="openstack/cinder-scheduler-0" Dec 01 16:12:46 crc kubenswrapper[4810]: I1201 16:12:46.955655 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lmp6\" (UniqueName: \"kubernetes.io/projected/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-kube-api-access-9lmp6\") pod \"cinder-scheduler-0\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " pod="openstack/cinder-scheduler-0" Dec 01 16:12:46 crc kubenswrapper[4810]: I1201 16:12:46.955728 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " pod="openstack/cinder-scheduler-0" Dec 01 16:12:46 crc kubenswrapper[4810]: I1201 16:12:46.955752 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-scripts\") pod \"cinder-scheduler-0\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " pod="openstack/cinder-scheduler-0" Dec 01 16:12:47 crc kubenswrapper[4810]: I1201 16:12:47.056718 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " pod="openstack/cinder-scheduler-0" Dec 01 16:12:47 crc kubenswrapper[4810]: I1201 16:12:47.056766 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-scripts\") pod \"cinder-scheduler-0\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " pod="openstack/cinder-scheduler-0" Dec 01 16:12:47 crc kubenswrapper[4810]: I1201 16:12:47.056814 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-config-data\") pod \"cinder-scheduler-0\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " pod="openstack/cinder-scheduler-0" Dec 01 16:12:47 crc kubenswrapper[4810]: I1201 16:12:47.056833 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " pod="openstack/cinder-scheduler-0" Dec 01 16:12:47 crc kubenswrapper[4810]: I1201 16:12:47.056863 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " pod="openstack/cinder-scheduler-0" Dec 01 16:12:47 crc kubenswrapper[4810]: I1201 16:12:47.056892 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lmp6\" (UniqueName: \"kubernetes.io/projected/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-kube-api-access-9lmp6\") pod \"cinder-scheduler-0\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " pod="openstack/cinder-scheduler-0" Dec 01 16:12:47 crc kubenswrapper[4810]: I1201 16:12:47.057028 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " pod="openstack/cinder-scheduler-0" Dec 01 16:12:47 crc kubenswrapper[4810]: I1201 16:12:47.064525 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " pod="openstack/cinder-scheduler-0" Dec 01 16:12:47 crc kubenswrapper[4810]: I1201 16:12:47.066967 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " pod="openstack/cinder-scheduler-0" Dec 01 16:12:47 crc kubenswrapper[4810]: I1201 16:12:47.074186 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-scripts\") pod \"cinder-scheduler-0\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " pod="openstack/cinder-scheduler-0" Dec 01 16:12:47 crc kubenswrapper[4810]: I1201 16:12:47.074378 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-config-data\") pod \"cinder-scheduler-0\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " pod="openstack/cinder-scheduler-0" Dec 01 16:12:47 crc kubenswrapper[4810]: I1201 16:12:47.090161 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lmp6\" (UniqueName: \"kubernetes.io/projected/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-kube-api-access-9lmp6\") pod \"cinder-scheduler-0\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " pod="openstack/cinder-scheduler-0" Dec 01 16:12:47 crc kubenswrapper[4810]: I1201 16:12:47.117119 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 16:12:47 crc kubenswrapper[4810]: I1201 16:12:47.611393 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 16:12:48 crc kubenswrapper[4810]: I1201 16:12:48.314995 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 16:12:48 crc kubenswrapper[4810]: I1201 16:12:48.322764 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0e609f03-2ec3-4488-aba2-7170f6b301e2" containerName="cinder-api" containerID="cri-o://5f22b7a6893cd7cc768e24d0ee86bcff916cc26bf3e5561d937f6d272f7fd993" gracePeriod=30 Dec 01 16:12:48 crc kubenswrapper[4810]: I1201 16:12:48.322407 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0e609f03-2ec3-4488-aba2-7170f6b301e2" containerName="cinder-api-log" containerID="cri-o://632d12484e59c135f099203905b141945f3ea14b7cef809b43c12179e5d02f0d" gracePeriod=30 Dec 01 16:12:48 crc kubenswrapper[4810]: I1201 16:12:48.417926 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8","Type":"ContainerStarted","Data":"43bc44484a0e957d08019ff79f3b1a244ce87c3308146ed9f249a3f47ca1ec4a"} Dec 01 16:12:48 crc kubenswrapper[4810]: I1201 16:12:48.417970 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8","Type":"ContainerStarted","Data":"3419b848aa74f137f2b35b94305321f8ef591d11a6e95b0c9243be005bb340ef"} Dec 01 16:12:49 crc kubenswrapper[4810]: I1201 16:12:49.429051 4810 generic.go:334] "Generic (PLEG): container finished" podID="0e609f03-2ec3-4488-aba2-7170f6b301e2" containerID="632d12484e59c135f099203905b141945f3ea14b7cef809b43c12179e5d02f0d" exitCode=143 Dec 01 16:12:49 crc kubenswrapper[4810]: I1201 16:12:49.429222 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0e609f03-2ec3-4488-aba2-7170f6b301e2","Type":"ContainerDied","Data":"632d12484e59c135f099203905b141945f3ea14b7cef809b43c12179e5d02f0d"} Dec 01 16:12:49 crc kubenswrapper[4810]: I1201 16:12:49.430707 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8","Type":"ContainerStarted","Data":"a3f74a11de5cf64b9ceca3f2734ec4fee5f98cfaf2ba9fe6e6dd89499e569f65"} Dec 01 16:12:49 crc kubenswrapper[4810]: I1201 16:12:49.453695 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.45367941 podStartE2EDuration="3.45367941s" podCreationTimestamp="2025-12-01 16:12:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:12:49.452268252 +0000 UTC m=+5935.215777865" watchObservedRunningTime="2025-12-01 16:12:49.45367941 +0000 UTC m=+5935.217189013" Dec 01 16:12:51 crc kubenswrapper[4810]: I1201 16:12:51.891627 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.049104 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-config-data-custom\") pod \"0e609f03-2ec3-4488-aba2-7170f6b301e2\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.049180 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e609f03-2ec3-4488-aba2-7170f6b301e2-logs\") pod \"0e609f03-2ec3-4488-aba2-7170f6b301e2\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.049227 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-config-data\") pod \"0e609f03-2ec3-4488-aba2-7170f6b301e2\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.049251 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-combined-ca-bundle\") pod \"0e609f03-2ec3-4488-aba2-7170f6b301e2\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.049273 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-public-tls-certs\") pod \"0e609f03-2ec3-4488-aba2-7170f6b301e2\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.049295 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6l8p\" (UniqueName: \"kubernetes.io/projected/0e609f03-2ec3-4488-aba2-7170f6b301e2-kube-api-access-r6l8p\") pod \"0e609f03-2ec3-4488-aba2-7170f6b301e2\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.049314 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e609f03-2ec3-4488-aba2-7170f6b301e2-etc-machine-id\") pod \"0e609f03-2ec3-4488-aba2-7170f6b301e2\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.049460 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-internal-tls-certs\") pod \"0e609f03-2ec3-4488-aba2-7170f6b301e2\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.049530 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-scripts\") pod \"0e609f03-2ec3-4488-aba2-7170f6b301e2\" (UID: \"0e609f03-2ec3-4488-aba2-7170f6b301e2\") " Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.050949 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e609f03-2ec3-4488-aba2-7170f6b301e2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0e609f03-2ec3-4488-aba2-7170f6b301e2" (UID: "0e609f03-2ec3-4488-aba2-7170f6b301e2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.051363 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e609f03-2ec3-4488-aba2-7170f6b301e2-logs" (OuterVolumeSpecName: "logs") pod "0e609f03-2ec3-4488-aba2-7170f6b301e2" (UID: "0e609f03-2ec3-4488-aba2-7170f6b301e2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.068765 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0e609f03-2ec3-4488-aba2-7170f6b301e2" (UID: "0e609f03-2ec3-4488-aba2-7170f6b301e2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.073078 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-scripts" (OuterVolumeSpecName: "scripts") pod "0e609f03-2ec3-4488-aba2-7170f6b301e2" (UID: "0e609f03-2ec3-4488-aba2-7170f6b301e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.074446 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e609f03-2ec3-4488-aba2-7170f6b301e2-kube-api-access-r6l8p" (OuterVolumeSpecName: "kube-api-access-r6l8p") pod "0e609f03-2ec3-4488-aba2-7170f6b301e2" (UID: "0e609f03-2ec3-4488-aba2-7170f6b301e2"). InnerVolumeSpecName "kube-api-access-r6l8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.077419 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e609f03-2ec3-4488-aba2-7170f6b301e2" (UID: "0e609f03-2ec3-4488-aba2-7170f6b301e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.096986 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0e609f03-2ec3-4488-aba2-7170f6b301e2" (UID: "0e609f03-2ec3-4488-aba2-7170f6b301e2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.100657 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0e609f03-2ec3-4488-aba2-7170f6b301e2" (UID: "0e609f03-2ec3-4488-aba2-7170f6b301e2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.112494 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-config-data" (OuterVolumeSpecName: "config-data") pod "0e609f03-2ec3-4488-aba2-7170f6b301e2" (UID: "0e609f03-2ec3-4488-aba2-7170f6b301e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.118430 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.151977 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.152043 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.152058 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.152071 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6l8p\" (UniqueName: \"kubernetes.io/projected/0e609f03-2ec3-4488-aba2-7170f6b301e2-kube-api-access-r6l8p\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.152083 4810 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e609f03-2ec3-4488-aba2-7170f6b301e2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.152115 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.152124 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.152132 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e609f03-2ec3-4488-aba2-7170f6b301e2-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.152140 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e609f03-2ec3-4488-aba2-7170f6b301e2-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.462603 4810 generic.go:334] "Generic (PLEG): container finished" podID="0e609f03-2ec3-4488-aba2-7170f6b301e2" containerID="5f22b7a6893cd7cc768e24d0ee86bcff916cc26bf3e5561d937f6d272f7fd993" exitCode=0 Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.462663 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0e609f03-2ec3-4488-aba2-7170f6b301e2","Type":"ContainerDied","Data":"5f22b7a6893cd7cc768e24d0ee86bcff916cc26bf3e5561d937f6d272f7fd993"} Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.462692 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0e609f03-2ec3-4488-aba2-7170f6b301e2","Type":"ContainerDied","Data":"3b4d40bcf29ef4dc56f4f305033af409f59a61aa629468624b874daaa55b4955"} Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.462740 4810 scope.go:117] "RemoveContainer" containerID="5f22b7a6893cd7cc768e24d0ee86bcff916cc26bf3e5561d937f6d272f7fd993" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.462918 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.496717 4810 scope.go:117] "RemoveContainer" containerID="632d12484e59c135f099203905b141945f3ea14b7cef809b43c12179e5d02f0d" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.509205 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.511714 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.523355 4810 scope.go:117] "RemoveContainer" containerID="5f22b7a6893cd7cc768e24d0ee86bcff916cc26bf3e5561d937f6d272f7fd993" Dec 01 16:12:52 crc kubenswrapper[4810]: E1201 16:12:52.523866 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f22b7a6893cd7cc768e24d0ee86bcff916cc26bf3e5561d937f6d272f7fd993\": container with ID starting with 5f22b7a6893cd7cc768e24d0ee86bcff916cc26bf3e5561d937f6d272f7fd993 not found: ID does not exist" containerID="5f22b7a6893cd7cc768e24d0ee86bcff916cc26bf3e5561d937f6d272f7fd993" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.523921 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f22b7a6893cd7cc768e24d0ee86bcff916cc26bf3e5561d937f6d272f7fd993"} err="failed to get container status \"5f22b7a6893cd7cc768e24d0ee86bcff916cc26bf3e5561d937f6d272f7fd993\": rpc error: code = NotFound desc = could not find container \"5f22b7a6893cd7cc768e24d0ee86bcff916cc26bf3e5561d937f6d272f7fd993\": container with ID starting with 5f22b7a6893cd7cc768e24d0ee86bcff916cc26bf3e5561d937f6d272f7fd993 not found: ID does not exist" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.523951 4810 scope.go:117] "RemoveContainer" containerID="632d12484e59c135f099203905b141945f3ea14b7cef809b43c12179e5d02f0d" Dec 01 16:12:52 crc kubenswrapper[4810]: E1201 16:12:52.525129 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"632d12484e59c135f099203905b141945f3ea14b7cef809b43c12179e5d02f0d\": container with ID starting with 632d12484e59c135f099203905b141945f3ea14b7cef809b43c12179e5d02f0d not found: ID does not exist" containerID="632d12484e59c135f099203905b141945f3ea14b7cef809b43c12179e5d02f0d" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.525160 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"632d12484e59c135f099203905b141945f3ea14b7cef809b43c12179e5d02f0d"} err="failed to get container status \"632d12484e59c135f099203905b141945f3ea14b7cef809b43c12179e5d02f0d\": rpc error: code = NotFound desc = could not find container \"632d12484e59c135f099203905b141945f3ea14b7cef809b43c12179e5d02f0d\": container with ID starting with 632d12484e59c135f099203905b141945f3ea14b7cef809b43c12179e5d02f0d not found: ID does not exist" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.527391 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 01 16:12:52 crc kubenswrapper[4810]: E1201 16:12:52.527851 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e609f03-2ec3-4488-aba2-7170f6b301e2" containerName="cinder-api" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.527873 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e609f03-2ec3-4488-aba2-7170f6b301e2" containerName="cinder-api" Dec 01 16:12:52 crc kubenswrapper[4810]: E1201 16:12:52.527895 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e609f03-2ec3-4488-aba2-7170f6b301e2" containerName="cinder-api-log" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.527903 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e609f03-2ec3-4488-aba2-7170f6b301e2" containerName="cinder-api-log" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.528071 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e609f03-2ec3-4488-aba2-7170f6b301e2" containerName="cinder-api-log" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.528101 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e609f03-2ec3-4488-aba2-7170f6b301e2" containerName="cinder-api" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.529108 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.534150 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.534331 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.534610 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.536271 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.669123 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4412031a-3547-49c1-907b-450db04bc075-config-data-custom\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.669274 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4412031a-3547-49c1-907b-450db04bc075-config-data\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.669303 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4412031a-3547-49c1-907b-450db04bc075-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.669340 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4412031a-3547-49c1-907b-450db04bc075-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.669379 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4412031a-3547-49c1-907b-450db04bc075-scripts\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.669413 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt7b6\" (UniqueName: \"kubernetes.io/projected/4412031a-3547-49c1-907b-450db04bc075-kube-api-access-tt7b6\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.669711 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4412031a-3547-49c1-907b-450db04bc075-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.669814 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4412031a-3547-49c1-907b-450db04bc075-logs\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.669875 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4412031a-3547-49c1-907b-450db04bc075-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.771390 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4412031a-3547-49c1-907b-450db04bc075-config-data-custom\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.771458 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4412031a-3547-49c1-907b-450db04bc075-config-data\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.771498 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4412031a-3547-49c1-907b-450db04bc075-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.771516 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4412031a-3547-49c1-907b-450db04bc075-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.771545 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4412031a-3547-49c1-907b-450db04bc075-scripts\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.771560 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt7b6\" (UniqueName: \"kubernetes.io/projected/4412031a-3547-49c1-907b-450db04bc075-kube-api-access-tt7b6\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.771599 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4412031a-3547-49c1-907b-450db04bc075-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.771647 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4412031a-3547-49c1-907b-450db04bc075-logs\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.771670 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4412031a-3547-49c1-907b-450db04bc075-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.771733 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4412031a-3547-49c1-907b-450db04bc075-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.772593 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4412031a-3547-49c1-907b-450db04bc075-logs\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.776099 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4412031a-3547-49c1-907b-450db04bc075-scripts\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.778708 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4412031a-3547-49c1-907b-450db04bc075-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.779160 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4412031a-3547-49c1-907b-450db04bc075-config-data\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.779343 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4412031a-3547-49c1-907b-450db04bc075-config-data-custom\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.779406 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4412031a-3547-49c1-907b-450db04bc075-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.780902 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4412031a-3547-49c1-907b-450db04bc075-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.789679 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt7b6\" (UniqueName: \"kubernetes.io/projected/4412031a-3547-49c1-907b-450db04bc075-kube-api-access-tt7b6\") pod \"cinder-api-0\" (UID: \"4412031a-3547-49c1-907b-450db04bc075\") " pod="openstack/cinder-api-0" Dec 01 16:12:52 crc kubenswrapper[4810]: I1201 16:12:52.854956 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 16:12:53 crc kubenswrapper[4810]: I1201 16:12:53.323043 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 16:12:53 crc kubenswrapper[4810]: W1201 16:12:53.324873 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4412031a_3547_49c1_907b_450db04bc075.slice/crio-3053788c125738ef2b61d168a12f6fcb3c19efa2413024b3c78182dbbae39848 WatchSource:0}: Error finding container 3053788c125738ef2b61d168a12f6fcb3c19efa2413024b3c78182dbbae39848: Status 404 returned error can't find the container with id 3053788c125738ef2b61d168a12f6fcb3c19efa2413024b3c78182dbbae39848 Dec 01 16:12:53 crc kubenswrapper[4810]: I1201 16:12:53.472544 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4412031a-3547-49c1-907b-450db04bc075","Type":"ContainerStarted","Data":"3053788c125738ef2b61d168a12f6fcb3c19efa2413024b3c78182dbbae39848"} Dec 01 16:12:54 crc kubenswrapper[4810]: I1201 16:12:54.484424 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4412031a-3547-49c1-907b-450db04bc075","Type":"ContainerStarted","Data":"50a5eae1dfd743ad1ab855e0caa920e2ef366a7f644704f0990d5a4baad9270c"} Dec 01 16:12:54 crc kubenswrapper[4810]: I1201 16:12:54.485027 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 01 16:12:54 crc kubenswrapper[4810]: I1201 16:12:54.485042 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4412031a-3547-49c1-907b-450db04bc075","Type":"ContainerStarted","Data":"3a3c0554478e1797799a28313a56dbc7caf1d5dc638afa5dec2a2000ca55465e"} Dec 01 16:12:54 crc kubenswrapper[4810]: I1201 16:12:54.508231 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e609f03-2ec3-4488-aba2-7170f6b301e2" path="/var/lib/kubelet/pods/0e609f03-2ec3-4488-aba2-7170f6b301e2/volumes" Dec 01 16:12:54 crc kubenswrapper[4810]: I1201 16:12:54.514685 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.51466601 podStartE2EDuration="2.51466601s" podCreationTimestamp="2025-12-01 16:12:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:12:54.507380723 +0000 UTC m=+5940.270890336" watchObservedRunningTime="2025-12-01 16:12:54.51466601 +0000 UTC m=+5940.278175613" Dec 01 16:12:57 crc kubenswrapper[4810]: I1201 16:12:57.321088 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 01 16:12:57 crc kubenswrapper[4810]: I1201 16:12:57.373279 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 16:12:57 crc kubenswrapper[4810]: I1201 16:12:57.512125 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1a8179a8-6b97-4ad0-93a9-2c73fb093cd8" containerName="cinder-scheduler" containerID="cri-o://43bc44484a0e957d08019ff79f3b1a244ce87c3308146ed9f249a3f47ca1ec4a" gracePeriod=30 Dec 01 16:12:57 crc kubenswrapper[4810]: I1201 16:12:57.512172 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1a8179a8-6b97-4ad0-93a9-2c73fb093cd8" containerName="probe" containerID="cri-o://a3f74a11de5cf64b9ceca3f2734ec4fee5f98cfaf2ba9fe6e6dd89499e569f65" gracePeriod=30 Dec 01 16:12:58 crc kubenswrapper[4810]: I1201 16:12:58.524130 4810 generic.go:334] "Generic (PLEG): container finished" podID="1a8179a8-6b97-4ad0-93a9-2c73fb093cd8" containerID="a3f74a11de5cf64b9ceca3f2734ec4fee5f98cfaf2ba9fe6e6dd89499e569f65" exitCode=0 Dec 01 16:12:58 crc kubenswrapper[4810]: I1201 16:12:58.525208 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8","Type":"ContainerDied","Data":"a3f74a11de5cf64b9ceca3f2734ec4fee5f98cfaf2ba9fe6e6dd89499e569f65"} Dec 01 16:12:59 crc kubenswrapper[4810]: I1201 16:12:59.548279 4810 generic.go:334] "Generic (PLEG): container finished" podID="1a8179a8-6b97-4ad0-93a9-2c73fb093cd8" containerID="43bc44484a0e957d08019ff79f3b1a244ce87c3308146ed9f249a3f47ca1ec4a" exitCode=0 Dec 01 16:12:59 crc kubenswrapper[4810]: I1201 16:12:59.548365 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8","Type":"ContainerDied","Data":"43bc44484a0e957d08019ff79f3b1a244ce87c3308146ed9f249a3f47ca1ec4a"} Dec 01 16:12:59 crc kubenswrapper[4810]: I1201 16:12:59.819032 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 16:12:59 crc kubenswrapper[4810]: I1201 16:12:59.997672 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-scripts\") pod \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " Dec 01 16:12:59 crc kubenswrapper[4810]: I1201 16:12:59.998206 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-config-data\") pod \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " Dec 01 16:12:59 crc kubenswrapper[4810]: I1201 16:12:59.998366 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-etc-machine-id\") pod \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " Dec 01 16:12:59 crc kubenswrapper[4810]: I1201 16:12:59.998498 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1a8179a8-6b97-4ad0-93a9-2c73fb093cd8" (UID: "1a8179a8-6b97-4ad0-93a9-2c73fb093cd8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 16:12:59 crc kubenswrapper[4810]: I1201 16:12:59.998665 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-combined-ca-bundle\") pod \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " Dec 01 16:12:59 crc kubenswrapper[4810]: I1201 16:12:59.998821 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-config-data-custom\") pod \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " Dec 01 16:12:59 crc kubenswrapper[4810]: I1201 16:12:59.998949 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lmp6\" (UniqueName: \"kubernetes.io/projected/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-kube-api-access-9lmp6\") pod \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\" (UID: \"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8\") " Dec 01 16:12:59 crc kubenswrapper[4810]: I1201 16:12:59.999650 4810 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.003636 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-scripts" (OuterVolumeSpecName: "scripts") pod "1a8179a8-6b97-4ad0-93a9-2c73fb093cd8" (UID: "1a8179a8-6b97-4ad0-93a9-2c73fb093cd8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.007389 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-kube-api-access-9lmp6" (OuterVolumeSpecName: "kube-api-access-9lmp6") pod "1a8179a8-6b97-4ad0-93a9-2c73fb093cd8" (UID: "1a8179a8-6b97-4ad0-93a9-2c73fb093cd8"). InnerVolumeSpecName "kube-api-access-9lmp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.009644 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1a8179a8-6b97-4ad0-93a9-2c73fb093cd8" (UID: "1a8179a8-6b97-4ad0-93a9-2c73fb093cd8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.049039 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a8179a8-6b97-4ad0-93a9-2c73fb093cd8" (UID: "1a8179a8-6b97-4ad0-93a9-2c73fb093cd8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.094834 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-config-data" (OuterVolumeSpecName: "config-data") pod "1a8179a8-6b97-4ad0-93a9-2c73fb093cd8" (UID: "1a8179a8-6b97-4ad0-93a9-2c73fb093cd8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.101711 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.101771 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.101790 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lmp6\" (UniqueName: \"kubernetes.io/projected/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-kube-api-access-9lmp6\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.101843 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.101862 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.559830 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1a8179a8-6b97-4ad0-93a9-2c73fb093cd8","Type":"ContainerDied","Data":"3419b848aa74f137f2b35b94305321f8ef591d11a6e95b0c9243be005bb340ef"} Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.559914 4810 scope.go:117] "RemoveContainer" containerID="a3f74a11de5cf64b9ceca3f2734ec4fee5f98cfaf2ba9fe6e6dd89499e569f65" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.559977 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.583283 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.588751 4810 scope.go:117] "RemoveContainer" containerID="43bc44484a0e957d08019ff79f3b1a244ce87c3308146ed9f249a3f47ca1ec4a" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.592509 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.614215 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 16:13:00 crc kubenswrapper[4810]: E1201 16:13:00.614621 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a8179a8-6b97-4ad0-93a9-2c73fb093cd8" containerName="cinder-scheduler" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.614635 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a8179a8-6b97-4ad0-93a9-2c73fb093cd8" containerName="cinder-scheduler" Dec 01 16:13:00 crc kubenswrapper[4810]: E1201 16:13:00.614656 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a8179a8-6b97-4ad0-93a9-2c73fb093cd8" containerName="probe" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.614661 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a8179a8-6b97-4ad0-93a9-2c73fb093cd8" containerName="probe" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.614827 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a8179a8-6b97-4ad0-93a9-2c73fb093cd8" containerName="probe" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.614852 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a8179a8-6b97-4ad0-93a9-2c73fb093cd8" containerName="cinder-scheduler" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.615741 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.635894 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.636135 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.716256 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/82892053-06d4-414e-9c79-4fd70eea44b3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"82892053-06d4-414e-9c79-4fd70eea44b3\") " pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.716300 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg9d5\" (UniqueName: \"kubernetes.io/projected/82892053-06d4-414e-9c79-4fd70eea44b3-kube-api-access-fg9d5\") pod \"cinder-scheduler-0\" (UID: \"82892053-06d4-414e-9c79-4fd70eea44b3\") " pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.716341 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82892053-06d4-414e-9c79-4fd70eea44b3-config-data\") pod \"cinder-scheduler-0\" (UID: \"82892053-06d4-414e-9c79-4fd70eea44b3\") " pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.716365 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82892053-06d4-414e-9c79-4fd70eea44b3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"82892053-06d4-414e-9c79-4fd70eea44b3\") " pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.716554 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82892053-06d4-414e-9c79-4fd70eea44b3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"82892053-06d4-414e-9c79-4fd70eea44b3\") " pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.716587 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82892053-06d4-414e-9c79-4fd70eea44b3-scripts\") pod \"cinder-scheduler-0\" (UID: \"82892053-06d4-414e-9c79-4fd70eea44b3\") " pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.818547 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82892053-06d4-414e-9c79-4fd70eea44b3-scripts\") pod \"cinder-scheduler-0\" (UID: \"82892053-06d4-414e-9c79-4fd70eea44b3\") " pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.818669 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/82892053-06d4-414e-9c79-4fd70eea44b3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"82892053-06d4-414e-9c79-4fd70eea44b3\") " pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.818701 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg9d5\" (UniqueName: \"kubernetes.io/projected/82892053-06d4-414e-9c79-4fd70eea44b3-kube-api-access-fg9d5\") pod \"cinder-scheduler-0\" (UID: \"82892053-06d4-414e-9c79-4fd70eea44b3\") " pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.818746 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82892053-06d4-414e-9c79-4fd70eea44b3-config-data\") pod \"cinder-scheduler-0\" (UID: \"82892053-06d4-414e-9c79-4fd70eea44b3\") " pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.818774 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82892053-06d4-414e-9c79-4fd70eea44b3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"82892053-06d4-414e-9c79-4fd70eea44b3\") " pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.818891 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82892053-06d4-414e-9c79-4fd70eea44b3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"82892053-06d4-414e-9c79-4fd70eea44b3\") " pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.818969 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82892053-06d4-414e-9c79-4fd70eea44b3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"82892053-06d4-414e-9c79-4fd70eea44b3\") " pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.822249 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/82892053-06d4-414e-9c79-4fd70eea44b3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"82892053-06d4-414e-9c79-4fd70eea44b3\") " pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.823623 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82892053-06d4-414e-9c79-4fd70eea44b3-config-data\") pod \"cinder-scheduler-0\" (UID: \"82892053-06d4-414e-9c79-4fd70eea44b3\") " pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.831866 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82892053-06d4-414e-9c79-4fd70eea44b3-scripts\") pod \"cinder-scheduler-0\" (UID: \"82892053-06d4-414e-9c79-4fd70eea44b3\") " pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.832182 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82892053-06d4-414e-9c79-4fd70eea44b3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"82892053-06d4-414e-9c79-4fd70eea44b3\") " pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.835558 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg9d5\" (UniqueName: \"kubernetes.io/projected/82892053-06d4-414e-9c79-4fd70eea44b3-kube-api-access-fg9d5\") pod \"cinder-scheduler-0\" (UID: \"82892053-06d4-414e-9c79-4fd70eea44b3\") " pod="openstack/cinder-scheduler-0" Dec 01 16:13:00 crc kubenswrapper[4810]: I1201 16:13:00.958828 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 16:13:01 crc kubenswrapper[4810]: I1201 16:13:01.366819 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 16:13:01 crc kubenswrapper[4810]: I1201 16:13:01.570722 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"82892053-06d4-414e-9c79-4fd70eea44b3","Type":"ContainerStarted","Data":"4c35392f142e0726ef7b544156b0621a0743562e42c3818f23afb5a94acd27ea"} Dec 01 16:13:02 crc kubenswrapper[4810]: I1201 16:13:02.508889 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a8179a8-6b97-4ad0-93a9-2c73fb093cd8" path="/var/lib/kubelet/pods/1a8179a8-6b97-4ad0-93a9-2c73fb093cd8/volumes" Dec 01 16:13:02 crc kubenswrapper[4810]: I1201 16:13:02.582630 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"82892053-06d4-414e-9c79-4fd70eea44b3","Type":"ContainerStarted","Data":"65fbcd3d34c46efebdc56dead86b434f220b6894f2fe802d9a9be5e9e787ac1b"} Dec 01 16:13:02 crc kubenswrapper[4810]: I1201 16:13:02.582678 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"82892053-06d4-414e-9c79-4fd70eea44b3","Type":"ContainerStarted","Data":"8afbd47299a727e9bfa733d5cfc262e59fec52201509f8d310638d379e22ff92"} Dec 01 16:13:02 crc kubenswrapper[4810]: I1201 16:13:02.608065 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.608042857 podStartE2EDuration="2.608042857s" podCreationTimestamp="2025-12-01 16:13:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:13:02.60118256 +0000 UTC m=+5948.364692163" watchObservedRunningTime="2025-12-01 16:13:02.608042857 +0000 UTC m=+5948.371552470" Dec 01 16:13:02 crc kubenswrapper[4810]: I1201 16:13:02.972197 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:13:02 crc kubenswrapper[4810]: I1201 16:13:02.972252 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:13:04 crc kubenswrapper[4810]: I1201 16:13:04.737103 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 01 16:13:05 crc kubenswrapper[4810]: I1201 16:13:05.959739 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 01 16:13:11 crc kubenswrapper[4810]: I1201 16:13:11.155975 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 01 16:13:13 crc kubenswrapper[4810]: I1201 16:13:13.971235 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-9lwc4"] Dec 01 16:13:13 crc kubenswrapper[4810]: I1201 16:13:13.973147 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9lwc4" Dec 01 16:13:13 crc kubenswrapper[4810]: I1201 16:13:13.982410 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-9lwc4"] Dec 01 16:13:14 crc kubenswrapper[4810]: I1201 16:13:14.076778 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-481d-account-create-update-2rdl9"] Dec 01 16:13:14 crc kubenswrapper[4810]: I1201 16:13:14.077967 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-481d-account-create-update-2rdl9" Dec 01 16:13:14 crc kubenswrapper[4810]: I1201 16:13:14.080251 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/607341d7-10b4-4d40-a2a2-7647ceff0c14-operator-scripts\") pod \"glance-db-create-9lwc4\" (UID: \"607341d7-10b4-4d40-a2a2-7647ceff0c14\") " pod="openstack/glance-db-create-9lwc4" Dec 01 16:13:14 crc kubenswrapper[4810]: I1201 16:13:14.080345 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w9mf\" (UniqueName: \"kubernetes.io/projected/607341d7-10b4-4d40-a2a2-7647ceff0c14-kube-api-access-9w9mf\") pod \"glance-db-create-9lwc4\" (UID: \"607341d7-10b4-4d40-a2a2-7647ceff0c14\") " pod="openstack/glance-db-create-9lwc4" Dec 01 16:13:14 crc kubenswrapper[4810]: I1201 16:13:14.082858 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 01 16:13:14 crc kubenswrapper[4810]: I1201 16:13:14.128021 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-481d-account-create-update-2rdl9"] Dec 01 16:13:14 crc kubenswrapper[4810]: I1201 16:13:14.181601 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25c3e988-ae27-4369-aa98-4bd7a3c5848a-operator-scripts\") pod \"glance-481d-account-create-update-2rdl9\" (UID: \"25c3e988-ae27-4369-aa98-4bd7a3c5848a\") " pod="openstack/glance-481d-account-create-update-2rdl9" Dec 01 16:13:14 crc kubenswrapper[4810]: I1201 16:13:14.181665 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w9mf\" (UniqueName: \"kubernetes.io/projected/607341d7-10b4-4d40-a2a2-7647ceff0c14-kube-api-access-9w9mf\") pod \"glance-db-create-9lwc4\" (UID: \"607341d7-10b4-4d40-a2a2-7647ceff0c14\") " pod="openstack/glance-db-create-9lwc4" Dec 01 16:13:14 crc kubenswrapper[4810]: I1201 16:13:14.182056 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/607341d7-10b4-4d40-a2a2-7647ceff0c14-operator-scripts\") pod \"glance-db-create-9lwc4\" (UID: \"607341d7-10b4-4d40-a2a2-7647ceff0c14\") " pod="openstack/glance-db-create-9lwc4" Dec 01 16:13:14 crc kubenswrapper[4810]: I1201 16:13:14.182109 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2gng\" (UniqueName: \"kubernetes.io/projected/25c3e988-ae27-4369-aa98-4bd7a3c5848a-kube-api-access-x2gng\") pod \"glance-481d-account-create-update-2rdl9\" (UID: \"25c3e988-ae27-4369-aa98-4bd7a3c5848a\") " pod="openstack/glance-481d-account-create-update-2rdl9" Dec 01 16:13:14 crc kubenswrapper[4810]: I1201 16:13:14.182838 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/607341d7-10b4-4d40-a2a2-7647ceff0c14-operator-scripts\") pod \"glance-db-create-9lwc4\" (UID: \"607341d7-10b4-4d40-a2a2-7647ceff0c14\") " pod="openstack/glance-db-create-9lwc4" Dec 01 16:13:14 crc kubenswrapper[4810]: I1201 16:13:14.208723 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w9mf\" (UniqueName: \"kubernetes.io/projected/607341d7-10b4-4d40-a2a2-7647ceff0c14-kube-api-access-9w9mf\") pod \"glance-db-create-9lwc4\" (UID: \"607341d7-10b4-4d40-a2a2-7647ceff0c14\") " pod="openstack/glance-db-create-9lwc4" Dec 01 16:13:14 crc kubenswrapper[4810]: I1201 16:13:14.283553 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2gng\" (UniqueName: \"kubernetes.io/projected/25c3e988-ae27-4369-aa98-4bd7a3c5848a-kube-api-access-x2gng\") pod \"glance-481d-account-create-update-2rdl9\" (UID: \"25c3e988-ae27-4369-aa98-4bd7a3c5848a\") " pod="openstack/glance-481d-account-create-update-2rdl9" Dec 01 16:13:14 crc kubenswrapper[4810]: I1201 16:13:14.283643 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25c3e988-ae27-4369-aa98-4bd7a3c5848a-operator-scripts\") pod \"glance-481d-account-create-update-2rdl9\" (UID: \"25c3e988-ae27-4369-aa98-4bd7a3c5848a\") " pod="openstack/glance-481d-account-create-update-2rdl9" Dec 01 16:13:14 crc kubenswrapper[4810]: I1201 16:13:14.284753 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25c3e988-ae27-4369-aa98-4bd7a3c5848a-operator-scripts\") pod \"glance-481d-account-create-update-2rdl9\" (UID: \"25c3e988-ae27-4369-aa98-4bd7a3c5848a\") " pod="openstack/glance-481d-account-create-update-2rdl9" Dec 01 16:13:14 crc kubenswrapper[4810]: I1201 16:13:14.304215 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2gng\" (UniqueName: \"kubernetes.io/projected/25c3e988-ae27-4369-aa98-4bd7a3c5848a-kube-api-access-x2gng\") pod \"glance-481d-account-create-update-2rdl9\" (UID: \"25c3e988-ae27-4369-aa98-4bd7a3c5848a\") " pod="openstack/glance-481d-account-create-update-2rdl9" Dec 01 16:13:14 crc kubenswrapper[4810]: I1201 16:13:14.305086 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9lwc4" Dec 01 16:13:14 crc kubenswrapper[4810]: I1201 16:13:14.394837 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-481d-account-create-update-2rdl9" Dec 01 16:13:15 crc kubenswrapper[4810]: W1201 16:13:15.046603 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod607341d7_10b4_4d40_a2a2_7647ceff0c14.slice/crio-bd58bff5762edaa305510da137be49450ec3edcb712d5ef84473a1df00b19f59 WatchSource:0}: Error finding container bd58bff5762edaa305510da137be49450ec3edcb712d5ef84473a1df00b19f59: Status 404 returned error can't find the container with id bd58bff5762edaa305510da137be49450ec3edcb712d5ef84473a1df00b19f59 Dec 01 16:13:15 crc kubenswrapper[4810]: I1201 16:13:15.048215 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-9lwc4"] Dec 01 16:13:15 crc kubenswrapper[4810]: I1201 16:13:15.064316 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-481d-account-create-update-2rdl9"] Dec 01 16:13:15 crc kubenswrapper[4810]: I1201 16:13:15.744536 4810 generic.go:334] "Generic (PLEG): container finished" podID="607341d7-10b4-4d40-a2a2-7647ceff0c14" containerID="21be20ccde4a34f9f5035ce0cc253e3abf1c6cea80658e626418f9db0ee7a24c" exitCode=0 Dec 01 16:13:15 crc kubenswrapper[4810]: I1201 16:13:15.744585 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9lwc4" event={"ID":"607341d7-10b4-4d40-a2a2-7647ceff0c14","Type":"ContainerDied","Data":"21be20ccde4a34f9f5035ce0cc253e3abf1c6cea80658e626418f9db0ee7a24c"} Dec 01 16:13:15 crc kubenswrapper[4810]: I1201 16:13:15.744630 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9lwc4" event={"ID":"607341d7-10b4-4d40-a2a2-7647ceff0c14","Type":"ContainerStarted","Data":"bd58bff5762edaa305510da137be49450ec3edcb712d5ef84473a1df00b19f59"} Dec 01 16:13:15 crc kubenswrapper[4810]: I1201 16:13:15.746350 4810 generic.go:334] "Generic (PLEG): container finished" podID="25c3e988-ae27-4369-aa98-4bd7a3c5848a" containerID="bb9403ef5d0fd0c4f80ae6cc912979f62b9a82165899bfe05216834631ccc2a8" exitCode=0 Dec 01 16:13:15 crc kubenswrapper[4810]: I1201 16:13:15.746398 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-481d-account-create-update-2rdl9" event={"ID":"25c3e988-ae27-4369-aa98-4bd7a3c5848a","Type":"ContainerDied","Data":"bb9403ef5d0fd0c4f80ae6cc912979f62b9a82165899bfe05216834631ccc2a8"} Dec 01 16:13:15 crc kubenswrapper[4810]: I1201 16:13:15.746425 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-481d-account-create-update-2rdl9" event={"ID":"25c3e988-ae27-4369-aa98-4bd7a3c5848a","Type":"ContainerStarted","Data":"6302902a1abb26b7c265c363da1edb911b21baa6aa2370688f75716b518ffac4"} Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.089170 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-481d-account-create-update-2rdl9" Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.095102 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9lwc4" Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.196067 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9w9mf\" (UniqueName: \"kubernetes.io/projected/607341d7-10b4-4d40-a2a2-7647ceff0c14-kube-api-access-9w9mf\") pod \"607341d7-10b4-4d40-a2a2-7647ceff0c14\" (UID: \"607341d7-10b4-4d40-a2a2-7647ceff0c14\") " Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.196142 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25c3e988-ae27-4369-aa98-4bd7a3c5848a-operator-scripts\") pod \"25c3e988-ae27-4369-aa98-4bd7a3c5848a\" (UID: \"25c3e988-ae27-4369-aa98-4bd7a3c5848a\") " Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.196187 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/607341d7-10b4-4d40-a2a2-7647ceff0c14-operator-scripts\") pod \"607341d7-10b4-4d40-a2a2-7647ceff0c14\" (UID: \"607341d7-10b4-4d40-a2a2-7647ceff0c14\") " Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.196310 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2gng\" (UniqueName: \"kubernetes.io/projected/25c3e988-ae27-4369-aa98-4bd7a3c5848a-kube-api-access-x2gng\") pod \"25c3e988-ae27-4369-aa98-4bd7a3c5848a\" (UID: \"25c3e988-ae27-4369-aa98-4bd7a3c5848a\") " Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.197017 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25c3e988-ae27-4369-aa98-4bd7a3c5848a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "25c3e988-ae27-4369-aa98-4bd7a3c5848a" (UID: "25c3e988-ae27-4369-aa98-4bd7a3c5848a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.197835 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/607341d7-10b4-4d40-a2a2-7647ceff0c14-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "607341d7-10b4-4d40-a2a2-7647ceff0c14" (UID: "607341d7-10b4-4d40-a2a2-7647ceff0c14"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.202840 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25c3e988-ae27-4369-aa98-4bd7a3c5848a-kube-api-access-x2gng" (OuterVolumeSpecName: "kube-api-access-x2gng") pod "25c3e988-ae27-4369-aa98-4bd7a3c5848a" (UID: "25c3e988-ae27-4369-aa98-4bd7a3c5848a"). InnerVolumeSpecName "kube-api-access-x2gng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.204007 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/607341d7-10b4-4d40-a2a2-7647ceff0c14-kube-api-access-9w9mf" (OuterVolumeSpecName: "kube-api-access-9w9mf") pod "607341d7-10b4-4d40-a2a2-7647ceff0c14" (UID: "607341d7-10b4-4d40-a2a2-7647ceff0c14"). InnerVolumeSpecName "kube-api-access-9w9mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.297839 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/607341d7-10b4-4d40-a2a2-7647ceff0c14-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.297893 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2gng\" (UniqueName: \"kubernetes.io/projected/25c3e988-ae27-4369-aa98-4bd7a3c5848a-kube-api-access-x2gng\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.297907 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9w9mf\" (UniqueName: \"kubernetes.io/projected/607341d7-10b4-4d40-a2a2-7647ceff0c14-kube-api-access-9w9mf\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.297916 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25c3e988-ae27-4369-aa98-4bd7a3c5848a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.768287 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9lwc4" Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.768290 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9lwc4" event={"ID":"607341d7-10b4-4d40-a2a2-7647ceff0c14","Type":"ContainerDied","Data":"bd58bff5762edaa305510da137be49450ec3edcb712d5ef84473a1df00b19f59"} Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.768933 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd58bff5762edaa305510da137be49450ec3edcb712d5ef84473a1df00b19f59" Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.770019 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-481d-account-create-update-2rdl9" event={"ID":"25c3e988-ae27-4369-aa98-4bd7a3c5848a","Type":"ContainerDied","Data":"6302902a1abb26b7c265c363da1edb911b21baa6aa2370688f75716b518ffac4"} Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.770050 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6302902a1abb26b7c265c363da1edb911b21baa6aa2370688f75716b518ffac4" Dec 01 16:13:17 crc kubenswrapper[4810]: I1201 16:13:17.770107 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-481d-account-create-update-2rdl9" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.142043 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-4q58g"] Dec 01 16:13:19 crc kubenswrapper[4810]: E1201 16:13:19.142491 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="607341d7-10b4-4d40-a2a2-7647ceff0c14" containerName="mariadb-database-create" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.142506 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="607341d7-10b4-4d40-a2a2-7647ceff0c14" containerName="mariadb-database-create" Dec 01 16:13:19 crc kubenswrapper[4810]: E1201 16:13:19.142525 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25c3e988-ae27-4369-aa98-4bd7a3c5848a" containerName="mariadb-account-create-update" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.142531 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="25c3e988-ae27-4369-aa98-4bd7a3c5848a" containerName="mariadb-account-create-update" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.142694 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="25c3e988-ae27-4369-aa98-4bd7a3c5848a" containerName="mariadb-account-create-update" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.142718 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="607341d7-10b4-4d40-a2a2-7647ceff0c14" containerName="mariadb-database-create" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.143274 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4q58g" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.145179 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.145611 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-f847x" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.162742 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4q58g"] Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.333190 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/908c7ff6-8b25-4011-b1da-159a16755ac5-config-data\") pod \"glance-db-sync-4q58g\" (UID: \"908c7ff6-8b25-4011-b1da-159a16755ac5\") " pod="openstack/glance-db-sync-4q58g" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.333264 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcstm\" (UniqueName: \"kubernetes.io/projected/908c7ff6-8b25-4011-b1da-159a16755ac5-kube-api-access-rcstm\") pod \"glance-db-sync-4q58g\" (UID: \"908c7ff6-8b25-4011-b1da-159a16755ac5\") " pod="openstack/glance-db-sync-4q58g" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.333321 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/908c7ff6-8b25-4011-b1da-159a16755ac5-db-sync-config-data\") pod \"glance-db-sync-4q58g\" (UID: \"908c7ff6-8b25-4011-b1da-159a16755ac5\") " pod="openstack/glance-db-sync-4q58g" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.333350 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/908c7ff6-8b25-4011-b1da-159a16755ac5-combined-ca-bundle\") pod \"glance-db-sync-4q58g\" (UID: \"908c7ff6-8b25-4011-b1da-159a16755ac5\") " pod="openstack/glance-db-sync-4q58g" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.435202 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/908c7ff6-8b25-4011-b1da-159a16755ac5-config-data\") pod \"glance-db-sync-4q58g\" (UID: \"908c7ff6-8b25-4011-b1da-159a16755ac5\") " pod="openstack/glance-db-sync-4q58g" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.435265 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcstm\" (UniqueName: \"kubernetes.io/projected/908c7ff6-8b25-4011-b1da-159a16755ac5-kube-api-access-rcstm\") pod \"glance-db-sync-4q58g\" (UID: \"908c7ff6-8b25-4011-b1da-159a16755ac5\") " pod="openstack/glance-db-sync-4q58g" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.435357 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/908c7ff6-8b25-4011-b1da-159a16755ac5-db-sync-config-data\") pod \"glance-db-sync-4q58g\" (UID: \"908c7ff6-8b25-4011-b1da-159a16755ac5\") " pod="openstack/glance-db-sync-4q58g" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.435387 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/908c7ff6-8b25-4011-b1da-159a16755ac5-combined-ca-bundle\") pod \"glance-db-sync-4q58g\" (UID: \"908c7ff6-8b25-4011-b1da-159a16755ac5\") " pod="openstack/glance-db-sync-4q58g" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.440923 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/908c7ff6-8b25-4011-b1da-159a16755ac5-combined-ca-bundle\") pod \"glance-db-sync-4q58g\" (UID: \"908c7ff6-8b25-4011-b1da-159a16755ac5\") " pod="openstack/glance-db-sync-4q58g" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.440923 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/908c7ff6-8b25-4011-b1da-159a16755ac5-config-data\") pod \"glance-db-sync-4q58g\" (UID: \"908c7ff6-8b25-4011-b1da-159a16755ac5\") " pod="openstack/glance-db-sync-4q58g" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.442455 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/908c7ff6-8b25-4011-b1da-159a16755ac5-db-sync-config-data\") pod \"glance-db-sync-4q58g\" (UID: \"908c7ff6-8b25-4011-b1da-159a16755ac5\") " pod="openstack/glance-db-sync-4q58g" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.459932 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcstm\" (UniqueName: \"kubernetes.io/projected/908c7ff6-8b25-4011-b1da-159a16755ac5-kube-api-access-rcstm\") pod \"glance-db-sync-4q58g\" (UID: \"908c7ff6-8b25-4011-b1da-159a16755ac5\") " pod="openstack/glance-db-sync-4q58g" Dec 01 16:13:19 crc kubenswrapper[4810]: I1201 16:13:19.465808 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4q58g" Dec 01 16:13:20 crc kubenswrapper[4810]: I1201 16:13:20.029380 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4q58g"] Dec 01 16:13:20 crc kubenswrapper[4810]: W1201 16:13:20.032186 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod908c7ff6_8b25_4011_b1da_159a16755ac5.slice/crio-87b6f8d58088613a6bbea515aac919bd5c667055288fec69dde334b1a34d5ba7 WatchSource:0}: Error finding container 87b6f8d58088613a6bbea515aac919bd5c667055288fec69dde334b1a34d5ba7: Status 404 returned error can't find the container with id 87b6f8d58088613a6bbea515aac919bd5c667055288fec69dde334b1a34d5ba7 Dec 01 16:13:20 crc kubenswrapper[4810]: I1201 16:13:20.797653 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4q58g" event={"ID":"908c7ff6-8b25-4011-b1da-159a16755ac5","Type":"ContainerStarted","Data":"ac95e8544d2d9d177ad553c30d64e480d85ac3dfaf5c112025a6760b5aab16a2"} Dec 01 16:13:20 crc kubenswrapper[4810]: I1201 16:13:20.797985 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4q58g" event={"ID":"908c7ff6-8b25-4011-b1da-159a16755ac5","Type":"ContainerStarted","Data":"87b6f8d58088613a6bbea515aac919bd5c667055288fec69dde334b1a34d5ba7"} Dec 01 16:13:20 crc kubenswrapper[4810]: I1201 16:13:20.818404 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-4q58g" podStartSLOduration=1.81838175 podStartE2EDuration="1.81838175s" podCreationTimestamp="2025-12-01 16:13:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:13:20.813316672 +0000 UTC m=+5966.576826275" watchObservedRunningTime="2025-12-01 16:13:20.81838175 +0000 UTC m=+5966.581891363" Dec 01 16:13:23 crc kubenswrapper[4810]: I1201 16:13:23.823757 4810 generic.go:334] "Generic (PLEG): container finished" podID="908c7ff6-8b25-4011-b1da-159a16755ac5" containerID="ac95e8544d2d9d177ad553c30d64e480d85ac3dfaf5c112025a6760b5aab16a2" exitCode=0 Dec 01 16:13:23 crc kubenswrapper[4810]: I1201 16:13:23.823836 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4q58g" event={"ID":"908c7ff6-8b25-4011-b1da-159a16755ac5","Type":"ContainerDied","Data":"ac95e8544d2d9d177ad553c30d64e480d85ac3dfaf5c112025a6760b5aab16a2"} Dec 01 16:13:25 crc kubenswrapper[4810]: I1201 16:13:25.211242 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4q58g" Dec 01 16:13:25 crc kubenswrapper[4810]: I1201 16:13:25.337412 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/908c7ff6-8b25-4011-b1da-159a16755ac5-config-data\") pod \"908c7ff6-8b25-4011-b1da-159a16755ac5\" (UID: \"908c7ff6-8b25-4011-b1da-159a16755ac5\") " Dec 01 16:13:25 crc kubenswrapper[4810]: I1201 16:13:25.337638 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/908c7ff6-8b25-4011-b1da-159a16755ac5-combined-ca-bundle\") pod \"908c7ff6-8b25-4011-b1da-159a16755ac5\" (UID: \"908c7ff6-8b25-4011-b1da-159a16755ac5\") " Dec 01 16:13:25 crc kubenswrapper[4810]: I1201 16:13:25.337682 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/908c7ff6-8b25-4011-b1da-159a16755ac5-db-sync-config-data\") pod \"908c7ff6-8b25-4011-b1da-159a16755ac5\" (UID: \"908c7ff6-8b25-4011-b1da-159a16755ac5\") " Dec 01 16:13:25 crc kubenswrapper[4810]: I1201 16:13:25.337710 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcstm\" (UniqueName: \"kubernetes.io/projected/908c7ff6-8b25-4011-b1da-159a16755ac5-kube-api-access-rcstm\") pod \"908c7ff6-8b25-4011-b1da-159a16755ac5\" (UID: \"908c7ff6-8b25-4011-b1da-159a16755ac5\") " Dec 01 16:13:25 crc kubenswrapper[4810]: I1201 16:13:25.342346 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/908c7ff6-8b25-4011-b1da-159a16755ac5-kube-api-access-rcstm" (OuterVolumeSpecName: "kube-api-access-rcstm") pod "908c7ff6-8b25-4011-b1da-159a16755ac5" (UID: "908c7ff6-8b25-4011-b1da-159a16755ac5"). InnerVolumeSpecName "kube-api-access-rcstm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:13:25 crc kubenswrapper[4810]: I1201 16:13:25.343248 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/908c7ff6-8b25-4011-b1da-159a16755ac5-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "908c7ff6-8b25-4011-b1da-159a16755ac5" (UID: "908c7ff6-8b25-4011-b1da-159a16755ac5"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:25 crc kubenswrapper[4810]: I1201 16:13:25.362315 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/908c7ff6-8b25-4011-b1da-159a16755ac5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "908c7ff6-8b25-4011-b1da-159a16755ac5" (UID: "908c7ff6-8b25-4011-b1da-159a16755ac5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:25 crc kubenswrapper[4810]: I1201 16:13:25.383436 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/908c7ff6-8b25-4011-b1da-159a16755ac5-config-data" (OuterVolumeSpecName: "config-data") pod "908c7ff6-8b25-4011-b1da-159a16755ac5" (UID: "908c7ff6-8b25-4011-b1da-159a16755ac5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:25 crc kubenswrapper[4810]: I1201 16:13:25.439971 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/908c7ff6-8b25-4011-b1da-159a16755ac5-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:25 crc kubenswrapper[4810]: I1201 16:13:25.440006 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/908c7ff6-8b25-4011-b1da-159a16755ac5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:25 crc kubenswrapper[4810]: I1201 16:13:25.440025 4810 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/908c7ff6-8b25-4011-b1da-159a16755ac5-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:25 crc kubenswrapper[4810]: I1201 16:13:25.440037 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcstm\" (UniqueName: \"kubernetes.io/projected/908c7ff6-8b25-4011-b1da-159a16755ac5-kube-api-access-rcstm\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:25 crc kubenswrapper[4810]: I1201 16:13:25.843090 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4q58g" event={"ID":"908c7ff6-8b25-4011-b1da-159a16755ac5","Type":"ContainerDied","Data":"87b6f8d58088613a6bbea515aac919bd5c667055288fec69dde334b1a34d5ba7"} Dec 01 16:13:25 crc kubenswrapper[4810]: I1201 16:13:25.843412 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87b6f8d58088613a6bbea515aac919bd5c667055288fec69dde334b1a34d5ba7" Dec 01 16:13:25 crc kubenswrapper[4810]: I1201 16:13:25.843275 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4q58g" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.126070 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:13:26 crc kubenswrapper[4810]: E1201 16:13:26.126847 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="908c7ff6-8b25-4011-b1da-159a16755ac5" containerName="glance-db-sync" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.126865 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="908c7ff6-8b25-4011-b1da-159a16755ac5" containerName="glance-db-sync" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.127074 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="908c7ff6-8b25-4011-b1da-159a16755ac5" containerName="glance-db-sync" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.128259 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.136335 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.136582 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-f847x" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.136706 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.147293 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.235815 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9d778847c-h4rtf"] Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.243027 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.257336 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9d778847c-h4rtf"] Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.269156 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11a558de-f7fe-46a8-9bba-2a50e6a5b538-scripts\") pod \"glance-default-external-api-0\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.269355 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11a558de-f7fe-46a8-9bba-2a50e6a5b538-config-data\") pod \"glance-default-external-api-0\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.269603 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/11a558de-f7fe-46a8-9bba-2a50e6a5b538-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.269750 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11a558de-f7fe-46a8-9bba-2a50e6a5b538-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.269902 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqbxw\" (UniqueName: \"kubernetes.io/projected/11a558de-f7fe-46a8-9bba-2a50e6a5b538-kube-api-access-hqbxw\") pod \"glance-default-external-api-0\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.270011 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11a558de-f7fe-46a8-9bba-2a50e6a5b538-logs\") pod \"glance-default-external-api-0\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.335610 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.337461 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.341947 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.353084 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.371946 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqbxw\" (UniqueName: \"kubernetes.io/projected/11a558de-f7fe-46a8-9bba-2a50e6a5b538-kube-api-access-hqbxw\") pod \"glance-default-external-api-0\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.372005 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-ovsdbserver-nb\") pod \"dnsmasq-dns-9d778847c-h4rtf\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.372039 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11a558de-f7fe-46a8-9bba-2a50e6a5b538-logs\") pod \"glance-default-external-api-0\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.372102 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-ovsdbserver-sb\") pod \"dnsmasq-dns-9d778847c-h4rtf\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.372130 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11a558de-f7fe-46a8-9bba-2a50e6a5b538-scripts\") pod \"glance-default-external-api-0\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.372160 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11a558de-f7fe-46a8-9bba-2a50e6a5b538-config-data\") pod \"glance-default-external-api-0\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.372190 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-dns-svc\") pod \"dnsmasq-dns-9d778847c-h4rtf\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.372270 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/11a558de-f7fe-46a8-9bba-2a50e6a5b538-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.372309 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11a558de-f7fe-46a8-9bba-2a50e6a5b538-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.372358 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-config\") pod \"dnsmasq-dns-9d778847c-h4rtf\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.372383 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cqzk\" (UniqueName: \"kubernetes.io/projected/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-kube-api-access-7cqzk\") pod \"dnsmasq-dns-9d778847c-h4rtf\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.373271 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11a558de-f7fe-46a8-9bba-2a50e6a5b538-logs\") pod \"glance-default-external-api-0\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.373597 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/11a558de-f7fe-46a8-9bba-2a50e6a5b538-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.398276 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11a558de-f7fe-46a8-9bba-2a50e6a5b538-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.403365 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqbxw\" (UniqueName: \"kubernetes.io/projected/11a558de-f7fe-46a8-9bba-2a50e6a5b538-kube-api-access-hqbxw\") pod \"glance-default-external-api-0\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.406247 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11a558de-f7fe-46a8-9bba-2a50e6a5b538-config-data\") pod \"glance-default-external-api-0\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.406302 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11a558de-f7fe-46a8-9bba-2a50e6a5b538-scripts\") pod \"glance-default-external-api-0\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.460078 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.473932 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/571a4966-bf13-46d1-835f-2b48e49026d8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.474149 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/571a4966-bf13-46d1-835f-2b48e49026d8-logs\") pod \"glance-default-internal-api-0\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.474167 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/571a4966-bf13-46d1-835f-2b48e49026d8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.474202 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-ovsdbserver-sb\") pod \"dnsmasq-dns-9d778847c-h4rtf\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.474239 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-dns-svc\") pod \"dnsmasq-dns-9d778847c-h4rtf\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.474295 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/571a4966-bf13-46d1-835f-2b48e49026d8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.474348 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-config\") pod \"dnsmasq-dns-9d778847c-h4rtf\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.474367 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cqzk\" (UniqueName: \"kubernetes.io/projected/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-kube-api-access-7cqzk\") pod \"dnsmasq-dns-9d778847c-h4rtf\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.474385 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtdjw\" (UniqueName: \"kubernetes.io/projected/571a4966-bf13-46d1-835f-2b48e49026d8-kube-api-access-mtdjw\") pod \"glance-default-internal-api-0\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.474405 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/571a4966-bf13-46d1-835f-2b48e49026d8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.474421 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-ovsdbserver-nb\") pod \"dnsmasq-dns-9d778847c-h4rtf\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.474972 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-ovsdbserver-sb\") pod \"dnsmasq-dns-9d778847c-h4rtf\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.475173 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-ovsdbserver-nb\") pod \"dnsmasq-dns-9d778847c-h4rtf\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.475327 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-dns-svc\") pod \"dnsmasq-dns-9d778847c-h4rtf\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.475618 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-config\") pod \"dnsmasq-dns-9d778847c-h4rtf\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.496176 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cqzk\" (UniqueName: \"kubernetes.io/projected/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-kube-api-access-7cqzk\") pod \"dnsmasq-dns-9d778847c-h4rtf\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.577658 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/571a4966-bf13-46d1-835f-2b48e49026d8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.577813 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtdjw\" (UniqueName: \"kubernetes.io/projected/571a4966-bf13-46d1-835f-2b48e49026d8-kube-api-access-mtdjw\") pod \"glance-default-internal-api-0\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.577841 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/571a4966-bf13-46d1-835f-2b48e49026d8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.577881 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/571a4966-bf13-46d1-835f-2b48e49026d8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.577938 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/571a4966-bf13-46d1-835f-2b48e49026d8-logs\") pod \"glance-default-internal-api-0\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.577968 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/571a4966-bf13-46d1-835f-2b48e49026d8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.581381 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/571a4966-bf13-46d1-835f-2b48e49026d8-logs\") pod \"glance-default-internal-api-0\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.581446 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/571a4966-bf13-46d1-835f-2b48e49026d8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.584240 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.590729 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/571a4966-bf13-46d1-835f-2b48e49026d8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.594228 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/571a4966-bf13-46d1-835f-2b48e49026d8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.594712 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/571a4966-bf13-46d1-835f-2b48e49026d8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.605682 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtdjw\" (UniqueName: \"kubernetes.io/projected/571a4966-bf13-46d1-835f-2b48e49026d8-kube-api-access-mtdjw\") pod \"glance-default-internal-api-0\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:26 crc kubenswrapper[4810]: I1201 16:13:26.677945 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 16:13:27 crc kubenswrapper[4810]: I1201 16:13:27.064124 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:13:27 crc kubenswrapper[4810]: W1201 16:13:27.149099 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba2acf0a_12c2_4ab0_9f5f_31f80a281b1c.slice/crio-437bcdee2132857316e182d4df9ac56aa2dfffa66b661eebacbee09d7091b81d WatchSource:0}: Error finding container 437bcdee2132857316e182d4df9ac56aa2dfffa66b661eebacbee09d7091b81d: Status 404 returned error can't find the container with id 437bcdee2132857316e182d4df9ac56aa2dfffa66b661eebacbee09d7091b81d Dec 01 16:13:27 crc kubenswrapper[4810]: I1201 16:13:27.154301 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9d778847c-h4rtf"] Dec 01 16:13:27 crc kubenswrapper[4810]: I1201 16:13:27.289920 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:13:27 crc kubenswrapper[4810]: I1201 16:13:27.387147 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:13:27 crc kubenswrapper[4810]: W1201 16:13:27.421458 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod571a4966_bf13_46d1_835f_2b48e49026d8.slice/crio-6680e2d848418d8348d8449ecdfb8d6d5c84638d536949ecbd7f3d59d5384dd6 WatchSource:0}: Error finding container 6680e2d848418d8348d8449ecdfb8d6d5c84638d536949ecbd7f3d59d5384dd6: Status 404 returned error can't find the container with id 6680e2d848418d8348d8449ecdfb8d6d5c84638d536949ecbd7f3d59d5384dd6 Dec 01 16:13:27 crc kubenswrapper[4810]: I1201 16:13:27.868089 4810 generic.go:334] "Generic (PLEG): container finished" podID="ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c" containerID="89e28ad83ce9149c604aea27f51aaffb5aa5ae0fea9e7d56a806354d291b4ab2" exitCode=0 Dec 01 16:13:27 crc kubenswrapper[4810]: I1201 16:13:27.868401 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9d778847c-h4rtf" event={"ID":"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c","Type":"ContainerDied","Data":"89e28ad83ce9149c604aea27f51aaffb5aa5ae0fea9e7d56a806354d291b4ab2"} Dec 01 16:13:27 crc kubenswrapper[4810]: I1201 16:13:27.868940 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9d778847c-h4rtf" event={"ID":"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c","Type":"ContainerStarted","Data":"437bcdee2132857316e182d4df9ac56aa2dfffa66b661eebacbee09d7091b81d"} Dec 01 16:13:27 crc kubenswrapper[4810]: I1201 16:13:27.871812 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"571a4966-bf13-46d1-835f-2b48e49026d8","Type":"ContainerStarted","Data":"6680e2d848418d8348d8449ecdfb8d6d5c84638d536949ecbd7f3d59d5384dd6"} Dec 01 16:13:27 crc kubenswrapper[4810]: I1201 16:13:27.880332 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"11a558de-f7fe-46a8-9bba-2a50e6a5b538","Type":"ContainerStarted","Data":"086ece90306279f8400c73cab985107d063fac0d2df2a38ff5b7abbd53c6d25f"} Dec 01 16:13:27 crc kubenswrapper[4810]: I1201 16:13:27.880366 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"11a558de-f7fe-46a8-9bba-2a50e6a5b538","Type":"ContainerStarted","Data":"13ae56c9da5c47b6f5c5fa46b84a4d645f1cdee44562bea87995c16484481484"} Dec 01 16:13:28 crc kubenswrapper[4810]: I1201 16:13:28.418873 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:13:28 crc kubenswrapper[4810]: I1201 16:13:28.915952 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"571a4966-bf13-46d1-835f-2b48e49026d8","Type":"ContainerStarted","Data":"1b1ccf698e5c9c4b6c81c4e029a18bd37c5e89aa989b0cea0dac9166e344c1e4"} Dec 01 16:13:28 crc kubenswrapper[4810]: I1201 16:13:28.916320 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"571a4966-bf13-46d1-835f-2b48e49026d8","Type":"ContainerStarted","Data":"4168216fde749b55a21ef4efe91c73ec297ea6acc6c91a789334debc56ba6fd1"} Dec 01 16:13:28 crc kubenswrapper[4810]: I1201 16:13:28.916460 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="571a4966-bf13-46d1-835f-2b48e49026d8" containerName="glance-log" containerID="cri-o://4168216fde749b55a21ef4efe91c73ec297ea6acc6c91a789334debc56ba6fd1" gracePeriod=30 Dec 01 16:13:28 crc kubenswrapper[4810]: I1201 16:13:28.916820 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="571a4966-bf13-46d1-835f-2b48e49026d8" containerName="glance-httpd" containerID="cri-o://1b1ccf698e5c9c4b6c81c4e029a18bd37c5e89aa989b0cea0dac9166e344c1e4" gracePeriod=30 Dec 01 16:13:28 crc kubenswrapper[4810]: I1201 16:13:28.939765 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="11a558de-f7fe-46a8-9bba-2a50e6a5b538" containerName="glance-log" containerID="cri-o://086ece90306279f8400c73cab985107d063fac0d2df2a38ff5b7abbd53c6d25f" gracePeriod=30 Dec 01 16:13:28 crc kubenswrapper[4810]: I1201 16:13:28.940107 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"11a558de-f7fe-46a8-9bba-2a50e6a5b538","Type":"ContainerStarted","Data":"e81aa157c0ccd3b1e4d888c13bc7fd3d52d578db1d3d09a4f9e8551ad7766e04"} Dec 01 16:13:28 crc kubenswrapper[4810]: I1201 16:13:28.940168 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="11a558de-f7fe-46a8-9bba-2a50e6a5b538" containerName="glance-httpd" containerID="cri-o://e81aa157c0ccd3b1e4d888c13bc7fd3d52d578db1d3d09a4f9e8551ad7766e04" gracePeriod=30 Dec 01 16:13:28 crc kubenswrapper[4810]: I1201 16:13:28.971257 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9d778847c-h4rtf" event={"ID":"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c","Type":"ContainerStarted","Data":"ff9363d79679c72ac1691d510f17ab6f5e9d55f30422534462c18f9889dd12f7"} Dec 01 16:13:28 crc kubenswrapper[4810]: I1201 16:13:28.972148 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.006723 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.006706537 podStartE2EDuration="3.006706537s" podCreationTimestamp="2025-12-01 16:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:13:28.96628471 +0000 UTC m=+5974.729794333" watchObservedRunningTime="2025-12-01 16:13:29.006706537 +0000 UTC m=+5974.770216140" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.038011 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.037994957 podStartE2EDuration="3.037994957s" podCreationTimestamp="2025-12-01 16:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:13:29.037850583 +0000 UTC m=+5974.801360186" watchObservedRunningTime="2025-12-01 16:13:29.037994957 +0000 UTC m=+5974.801504560" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.099340 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9d778847c-h4rtf" podStartSLOduration=3.099319841 podStartE2EDuration="3.099319841s" podCreationTimestamp="2025-12-01 16:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:13:29.087397337 +0000 UTC m=+5974.850906940" watchObservedRunningTime="2025-12-01 16:13:29.099319841 +0000 UTC m=+5974.862829444" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.531328 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.664142 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/571a4966-bf13-46d1-835f-2b48e49026d8-combined-ca-bundle\") pod \"571a4966-bf13-46d1-835f-2b48e49026d8\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.664226 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/571a4966-bf13-46d1-835f-2b48e49026d8-scripts\") pod \"571a4966-bf13-46d1-835f-2b48e49026d8\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.664270 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/571a4966-bf13-46d1-835f-2b48e49026d8-config-data\") pod \"571a4966-bf13-46d1-835f-2b48e49026d8\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.664366 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtdjw\" (UniqueName: \"kubernetes.io/projected/571a4966-bf13-46d1-835f-2b48e49026d8-kube-api-access-mtdjw\") pod \"571a4966-bf13-46d1-835f-2b48e49026d8\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.664402 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/571a4966-bf13-46d1-835f-2b48e49026d8-httpd-run\") pod \"571a4966-bf13-46d1-835f-2b48e49026d8\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.664461 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/571a4966-bf13-46d1-835f-2b48e49026d8-logs\") pod \"571a4966-bf13-46d1-835f-2b48e49026d8\" (UID: \"571a4966-bf13-46d1-835f-2b48e49026d8\") " Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.666271 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/571a4966-bf13-46d1-835f-2b48e49026d8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "571a4966-bf13-46d1-835f-2b48e49026d8" (UID: "571a4966-bf13-46d1-835f-2b48e49026d8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.666327 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/571a4966-bf13-46d1-835f-2b48e49026d8-logs" (OuterVolumeSpecName: "logs") pod "571a4966-bf13-46d1-835f-2b48e49026d8" (UID: "571a4966-bf13-46d1-835f-2b48e49026d8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.675538 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/571a4966-bf13-46d1-835f-2b48e49026d8-kube-api-access-mtdjw" (OuterVolumeSpecName: "kube-api-access-mtdjw") pod "571a4966-bf13-46d1-835f-2b48e49026d8" (UID: "571a4966-bf13-46d1-835f-2b48e49026d8"). InnerVolumeSpecName "kube-api-access-mtdjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.675614 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/571a4966-bf13-46d1-835f-2b48e49026d8-scripts" (OuterVolumeSpecName: "scripts") pod "571a4966-bf13-46d1-835f-2b48e49026d8" (UID: "571a4966-bf13-46d1-835f-2b48e49026d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.700836 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/571a4966-bf13-46d1-835f-2b48e49026d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "571a4966-bf13-46d1-835f-2b48e49026d8" (UID: "571a4966-bf13-46d1-835f-2b48e49026d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.723178 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/571a4966-bf13-46d1-835f-2b48e49026d8-config-data" (OuterVolumeSpecName: "config-data") pod "571a4966-bf13-46d1-835f-2b48e49026d8" (UID: "571a4966-bf13-46d1-835f-2b48e49026d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.766252 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/571a4966-bf13-46d1-835f-2b48e49026d8-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.766290 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/571a4966-bf13-46d1-835f-2b48e49026d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.766305 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/571a4966-bf13-46d1-835f-2b48e49026d8-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.766314 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/571a4966-bf13-46d1-835f-2b48e49026d8-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.766324 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtdjw\" (UniqueName: \"kubernetes.io/projected/571a4966-bf13-46d1-835f-2b48e49026d8-kube-api-access-mtdjw\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.766332 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/571a4966-bf13-46d1-835f-2b48e49026d8-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.797754 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.969940 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqbxw\" (UniqueName: \"kubernetes.io/projected/11a558de-f7fe-46a8-9bba-2a50e6a5b538-kube-api-access-hqbxw\") pod \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.970112 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11a558de-f7fe-46a8-9bba-2a50e6a5b538-scripts\") pod \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.970198 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11a558de-f7fe-46a8-9bba-2a50e6a5b538-combined-ca-bundle\") pod \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.970248 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11a558de-f7fe-46a8-9bba-2a50e6a5b538-logs\") pod \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.970268 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/11a558de-f7fe-46a8-9bba-2a50e6a5b538-httpd-run\") pod \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.970284 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11a558de-f7fe-46a8-9bba-2a50e6a5b538-config-data\") pod \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\" (UID: \"11a558de-f7fe-46a8-9bba-2a50e6a5b538\") " Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.970622 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11a558de-f7fe-46a8-9bba-2a50e6a5b538-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "11a558de-f7fe-46a8-9bba-2a50e6a5b538" (UID: "11a558de-f7fe-46a8-9bba-2a50e6a5b538"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.970648 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11a558de-f7fe-46a8-9bba-2a50e6a5b538-logs" (OuterVolumeSpecName: "logs") pod "11a558de-f7fe-46a8-9bba-2a50e6a5b538" (UID: "11a558de-f7fe-46a8-9bba-2a50e6a5b538"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.970857 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11a558de-f7fe-46a8-9bba-2a50e6a5b538-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.970872 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/11a558de-f7fe-46a8-9bba-2a50e6a5b538-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.977262 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11a558de-f7fe-46a8-9bba-2a50e6a5b538-kube-api-access-hqbxw" (OuterVolumeSpecName: "kube-api-access-hqbxw") pod "11a558de-f7fe-46a8-9bba-2a50e6a5b538" (UID: "11a558de-f7fe-46a8-9bba-2a50e6a5b538"). InnerVolumeSpecName "kube-api-access-hqbxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.978898 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11a558de-f7fe-46a8-9bba-2a50e6a5b538-scripts" (OuterVolumeSpecName: "scripts") pod "11a558de-f7fe-46a8-9bba-2a50e6a5b538" (UID: "11a558de-f7fe-46a8-9bba-2a50e6a5b538"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.983004 4810 generic.go:334] "Generic (PLEG): container finished" podID="11a558de-f7fe-46a8-9bba-2a50e6a5b538" containerID="e81aa157c0ccd3b1e4d888c13bc7fd3d52d578db1d3d09a4f9e8551ad7766e04" exitCode=0 Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.983037 4810 generic.go:334] "Generic (PLEG): container finished" podID="11a558de-f7fe-46a8-9bba-2a50e6a5b538" containerID="086ece90306279f8400c73cab985107d063fac0d2df2a38ff5b7abbd53c6d25f" exitCode=143 Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.983088 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"11a558de-f7fe-46a8-9bba-2a50e6a5b538","Type":"ContainerDied","Data":"e81aa157c0ccd3b1e4d888c13bc7fd3d52d578db1d3d09a4f9e8551ad7766e04"} Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.983117 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"11a558de-f7fe-46a8-9bba-2a50e6a5b538","Type":"ContainerDied","Data":"086ece90306279f8400c73cab985107d063fac0d2df2a38ff5b7abbd53c6d25f"} Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.983128 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"11a558de-f7fe-46a8-9bba-2a50e6a5b538","Type":"ContainerDied","Data":"13ae56c9da5c47b6f5c5fa46b84a4d645f1cdee44562bea87995c16484481484"} Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.983089 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.983135 4810 scope.go:117] "RemoveContainer" containerID="e81aa157c0ccd3b1e4d888c13bc7fd3d52d578db1d3d09a4f9e8551ad7766e04" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.987254 4810 generic.go:334] "Generic (PLEG): container finished" podID="571a4966-bf13-46d1-835f-2b48e49026d8" containerID="1b1ccf698e5c9c4b6c81c4e029a18bd37c5e89aa989b0cea0dac9166e344c1e4" exitCode=143 Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.987298 4810 generic.go:334] "Generic (PLEG): container finished" podID="571a4966-bf13-46d1-835f-2b48e49026d8" containerID="4168216fde749b55a21ef4efe91c73ec297ea6acc6c91a789334debc56ba6fd1" exitCode=143 Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.987310 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.987339 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"571a4966-bf13-46d1-835f-2b48e49026d8","Type":"ContainerDied","Data":"1b1ccf698e5c9c4b6c81c4e029a18bd37c5e89aa989b0cea0dac9166e344c1e4"} Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.987370 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"571a4966-bf13-46d1-835f-2b48e49026d8","Type":"ContainerDied","Data":"4168216fde749b55a21ef4efe91c73ec297ea6acc6c91a789334debc56ba6fd1"} Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.987385 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"571a4966-bf13-46d1-835f-2b48e49026d8","Type":"ContainerDied","Data":"6680e2d848418d8348d8449ecdfb8d6d5c84638d536949ecbd7f3d59d5384dd6"} Dec 01 16:13:29 crc kubenswrapper[4810]: I1201 16:13:29.997752 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11a558de-f7fe-46a8-9bba-2a50e6a5b538-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11a558de-f7fe-46a8-9bba-2a50e6a5b538" (UID: "11a558de-f7fe-46a8-9bba-2a50e6a5b538"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.027765 4810 scope.go:117] "RemoveContainer" containerID="086ece90306279f8400c73cab985107d063fac0d2df2a38ff5b7abbd53c6d25f" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.033088 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.036098 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11a558de-f7fe-46a8-9bba-2a50e6a5b538-config-data" (OuterVolumeSpecName: "config-data") pod "11a558de-f7fe-46a8-9bba-2a50e6a5b538" (UID: "11a558de-f7fe-46a8-9bba-2a50e6a5b538"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.046906 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.055898 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:13:30 crc kubenswrapper[4810]: E1201 16:13:30.056377 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="571a4966-bf13-46d1-835f-2b48e49026d8" containerName="glance-log" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.056403 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="571a4966-bf13-46d1-835f-2b48e49026d8" containerName="glance-log" Dec 01 16:13:30 crc kubenswrapper[4810]: E1201 16:13:30.056424 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11a558de-f7fe-46a8-9bba-2a50e6a5b538" containerName="glance-log" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.056436 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="11a558de-f7fe-46a8-9bba-2a50e6a5b538" containerName="glance-log" Dec 01 16:13:30 crc kubenswrapper[4810]: E1201 16:13:30.056453 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11a558de-f7fe-46a8-9bba-2a50e6a5b538" containerName="glance-httpd" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.056460 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="11a558de-f7fe-46a8-9bba-2a50e6a5b538" containerName="glance-httpd" Dec 01 16:13:30 crc kubenswrapper[4810]: E1201 16:13:30.056498 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="571a4966-bf13-46d1-835f-2b48e49026d8" containerName="glance-httpd" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.056506 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="571a4966-bf13-46d1-835f-2b48e49026d8" containerName="glance-httpd" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.056695 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="11a558de-f7fe-46a8-9bba-2a50e6a5b538" containerName="glance-httpd" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.056728 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="11a558de-f7fe-46a8-9bba-2a50e6a5b538" containerName="glance-log" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.056741 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="571a4966-bf13-46d1-835f-2b48e49026d8" containerName="glance-httpd" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.056753 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="571a4966-bf13-46d1-835f-2b48e49026d8" containerName="glance-log" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.057704 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.060265 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.061644 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.064604 4810 scope.go:117] "RemoveContainer" containerID="e81aa157c0ccd3b1e4d888c13bc7fd3d52d578db1d3d09a4f9e8551ad7766e04" Dec 01 16:13:30 crc kubenswrapper[4810]: E1201 16:13:30.065021 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e81aa157c0ccd3b1e4d888c13bc7fd3d52d578db1d3d09a4f9e8551ad7766e04\": container with ID starting with e81aa157c0ccd3b1e4d888c13bc7fd3d52d578db1d3d09a4f9e8551ad7766e04 not found: ID does not exist" containerID="e81aa157c0ccd3b1e4d888c13bc7fd3d52d578db1d3d09a4f9e8551ad7766e04" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.065054 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e81aa157c0ccd3b1e4d888c13bc7fd3d52d578db1d3d09a4f9e8551ad7766e04"} err="failed to get container status \"e81aa157c0ccd3b1e4d888c13bc7fd3d52d578db1d3d09a4f9e8551ad7766e04\": rpc error: code = NotFound desc = could not find container \"e81aa157c0ccd3b1e4d888c13bc7fd3d52d578db1d3d09a4f9e8551ad7766e04\": container with ID starting with e81aa157c0ccd3b1e4d888c13bc7fd3d52d578db1d3d09a4f9e8551ad7766e04 not found: ID does not exist" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.065080 4810 scope.go:117] "RemoveContainer" containerID="086ece90306279f8400c73cab985107d063fac0d2df2a38ff5b7abbd53c6d25f" Dec 01 16:13:30 crc kubenswrapper[4810]: E1201 16:13:30.065307 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"086ece90306279f8400c73cab985107d063fac0d2df2a38ff5b7abbd53c6d25f\": container with ID starting with 086ece90306279f8400c73cab985107d063fac0d2df2a38ff5b7abbd53c6d25f not found: ID does not exist" containerID="086ece90306279f8400c73cab985107d063fac0d2df2a38ff5b7abbd53c6d25f" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.065338 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"086ece90306279f8400c73cab985107d063fac0d2df2a38ff5b7abbd53c6d25f"} err="failed to get container status \"086ece90306279f8400c73cab985107d063fac0d2df2a38ff5b7abbd53c6d25f\": rpc error: code = NotFound desc = could not find container \"086ece90306279f8400c73cab985107d063fac0d2df2a38ff5b7abbd53c6d25f\": container with ID starting with 086ece90306279f8400c73cab985107d063fac0d2df2a38ff5b7abbd53c6d25f not found: ID does not exist" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.065354 4810 scope.go:117] "RemoveContainer" containerID="e81aa157c0ccd3b1e4d888c13bc7fd3d52d578db1d3d09a4f9e8551ad7766e04" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.065885 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.070747 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e81aa157c0ccd3b1e4d888c13bc7fd3d52d578db1d3d09a4f9e8551ad7766e04"} err="failed to get container status \"e81aa157c0ccd3b1e4d888c13bc7fd3d52d578db1d3d09a4f9e8551ad7766e04\": rpc error: code = NotFound desc = could not find container \"e81aa157c0ccd3b1e4d888c13bc7fd3d52d578db1d3d09a4f9e8551ad7766e04\": container with ID starting with e81aa157c0ccd3b1e4d888c13bc7fd3d52d578db1d3d09a4f9e8551ad7766e04 not found: ID does not exist" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.070947 4810 scope.go:117] "RemoveContainer" containerID="086ece90306279f8400c73cab985107d063fac0d2df2a38ff5b7abbd53c6d25f" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.071558 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"086ece90306279f8400c73cab985107d063fac0d2df2a38ff5b7abbd53c6d25f"} err="failed to get container status \"086ece90306279f8400c73cab985107d063fac0d2df2a38ff5b7abbd53c6d25f\": rpc error: code = NotFound desc = could not find container \"086ece90306279f8400c73cab985107d063fac0d2df2a38ff5b7abbd53c6d25f\": container with ID starting with 086ece90306279f8400c73cab985107d063fac0d2df2a38ff5b7abbd53c6d25f not found: ID does not exist" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.071623 4810 scope.go:117] "RemoveContainer" containerID="1b1ccf698e5c9c4b6c81c4e029a18bd37c5e89aa989b0cea0dac9166e344c1e4" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.072242 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11a558de-f7fe-46a8-9bba-2a50e6a5b538-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.072285 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11a558de-f7fe-46a8-9bba-2a50e6a5b538-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.072299 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11a558de-f7fe-46a8-9bba-2a50e6a5b538-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.072313 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqbxw\" (UniqueName: \"kubernetes.io/projected/11a558de-f7fe-46a8-9bba-2a50e6a5b538-kube-api-access-hqbxw\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.104461 4810 scope.go:117] "RemoveContainer" containerID="4168216fde749b55a21ef4efe91c73ec297ea6acc6c91a789334debc56ba6fd1" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.126709 4810 scope.go:117] "RemoveContainer" containerID="1b1ccf698e5c9c4b6c81c4e029a18bd37c5e89aa989b0cea0dac9166e344c1e4" Dec 01 16:13:30 crc kubenswrapper[4810]: E1201 16:13:30.132034 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b1ccf698e5c9c4b6c81c4e029a18bd37c5e89aa989b0cea0dac9166e344c1e4\": container with ID starting with 1b1ccf698e5c9c4b6c81c4e029a18bd37c5e89aa989b0cea0dac9166e344c1e4 not found: ID does not exist" containerID="1b1ccf698e5c9c4b6c81c4e029a18bd37c5e89aa989b0cea0dac9166e344c1e4" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.132070 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b1ccf698e5c9c4b6c81c4e029a18bd37c5e89aa989b0cea0dac9166e344c1e4"} err="failed to get container status \"1b1ccf698e5c9c4b6c81c4e029a18bd37c5e89aa989b0cea0dac9166e344c1e4\": rpc error: code = NotFound desc = could not find container \"1b1ccf698e5c9c4b6c81c4e029a18bd37c5e89aa989b0cea0dac9166e344c1e4\": container with ID starting with 1b1ccf698e5c9c4b6c81c4e029a18bd37c5e89aa989b0cea0dac9166e344c1e4 not found: ID does not exist" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.132094 4810 scope.go:117] "RemoveContainer" containerID="4168216fde749b55a21ef4efe91c73ec297ea6acc6c91a789334debc56ba6fd1" Dec 01 16:13:30 crc kubenswrapper[4810]: E1201 16:13:30.132806 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4168216fde749b55a21ef4efe91c73ec297ea6acc6c91a789334debc56ba6fd1\": container with ID starting with 4168216fde749b55a21ef4efe91c73ec297ea6acc6c91a789334debc56ba6fd1 not found: ID does not exist" containerID="4168216fde749b55a21ef4efe91c73ec297ea6acc6c91a789334debc56ba6fd1" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.132863 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4168216fde749b55a21ef4efe91c73ec297ea6acc6c91a789334debc56ba6fd1"} err="failed to get container status \"4168216fde749b55a21ef4efe91c73ec297ea6acc6c91a789334debc56ba6fd1\": rpc error: code = NotFound desc = could not find container \"4168216fde749b55a21ef4efe91c73ec297ea6acc6c91a789334debc56ba6fd1\": container with ID starting with 4168216fde749b55a21ef4efe91c73ec297ea6acc6c91a789334debc56ba6fd1 not found: ID does not exist" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.132897 4810 scope.go:117] "RemoveContainer" containerID="1b1ccf698e5c9c4b6c81c4e029a18bd37c5e89aa989b0cea0dac9166e344c1e4" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.133162 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b1ccf698e5c9c4b6c81c4e029a18bd37c5e89aa989b0cea0dac9166e344c1e4"} err="failed to get container status \"1b1ccf698e5c9c4b6c81c4e029a18bd37c5e89aa989b0cea0dac9166e344c1e4\": rpc error: code = NotFound desc = could not find container \"1b1ccf698e5c9c4b6c81c4e029a18bd37c5e89aa989b0cea0dac9166e344c1e4\": container with ID starting with 1b1ccf698e5c9c4b6c81c4e029a18bd37c5e89aa989b0cea0dac9166e344c1e4 not found: ID does not exist" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.133187 4810 scope.go:117] "RemoveContainer" containerID="4168216fde749b55a21ef4efe91c73ec297ea6acc6c91a789334debc56ba6fd1" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.133386 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4168216fde749b55a21ef4efe91c73ec297ea6acc6c91a789334debc56ba6fd1"} err="failed to get container status \"4168216fde749b55a21ef4efe91c73ec297ea6acc6c91a789334debc56ba6fd1\": rpc error: code = NotFound desc = could not find container \"4168216fde749b55a21ef4efe91c73ec297ea6acc6c91a789334debc56ba6fd1\": container with ID starting with 4168216fde749b55a21ef4efe91c73ec297ea6acc6c91a789334debc56ba6fd1 not found: ID does not exist" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.174345 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.174404 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.174591 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.174765 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d401e073-d123-406b-8282-7e25300eea65-logs\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.175035 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.175171 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d401e073-d123-406b-8282-7e25300eea65-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.175233 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5dmg\" (UniqueName: \"kubernetes.io/projected/d401e073-d123-406b-8282-7e25300eea65-kube-api-access-v5dmg\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.277394 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d401e073-d123-406b-8282-7e25300eea65-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.277463 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5dmg\" (UniqueName: \"kubernetes.io/projected/d401e073-d123-406b-8282-7e25300eea65-kube-api-access-v5dmg\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.277545 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.277574 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.277627 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.277670 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d401e073-d123-406b-8282-7e25300eea65-logs\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.277943 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d401e073-d123-406b-8282-7e25300eea65-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.278132 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d401e073-d123-406b-8282-7e25300eea65-logs\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.278335 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.281201 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.281414 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.281914 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.285184 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.293959 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5dmg\" (UniqueName: \"kubernetes.io/projected/d401e073-d123-406b-8282-7e25300eea65-kube-api-access-v5dmg\") pod \"glance-default-internal-api-0\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.317369 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.327630 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.336226 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.337921 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.340670 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.340896 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.346207 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.377931 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.481845 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-scripts\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.481919 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.481994 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76104e9f-958f-4ca1-bc3a-1a15e635ff48-logs\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.482022 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.482096 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24f89\" (UniqueName: \"kubernetes.io/projected/76104e9f-958f-4ca1-bc3a-1a15e635ff48-kube-api-access-24f89\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.482184 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/76104e9f-958f-4ca1-bc3a-1a15e635ff48-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.482318 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-config-data\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.508820 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11a558de-f7fe-46a8-9bba-2a50e6a5b538" path="/var/lib/kubelet/pods/11a558de-f7fe-46a8-9bba-2a50e6a5b538/volumes" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.514493 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="571a4966-bf13-46d1-835f-2b48e49026d8" path="/var/lib/kubelet/pods/571a4966-bf13-46d1-835f-2b48e49026d8/volumes" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.586062 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-scripts\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.586123 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.586175 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76104e9f-958f-4ca1-bc3a-1a15e635ff48-logs\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.586196 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.586228 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24f89\" (UniqueName: \"kubernetes.io/projected/76104e9f-958f-4ca1-bc3a-1a15e635ff48-kube-api-access-24f89\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.586267 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/76104e9f-958f-4ca1-bc3a-1a15e635ff48-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.586292 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-config-data\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.587145 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76104e9f-958f-4ca1-bc3a-1a15e635ff48-logs\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.587160 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/76104e9f-958f-4ca1-bc3a-1a15e635ff48-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.590873 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-config-data\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.597633 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.598204 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.599206 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-scripts\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.605725 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24f89\" (UniqueName: \"kubernetes.io/projected/76104e9f-958f-4ca1-bc3a-1a15e635ff48-kube-api-access-24f89\") pod \"glance-default-external-api-0\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " pod="openstack/glance-default-external-api-0" Dec 01 16:13:30 crc kubenswrapper[4810]: I1201 16:13:30.656120 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 16:13:31 crc kubenswrapper[4810]: I1201 16:13:30.940211 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:13:31 crc kubenswrapper[4810]: W1201 16:13:30.943331 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd401e073_d123_406b_8282_7e25300eea65.slice/crio-109b8610ffd8f7d330500c7934091e3a004d95ac6cf6961dfa9742a57adec879 WatchSource:0}: Error finding container 109b8610ffd8f7d330500c7934091e3a004d95ac6cf6961dfa9742a57adec879: Status 404 returned error can't find the container with id 109b8610ffd8f7d330500c7934091e3a004d95ac6cf6961dfa9742a57adec879 Dec 01 16:13:31 crc kubenswrapper[4810]: I1201 16:13:30.997901 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d401e073-d123-406b-8282-7e25300eea65","Type":"ContainerStarted","Data":"109b8610ffd8f7d330500c7934091e3a004d95ac6cf6961dfa9742a57adec879"} Dec 01 16:13:31 crc kubenswrapper[4810]: I1201 16:13:31.042792 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:13:31 crc kubenswrapper[4810]: W1201 16:13:31.051623 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76104e9f_958f_4ca1_bc3a_1a15e635ff48.slice/crio-c9746fb3c485f308dc5ce3e783d7f8d5df06eab6e8e078866a0379dca331d8dd WatchSource:0}: Error finding container c9746fb3c485f308dc5ce3e783d7f8d5df06eab6e8e078866a0379dca331d8dd: Status 404 returned error can't find the container with id c9746fb3c485f308dc5ce3e783d7f8d5df06eab6e8e078866a0379dca331d8dd Dec 01 16:13:32 crc kubenswrapper[4810]: I1201 16:13:32.015270 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"76104e9f-958f-4ca1-bc3a-1a15e635ff48","Type":"ContainerStarted","Data":"c39c26527d998183b88e7e60f894db477ff11d24b80a4b7263e61d7d2bcec83f"} Dec 01 16:13:32 crc kubenswrapper[4810]: I1201 16:13:32.015645 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"76104e9f-958f-4ca1-bc3a-1a15e635ff48","Type":"ContainerStarted","Data":"c9746fb3c485f308dc5ce3e783d7f8d5df06eab6e8e078866a0379dca331d8dd"} Dec 01 16:13:32 crc kubenswrapper[4810]: I1201 16:13:32.017035 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d401e073-d123-406b-8282-7e25300eea65","Type":"ContainerStarted","Data":"1d7fedbac8c47a4792657340b81ad6aa5cbff1c15d61f7b1aa7699324bc89fce"} Dec 01 16:13:32 crc kubenswrapper[4810]: I1201 16:13:32.971821 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:13:32 crc kubenswrapper[4810]: I1201 16:13:32.972156 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:13:33 crc kubenswrapper[4810]: I1201 16:13:33.027233 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d401e073-d123-406b-8282-7e25300eea65","Type":"ContainerStarted","Data":"7cb96c27e96c8a68f6eeeb30ff69f968200aead5bc35e93c92675395d16ee569"} Dec 01 16:13:33 crc kubenswrapper[4810]: I1201 16:13:33.029328 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"76104e9f-958f-4ca1-bc3a-1a15e635ff48","Type":"ContainerStarted","Data":"9df7f870b2d4de005275809ca768bdb334b257b2b522ea3cda6a960246b34076"} Dec 01 16:13:33 crc kubenswrapper[4810]: I1201 16:13:33.046919 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.046896685 podStartE2EDuration="3.046896685s" podCreationTimestamp="2025-12-01 16:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:13:33.044407768 +0000 UTC m=+5978.807917391" watchObservedRunningTime="2025-12-01 16:13:33.046896685 +0000 UTC m=+5978.810406288" Dec 01 16:13:33 crc kubenswrapper[4810]: I1201 16:13:33.069202 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.06918093 podStartE2EDuration="3.06918093s" podCreationTimestamp="2025-12-01 16:13:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:13:33.063791064 +0000 UTC m=+5978.827300677" watchObservedRunningTime="2025-12-01 16:13:33.06918093 +0000 UTC m=+5978.832690533" Dec 01 16:13:36 crc kubenswrapper[4810]: I1201 16:13:36.586753 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:13:36 crc kubenswrapper[4810]: I1201 16:13:36.658533 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-655d5b56f-qjvr4"] Dec 01 16:13:36 crc kubenswrapper[4810]: I1201 16:13:36.658868 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" podUID="0d5dc495-73c7-4eaa-b9be-61f8360ae8dc" containerName="dnsmasq-dns" containerID="cri-o://4bfbdb437e9e428fbacca57469e1b01885410191d71e644ae4e97e13a8b2ceb0" gracePeriod=10 Dec 01 16:13:37 crc kubenswrapper[4810]: I1201 16:13:37.064898 4810 generic.go:334] "Generic (PLEG): container finished" podID="0d5dc495-73c7-4eaa-b9be-61f8360ae8dc" containerID="4bfbdb437e9e428fbacca57469e1b01885410191d71e644ae4e97e13a8b2ceb0" exitCode=0 Dec 01 16:13:37 crc kubenswrapper[4810]: I1201 16:13:37.065001 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" event={"ID":"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc","Type":"ContainerDied","Data":"4bfbdb437e9e428fbacca57469e1b01885410191d71e644ae4e97e13a8b2ceb0"} Dec 01 16:13:37 crc kubenswrapper[4810]: I1201 16:13:37.197907 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:13:37 crc kubenswrapper[4810]: I1201 16:13:37.300336 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-dns-svc\") pod \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " Dec 01 16:13:37 crc kubenswrapper[4810]: I1201 16:13:37.300384 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6xcw\" (UniqueName: \"kubernetes.io/projected/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-kube-api-access-r6xcw\") pod \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " Dec 01 16:13:37 crc kubenswrapper[4810]: I1201 16:13:37.300439 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-ovsdbserver-nb\") pod \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " Dec 01 16:13:37 crc kubenswrapper[4810]: I1201 16:13:37.300456 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-ovsdbserver-sb\") pod \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " Dec 01 16:13:37 crc kubenswrapper[4810]: I1201 16:13:37.300711 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-config\") pod \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\" (UID: \"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc\") " Dec 01 16:13:37 crc kubenswrapper[4810]: I1201 16:13:37.305820 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-kube-api-access-r6xcw" (OuterVolumeSpecName: "kube-api-access-r6xcw") pod "0d5dc495-73c7-4eaa-b9be-61f8360ae8dc" (UID: "0d5dc495-73c7-4eaa-b9be-61f8360ae8dc"). InnerVolumeSpecName "kube-api-access-r6xcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:13:37 crc kubenswrapper[4810]: I1201 16:13:37.345313 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0d5dc495-73c7-4eaa-b9be-61f8360ae8dc" (UID: "0d5dc495-73c7-4eaa-b9be-61f8360ae8dc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:13:37 crc kubenswrapper[4810]: I1201 16:13:37.350882 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-config" (OuterVolumeSpecName: "config") pod "0d5dc495-73c7-4eaa-b9be-61f8360ae8dc" (UID: "0d5dc495-73c7-4eaa-b9be-61f8360ae8dc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:13:37 crc kubenswrapper[4810]: I1201 16:13:37.352442 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0d5dc495-73c7-4eaa-b9be-61f8360ae8dc" (UID: "0d5dc495-73c7-4eaa-b9be-61f8360ae8dc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:13:37 crc kubenswrapper[4810]: I1201 16:13:37.361693 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0d5dc495-73c7-4eaa-b9be-61f8360ae8dc" (UID: "0d5dc495-73c7-4eaa-b9be-61f8360ae8dc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:13:37 crc kubenswrapper[4810]: I1201 16:13:37.403249 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:37 crc kubenswrapper[4810]: I1201 16:13:37.403283 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:37 crc kubenswrapper[4810]: I1201 16:13:37.403293 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6xcw\" (UniqueName: \"kubernetes.io/projected/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-kube-api-access-r6xcw\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:37 crc kubenswrapper[4810]: I1201 16:13:37.403304 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:37 crc kubenswrapper[4810]: I1201 16:13:37.403315 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:38 crc kubenswrapper[4810]: I1201 16:13:38.073443 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" event={"ID":"0d5dc495-73c7-4eaa-b9be-61f8360ae8dc","Type":"ContainerDied","Data":"18a9dfb7b3ef4c2d128cc7ecad4dcd48c04f858f7d19b1b08ddf92f85a348aed"} Dec 01 16:13:38 crc kubenswrapper[4810]: I1201 16:13:38.073523 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-655d5b56f-qjvr4" Dec 01 16:13:38 crc kubenswrapper[4810]: I1201 16:13:38.073704 4810 scope.go:117] "RemoveContainer" containerID="4bfbdb437e9e428fbacca57469e1b01885410191d71e644ae4e97e13a8b2ceb0" Dec 01 16:13:38 crc kubenswrapper[4810]: I1201 16:13:38.097232 4810 scope.go:117] "RemoveContainer" containerID="f21ef54d9cfd076fae9a3da2004aa9e803aedc1288e566da226ece0ed47eae3a" Dec 01 16:13:38 crc kubenswrapper[4810]: I1201 16:13:38.101759 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-655d5b56f-qjvr4"] Dec 01 16:13:38 crc kubenswrapper[4810]: I1201 16:13:38.109528 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-655d5b56f-qjvr4"] Dec 01 16:13:38 crc kubenswrapper[4810]: I1201 16:13:38.509928 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d5dc495-73c7-4eaa-b9be-61f8360ae8dc" path="/var/lib/kubelet/pods/0d5dc495-73c7-4eaa-b9be-61f8360ae8dc/volumes" Dec 01 16:13:40 crc kubenswrapper[4810]: I1201 16:13:40.378682 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 16:13:40 crc kubenswrapper[4810]: I1201 16:13:40.379020 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 16:13:40 crc kubenswrapper[4810]: I1201 16:13:40.408944 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 16:13:40 crc kubenswrapper[4810]: I1201 16:13:40.427125 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 16:13:40 crc kubenswrapper[4810]: I1201 16:13:40.656375 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 16:13:40 crc kubenswrapper[4810]: I1201 16:13:40.656427 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 16:13:40 crc kubenswrapper[4810]: I1201 16:13:40.688228 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 16:13:40 crc kubenswrapper[4810]: I1201 16:13:40.699608 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 16:13:41 crc kubenswrapper[4810]: I1201 16:13:41.103531 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 16:13:41 crc kubenswrapper[4810]: I1201 16:13:41.103619 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 16:13:41 crc kubenswrapper[4810]: I1201 16:13:41.103663 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 16:13:41 crc kubenswrapper[4810]: I1201 16:13:41.103676 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 16:13:42 crc kubenswrapper[4810]: I1201 16:13:42.992028 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 16:13:43 crc kubenswrapper[4810]: I1201 16:13:43.007737 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 16:13:43 crc kubenswrapper[4810]: I1201 16:13:43.008258 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 16:13:43 crc kubenswrapper[4810]: I1201 16:13:43.013524 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.271661 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-2zlzh"] Dec 01 16:13:49 crc kubenswrapper[4810]: E1201 16:13:49.272558 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d5dc495-73c7-4eaa-b9be-61f8360ae8dc" containerName="dnsmasq-dns" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.272571 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d5dc495-73c7-4eaa-b9be-61f8360ae8dc" containerName="dnsmasq-dns" Dec 01 16:13:49 crc kubenswrapper[4810]: E1201 16:13:49.272599 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d5dc495-73c7-4eaa-b9be-61f8360ae8dc" containerName="init" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.272606 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d5dc495-73c7-4eaa-b9be-61f8360ae8dc" containerName="init" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.272770 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d5dc495-73c7-4eaa-b9be-61f8360ae8dc" containerName="dnsmasq-dns" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.273310 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2zlzh" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.283552 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-2zlzh"] Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.378875 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-2ca7-account-create-update-fkfmv"] Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.380002 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2ca7-account-create-update-fkfmv" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.383160 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.394908 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-2ca7-account-create-update-fkfmv"] Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.413942 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8hn5\" (UniqueName: \"kubernetes.io/projected/1c1aa2aa-21be-4db1-9b97-552b303cfa57-kube-api-access-p8hn5\") pod \"placement-db-create-2zlzh\" (UID: \"1c1aa2aa-21be-4db1-9b97-552b303cfa57\") " pod="openstack/placement-db-create-2zlzh" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.414119 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c1aa2aa-21be-4db1-9b97-552b303cfa57-operator-scripts\") pod \"placement-db-create-2zlzh\" (UID: \"1c1aa2aa-21be-4db1-9b97-552b303cfa57\") " pod="openstack/placement-db-create-2zlzh" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.516057 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txtzm\" (UniqueName: \"kubernetes.io/projected/b0f1746a-f775-4c67-9122-903cd261dc87-kube-api-access-txtzm\") pod \"placement-2ca7-account-create-update-fkfmv\" (UID: \"b0f1746a-f775-4c67-9122-903cd261dc87\") " pod="openstack/placement-2ca7-account-create-update-fkfmv" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.516137 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8hn5\" (UniqueName: \"kubernetes.io/projected/1c1aa2aa-21be-4db1-9b97-552b303cfa57-kube-api-access-p8hn5\") pod \"placement-db-create-2zlzh\" (UID: \"1c1aa2aa-21be-4db1-9b97-552b303cfa57\") " pod="openstack/placement-db-create-2zlzh" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.516174 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0f1746a-f775-4c67-9122-903cd261dc87-operator-scripts\") pod \"placement-2ca7-account-create-update-fkfmv\" (UID: \"b0f1746a-f775-4c67-9122-903cd261dc87\") " pod="openstack/placement-2ca7-account-create-update-fkfmv" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.516351 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c1aa2aa-21be-4db1-9b97-552b303cfa57-operator-scripts\") pod \"placement-db-create-2zlzh\" (UID: \"1c1aa2aa-21be-4db1-9b97-552b303cfa57\") " pod="openstack/placement-db-create-2zlzh" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.517459 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c1aa2aa-21be-4db1-9b97-552b303cfa57-operator-scripts\") pod \"placement-db-create-2zlzh\" (UID: \"1c1aa2aa-21be-4db1-9b97-552b303cfa57\") " pod="openstack/placement-db-create-2zlzh" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.534342 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8hn5\" (UniqueName: \"kubernetes.io/projected/1c1aa2aa-21be-4db1-9b97-552b303cfa57-kube-api-access-p8hn5\") pod \"placement-db-create-2zlzh\" (UID: \"1c1aa2aa-21be-4db1-9b97-552b303cfa57\") " pod="openstack/placement-db-create-2zlzh" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.596422 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2zlzh" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.618148 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0f1746a-f775-4c67-9122-903cd261dc87-operator-scripts\") pod \"placement-2ca7-account-create-update-fkfmv\" (UID: \"b0f1746a-f775-4c67-9122-903cd261dc87\") " pod="openstack/placement-2ca7-account-create-update-fkfmv" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.619101 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0f1746a-f775-4c67-9122-903cd261dc87-operator-scripts\") pod \"placement-2ca7-account-create-update-fkfmv\" (UID: \"b0f1746a-f775-4c67-9122-903cd261dc87\") " pod="openstack/placement-2ca7-account-create-update-fkfmv" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.619666 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txtzm\" (UniqueName: \"kubernetes.io/projected/b0f1746a-f775-4c67-9122-903cd261dc87-kube-api-access-txtzm\") pod \"placement-2ca7-account-create-update-fkfmv\" (UID: \"b0f1746a-f775-4c67-9122-903cd261dc87\") " pod="openstack/placement-2ca7-account-create-update-fkfmv" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.642357 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txtzm\" (UniqueName: \"kubernetes.io/projected/b0f1746a-f775-4c67-9122-903cd261dc87-kube-api-access-txtzm\") pod \"placement-2ca7-account-create-update-fkfmv\" (UID: \"b0f1746a-f775-4c67-9122-903cd261dc87\") " pod="openstack/placement-2ca7-account-create-update-fkfmv" Dec 01 16:13:49 crc kubenswrapper[4810]: I1201 16:13:49.701448 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2ca7-account-create-update-fkfmv" Dec 01 16:13:50 crc kubenswrapper[4810]: I1201 16:13:50.046388 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-2zlzh"] Dec 01 16:13:50 crc kubenswrapper[4810]: W1201 16:13:50.049636 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c1aa2aa_21be_4db1_9b97_552b303cfa57.slice/crio-971d3b11377178ae9554866fd8cb657df67d4ec567cac3c3a7df41599d5057a0 WatchSource:0}: Error finding container 971d3b11377178ae9554866fd8cb657df67d4ec567cac3c3a7df41599d5057a0: Status 404 returned error can't find the container with id 971d3b11377178ae9554866fd8cb657df67d4ec567cac3c3a7df41599d5057a0 Dec 01 16:13:50 crc kubenswrapper[4810]: I1201 16:13:50.178129 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2zlzh" event={"ID":"1c1aa2aa-21be-4db1-9b97-552b303cfa57","Type":"ContainerStarted","Data":"971d3b11377178ae9554866fd8cb657df67d4ec567cac3c3a7df41599d5057a0"} Dec 01 16:13:50 crc kubenswrapper[4810]: I1201 16:13:50.184646 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-2ca7-account-create-update-fkfmv"] Dec 01 16:13:50 crc kubenswrapper[4810]: W1201 16:13:50.194793 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0f1746a_f775_4c67_9122_903cd261dc87.slice/crio-d15300caf44fbef6eacea0d380947df70c25827103ff0c48f4a7898dd7c374b2 WatchSource:0}: Error finding container d15300caf44fbef6eacea0d380947df70c25827103ff0c48f4a7898dd7c374b2: Status 404 returned error can't find the container with id d15300caf44fbef6eacea0d380947df70c25827103ff0c48f4a7898dd7c374b2 Dec 01 16:13:51 crc kubenswrapper[4810]: I1201 16:13:51.187232 4810 generic.go:334] "Generic (PLEG): container finished" podID="1c1aa2aa-21be-4db1-9b97-552b303cfa57" containerID="2ffb281642b4705e534267c46825b7a1113616d4e1a85b7062d9692bd49640b1" exitCode=0 Dec 01 16:13:51 crc kubenswrapper[4810]: I1201 16:13:51.187286 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2zlzh" event={"ID":"1c1aa2aa-21be-4db1-9b97-552b303cfa57","Type":"ContainerDied","Data":"2ffb281642b4705e534267c46825b7a1113616d4e1a85b7062d9692bd49640b1"} Dec 01 16:13:51 crc kubenswrapper[4810]: I1201 16:13:51.190127 4810 generic.go:334] "Generic (PLEG): container finished" podID="b0f1746a-f775-4c67-9122-903cd261dc87" containerID="b17090bed39c4e268f652595deba6c7b1d0907729166a71e93979d05ea4332ec" exitCode=0 Dec 01 16:13:51 crc kubenswrapper[4810]: I1201 16:13:51.190152 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2ca7-account-create-update-fkfmv" event={"ID":"b0f1746a-f775-4c67-9122-903cd261dc87","Type":"ContainerDied","Data":"b17090bed39c4e268f652595deba6c7b1d0907729166a71e93979d05ea4332ec"} Dec 01 16:13:51 crc kubenswrapper[4810]: I1201 16:13:51.190167 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2ca7-account-create-update-fkfmv" event={"ID":"b0f1746a-f775-4c67-9122-903cd261dc87","Type":"ContainerStarted","Data":"d15300caf44fbef6eacea0d380947df70c25827103ff0c48f4a7898dd7c374b2"} Dec 01 16:13:52 crc kubenswrapper[4810]: I1201 16:13:52.610854 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2ca7-account-create-update-fkfmv" Dec 01 16:13:52 crc kubenswrapper[4810]: I1201 16:13:52.616213 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2zlzh" Dec 01 16:13:52 crc kubenswrapper[4810]: I1201 16:13:52.776770 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c1aa2aa-21be-4db1-9b97-552b303cfa57-operator-scripts\") pod \"1c1aa2aa-21be-4db1-9b97-552b303cfa57\" (UID: \"1c1aa2aa-21be-4db1-9b97-552b303cfa57\") " Dec 01 16:13:52 crc kubenswrapper[4810]: I1201 16:13:52.776903 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txtzm\" (UniqueName: \"kubernetes.io/projected/b0f1746a-f775-4c67-9122-903cd261dc87-kube-api-access-txtzm\") pod \"b0f1746a-f775-4c67-9122-903cd261dc87\" (UID: \"b0f1746a-f775-4c67-9122-903cd261dc87\") " Dec 01 16:13:52 crc kubenswrapper[4810]: I1201 16:13:52.776961 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0f1746a-f775-4c67-9122-903cd261dc87-operator-scripts\") pod \"b0f1746a-f775-4c67-9122-903cd261dc87\" (UID: \"b0f1746a-f775-4c67-9122-903cd261dc87\") " Dec 01 16:13:52 crc kubenswrapper[4810]: I1201 16:13:52.777110 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8hn5\" (UniqueName: \"kubernetes.io/projected/1c1aa2aa-21be-4db1-9b97-552b303cfa57-kube-api-access-p8hn5\") pod \"1c1aa2aa-21be-4db1-9b97-552b303cfa57\" (UID: \"1c1aa2aa-21be-4db1-9b97-552b303cfa57\") " Dec 01 16:13:52 crc kubenswrapper[4810]: I1201 16:13:52.777426 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c1aa2aa-21be-4db1-9b97-552b303cfa57-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1c1aa2aa-21be-4db1-9b97-552b303cfa57" (UID: "1c1aa2aa-21be-4db1-9b97-552b303cfa57"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:13:52 crc kubenswrapper[4810]: I1201 16:13:52.777689 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0f1746a-f775-4c67-9122-903cd261dc87-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b0f1746a-f775-4c67-9122-903cd261dc87" (UID: "b0f1746a-f775-4c67-9122-903cd261dc87"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:13:52 crc kubenswrapper[4810]: I1201 16:13:52.778304 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c1aa2aa-21be-4db1-9b97-552b303cfa57-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:52 crc kubenswrapper[4810]: I1201 16:13:52.778366 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0f1746a-f775-4c67-9122-903cd261dc87-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:52 crc kubenswrapper[4810]: I1201 16:13:52.783659 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c1aa2aa-21be-4db1-9b97-552b303cfa57-kube-api-access-p8hn5" (OuterVolumeSpecName: "kube-api-access-p8hn5") pod "1c1aa2aa-21be-4db1-9b97-552b303cfa57" (UID: "1c1aa2aa-21be-4db1-9b97-552b303cfa57"). InnerVolumeSpecName "kube-api-access-p8hn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:13:52 crc kubenswrapper[4810]: I1201 16:13:52.783822 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0f1746a-f775-4c67-9122-903cd261dc87-kube-api-access-txtzm" (OuterVolumeSpecName: "kube-api-access-txtzm") pod "b0f1746a-f775-4c67-9122-903cd261dc87" (UID: "b0f1746a-f775-4c67-9122-903cd261dc87"). InnerVolumeSpecName "kube-api-access-txtzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:13:52 crc kubenswrapper[4810]: I1201 16:13:52.880299 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txtzm\" (UniqueName: \"kubernetes.io/projected/b0f1746a-f775-4c67-9122-903cd261dc87-kube-api-access-txtzm\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:52 crc kubenswrapper[4810]: I1201 16:13:52.880507 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8hn5\" (UniqueName: \"kubernetes.io/projected/1c1aa2aa-21be-4db1-9b97-552b303cfa57-kube-api-access-p8hn5\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:53 crc kubenswrapper[4810]: I1201 16:13:53.209310 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-2zlzh" Dec 01 16:13:53 crc kubenswrapper[4810]: I1201 16:13:53.209337 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-2zlzh" event={"ID":"1c1aa2aa-21be-4db1-9b97-552b303cfa57","Type":"ContainerDied","Data":"971d3b11377178ae9554866fd8cb657df67d4ec567cac3c3a7df41599d5057a0"} Dec 01 16:13:53 crc kubenswrapper[4810]: I1201 16:13:53.209775 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="971d3b11377178ae9554866fd8cb657df67d4ec567cac3c3a7df41599d5057a0" Dec 01 16:13:53 crc kubenswrapper[4810]: I1201 16:13:53.210923 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2ca7-account-create-update-fkfmv" event={"ID":"b0f1746a-f775-4c67-9122-903cd261dc87","Type":"ContainerDied","Data":"d15300caf44fbef6eacea0d380947df70c25827103ff0c48f4a7898dd7c374b2"} Dec 01 16:13:53 crc kubenswrapper[4810]: I1201 16:13:53.210952 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d15300caf44fbef6eacea0d380947df70c25827103ff0c48f4a7898dd7c374b2" Dec 01 16:13:53 crc kubenswrapper[4810]: I1201 16:13:53.211088 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2ca7-account-create-update-fkfmv" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.636676 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86b95c6b9-2jj57"] Dec 01 16:13:54 crc kubenswrapper[4810]: E1201 16:13:54.637030 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c1aa2aa-21be-4db1-9b97-552b303cfa57" containerName="mariadb-database-create" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.637043 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c1aa2aa-21be-4db1-9b97-552b303cfa57" containerName="mariadb-database-create" Dec 01 16:13:54 crc kubenswrapper[4810]: E1201 16:13:54.637073 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0f1746a-f775-4c67-9122-903cd261dc87" containerName="mariadb-account-create-update" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.637079 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0f1746a-f775-4c67-9122-903cd261dc87" containerName="mariadb-account-create-update" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.637251 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0f1746a-f775-4c67-9122-903cd261dc87" containerName="mariadb-account-create-update" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.637262 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c1aa2aa-21be-4db1-9b97-552b303cfa57" containerName="mariadb-database-create" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.638167 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.660251 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86b95c6b9-2jj57"] Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.677527 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-dwrp5"] Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.679733 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.682287 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.690969 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.691112 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-qmdxl" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.714103 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dwrp5"] Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.715697 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-ovsdbserver-sb\") pod \"dnsmasq-dns-86b95c6b9-2jj57\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.715743 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-config\") pod \"dnsmasq-dns-86b95c6b9-2jj57\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.715787 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-dns-svc\") pod \"dnsmasq-dns-86b95c6b9-2jj57\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.715808 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-ovsdbserver-nb\") pod \"dnsmasq-dns-86b95c6b9-2jj57\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.715865 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhmr9\" (UniqueName: \"kubernetes.io/projected/bbd15d1f-1b0d-452d-8110-537140ccd88c-kube-api-access-lhmr9\") pod \"dnsmasq-dns-86b95c6b9-2jj57\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.817540 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhmr9\" (UniqueName: \"kubernetes.io/projected/bbd15d1f-1b0d-452d-8110-537140ccd88c-kube-api-access-lhmr9\") pod \"dnsmasq-dns-86b95c6b9-2jj57\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.817619 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84039e04-f053-496f-921a-18e12ff432e1-config-data\") pod \"placement-db-sync-dwrp5\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.817678 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84039e04-f053-496f-921a-18e12ff432e1-logs\") pod \"placement-db-sync-dwrp5\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.817696 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-ovsdbserver-sb\") pod \"dnsmasq-dns-86b95c6b9-2jj57\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.817715 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84039e04-f053-496f-921a-18e12ff432e1-combined-ca-bundle\") pod \"placement-db-sync-dwrp5\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.817741 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-config\") pod \"dnsmasq-dns-86b95c6b9-2jj57\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.817776 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzpcc\" (UniqueName: \"kubernetes.io/projected/84039e04-f053-496f-921a-18e12ff432e1-kube-api-access-qzpcc\") pod \"placement-db-sync-dwrp5\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.817810 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84039e04-f053-496f-921a-18e12ff432e1-scripts\") pod \"placement-db-sync-dwrp5\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.817839 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-dns-svc\") pod \"dnsmasq-dns-86b95c6b9-2jj57\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.817865 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-ovsdbserver-nb\") pod \"dnsmasq-dns-86b95c6b9-2jj57\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.818881 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-ovsdbserver-nb\") pod \"dnsmasq-dns-86b95c6b9-2jj57\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.819782 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-ovsdbserver-sb\") pod \"dnsmasq-dns-86b95c6b9-2jj57\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.820386 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-config\") pod \"dnsmasq-dns-86b95c6b9-2jj57\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.820913 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-dns-svc\") pod \"dnsmasq-dns-86b95c6b9-2jj57\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.836149 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhmr9\" (UniqueName: \"kubernetes.io/projected/bbd15d1f-1b0d-452d-8110-537140ccd88c-kube-api-access-lhmr9\") pod \"dnsmasq-dns-86b95c6b9-2jj57\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.919485 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84039e04-f053-496f-921a-18e12ff432e1-config-data\") pod \"placement-db-sync-dwrp5\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.919568 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84039e04-f053-496f-921a-18e12ff432e1-logs\") pod \"placement-db-sync-dwrp5\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.919592 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84039e04-f053-496f-921a-18e12ff432e1-combined-ca-bundle\") pod \"placement-db-sync-dwrp5\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.919617 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzpcc\" (UniqueName: \"kubernetes.io/projected/84039e04-f053-496f-921a-18e12ff432e1-kube-api-access-qzpcc\") pod \"placement-db-sync-dwrp5\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.919688 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84039e04-f053-496f-921a-18e12ff432e1-scripts\") pod \"placement-db-sync-dwrp5\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.920204 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84039e04-f053-496f-921a-18e12ff432e1-logs\") pod \"placement-db-sync-dwrp5\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.923237 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84039e04-f053-496f-921a-18e12ff432e1-scripts\") pod \"placement-db-sync-dwrp5\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.924225 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84039e04-f053-496f-921a-18e12ff432e1-combined-ca-bundle\") pod \"placement-db-sync-dwrp5\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.924391 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84039e04-f053-496f-921a-18e12ff432e1-config-data\") pod \"placement-db-sync-dwrp5\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.938643 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzpcc\" (UniqueName: \"kubernetes.io/projected/84039e04-f053-496f-921a-18e12ff432e1-kube-api-access-qzpcc\") pod \"placement-db-sync-dwrp5\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:54 crc kubenswrapper[4810]: I1201 16:13:54.967317 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:13:55 crc kubenswrapper[4810]: I1201 16:13:55.000258 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:55 crc kubenswrapper[4810]: I1201 16:13:55.465557 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86b95c6b9-2jj57"] Dec 01 16:13:55 crc kubenswrapper[4810]: I1201 16:13:55.540059 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dwrp5"] Dec 01 16:13:55 crc kubenswrapper[4810]: W1201 16:13:55.547581 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84039e04_f053_496f_921a_18e12ff432e1.slice/crio-449ca8d2f20cddf8ac79a7b9f6159a0924e91bdb481e3f90e3e382924f88814d WatchSource:0}: Error finding container 449ca8d2f20cddf8ac79a7b9f6159a0924e91bdb481e3f90e3e382924f88814d: Status 404 returned error can't find the container with id 449ca8d2f20cddf8ac79a7b9f6159a0924e91bdb481e3f90e3e382924f88814d Dec 01 16:13:56 crc kubenswrapper[4810]: I1201 16:13:56.243924 4810 generic.go:334] "Generic (PLEG): container finished" podID="bbd15d1f-1b0d-452d-8110-537140ccd88c" containerID="8d2cd0ef7276f18b52476ac042cd52cb32082380af17e26b8fcc01b21a3008d5" exitCode=0 Dec 01 16:13:56 crc kubenswrapper[4810]: I1201 16:13:56.244030 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" event={"ID":"bbd15d1f-1b0d-452d-8110-537140ccd88c","Type":"ContainerDied","Data":"8d2cd0ef7276f18b52476ac042cd52cb32082380af17e26b8fcc01b21a3008d5"} Dec 01 16:13:56 crc kubenswrapper[4810]: I1201 16:13:56.244568 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" event={"ID":"bbd15d1f-1b0d-452d-8110-537140ccd88c","Type":"ContainerStarted","Data":"71ab8379f9b94b4753bcc7ea33b16eb9aed2c61d429a0753954cb8b4d99ec4c4"} Dec 01 16:13:56 crc kubenswrapper[4810]: I1201 16:13:56.246430 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dwrp5" event={"ID":"84039e04-f053-496f-921a-18e12ff432e1","Type":"ContainerStarted","Data":"bc86f9e27fdb7f8ebcdb878b0e8d795189133c8c8c3c46b170bd0b3aae36e026"} Dec 01 16:13:56 crc kubenswrapper[4810]: I1201 16:13:56.246495 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dwrp5" event={"ID":"84039e04-f053-496f-921a-18e12ff432e1","Type":"ContainerStarted","Data":"449ca8d2f20cddf8ac79a7b9f6159a0924e91bdb481e3f90e3e382924f88814d"} Dec 01 16:13:56 crc kubenswrapper[4810]: I1201 16:13:56.304905 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-dwrp5" podStartSLOduration=2.304884528 podStartE2EDuration="2.304884528s" podCreationTimestamp="2025-12-01 16:13:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:13:56.290123797 +0000 UTC m=+6002.053633410" watchObservedRunningTime="2025-12-01 16:13:56.304884528 +0000 UTC m=+6002.068394151" Dec 01 16:13:57 crc kubenswrapper[4810]: I1201 16:13:57.255773 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" event={"ID":"bbd15d1f-1b0d-452d-8110-537140ccd88c","Type":"ContainerStarted","Data":"429b2ac6c979ceff96755077a8398310e6de3258b8bbbb97f9f5436478aa9d9c"} Dec 01 16:13:57 crc kubenswrapper[4810]: I1201 16:13:57.256229 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:13:57 crc kubenswrapper[4810]: I1201 16:13:57.258351 4810 generic.go:334] "Generic (PLEG): container finished" podID="84039e04-f053-496f-921a-18e12ff432e1" containerID="bc86f9e27fdb7f8ebcdb878b0e8d795189133c8c8c3c46b170bd0b3aae36e026" exitCode=0 Dec 01 16:13:57 crc kubenswrapper[4810]: I1201 16:13:57.258404 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dwrp5" event={"ID":"84039e04-f053-496f-921a-18e12ff432e1","Type":"ContainerDied","Data":"bc86f9e27fdb7f8ebcdb878b0e8d795189133c8c8c3c46b170bd0b3aae36e026"} Dec 01 16:13:57 crc kubenswrapper[4810]: I1201 16:13:57.283785 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" podStartSLOduration=3.283767352 podStartE2EDuration="3.283767352s" podCreationTimestamp="2025-12-01 16:13:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:13:57.277392439 +0000 UTC m=+6003.040902092" watchObservedRunningTime="2025-12-01 16:13:57.283767352 +0000 UTC m=+6003.047276965" Dec 01 16:13:58 crc kubenswrapper[4810]: I1201 16:13:58.607977 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:58 crc kubenswrapper[4810]: I1201 16:13:58.684450 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzpcc\" (UniqueName: \"kubernetes.io/projected/84039e04-f053-496f-921a-18e12ff432e1-kube-api-access-qzpcc\") pod \"84039e04-f053-496f-921a-18e12ff432e1\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " Dec 01 16:13:58 crc kubenswrapper[4810]: I1201 16:13:58.684630 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84039e04-f053-496f-921a-18e12ff432e1-logs\") pod \"84039e04-f053-496f-921a-18e12ff432e1\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " Dec 01 16:13:58 crc kubenswrapper[4810]: I1201 16:13:58.684903 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84039e04-f053-496f-921a-18e12ff432e1-combined-ca-bundle\") pod \"84039e04-f053-496f-921a-18e12ff432e1\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " Dec 01 16:13:58 crc kubenswrapper[4810]: I1201 16:13:58.684961 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84039e04-f053-496f-921a-18e12ff432e1-scripts\") pod \"84039e04-f053-496f-921a-18e12ff432e1\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " Dec 01 16:13:58 crc kubenswrapper[4810]: I1201 16:13:58.685018 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84039e04-f053-496f-921a-18e12ff432e1-config-data\") pod \"84039e04-f053-496f-921a-18e12ff432e1\" (UID: \"84039e04-f053-496f-921a-18e12ff432e1\") " Dec 01 16:13:58 crc kubenswrapper[4810]: I1201 16:13:58.685042 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84039e04-f053-496f-921a-18e12ff432e1-logs" (OuterVolumeSpecName: "logs") pod "84039e04-f053-496f-921a-18e12ff432e1" (UID: "84039e04-f053-496f-921a-18e12ff432e1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:13:58 crc kubenswrapper[4810]: I1201 16:13:58.685940 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84039e04-f053-496f-921a-18e12ff432e1-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:58 crc kubenswrapper[4810]: I1201 16:13:58.690953 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84039e04-f053-496f-921a-18e12ff432e1-scripts" (OuterVolumeSpecName: "scripts") pod "84039e04-f053-496f-921a-18e12ff432e1" (UID: "84039e04-f053-496f-921a-18e12ff432e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:58 crc kubenswrapper[4810]: I1201 16:13:58.691121 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84039e04-f053-496f-921a-18e12ff432e1-kube-api-access-qzpcc" (OuterVolumeSpecName: "kube-api-access-qzpcc") pod "84039e04-f053-496f-921a-18e12ff432e1" (UID: "84039e04-f053-496f-921a-18e12ff432e1"). InnerVolumeSpecName "kube-api-access-qzpcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:13:58 crc kubenswrapper[4810]: I1201 16:13:58.713782 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84039e04-f053-496f-921a-18e12ff432e1-config-data" (OuterVolumeSpecName: "config-data") pod "84039e04-f053-496f-921a-18e12ff432e1" (UID: "84039e04-f053-496f-921a-18e12ff432e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:58 crc kubenswrapper[4810]: I1201 16:13:58.715820 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84039e04-f053-496f-921a-18e12ff432e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84039e04-f053-496f-921a-18e12ff432e1" (UID: "84039e04-f053-496f-921a-18e12ff432e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:13:58 crc kubenswrapper[4810]: I1201 16:13:58.787364 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84039e04-f053-496f-921a-18e12ff432e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:58 crc kubenswrapper[4810]: I1201 16:13:58.787424 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84039e04-f053-496f-921a-18e12ff432e1-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:58 crc kubenswrapper[4810]: I1201 16:13:58.787437 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84039e04-f053-496f-921a-18e12ff432e1-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:58 crc kubenswrapper[4810]: I1201 16:13:58.787451 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzpcc\" (UniqueName: \"kubernetes.io/projected/84039e04-f053-496f-921a-18e12ff432e1-kube-api-access-qzpcc\") on node \"crc\" DevicePath \"\"" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.274499 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dwrp5" event={"ID":"84039e04-f053-496f-921a-18e12ff432e1","Type":"ContainerDied","Data":"449ca8d2f20cddf8ac79a7b9f6159a0924e91bdb481e3f90e3e382924f88814d"} Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.274535 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="449ca8d2f20cddf8ac79a7b9f6159a0924e91bdb481e3f90e3e382924f88814d" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.274588 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dwrp5" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.395706 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-668cb558f4-t5z5m"] Dec 01 16:13:59 crc kubenswrapper[4810]: E1201 16:13:59.412331 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84039e04-f053-496f-921a-18e12ff432e1" containerName="placement-db-sync" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.412378 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="84039e04-f053-496f-921a-18e12ff432e1" containerName="placement-db-sync" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.412734 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="84039e04-f053-496f-921a-18e12ff432e1" containerName="placement-db-sync" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.421209 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-668cb558f4-t5z5m"] Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.421399 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.425177 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.425557 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.425607 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-qmdxl" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.427892 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.428053 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.499424 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-scripts\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.499858 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-config-data\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.500172 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-internal-tls-certs\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.500325 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-combined-ca-bundle\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.500365 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-public-tls-certs\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.500453 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sl5t\" (UniqueName: \"kubernetes.io/projected/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-kube-api-access-4sl5t\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.500693 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-logs\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.602888 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-internal-tls-certs\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.602975 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-combined-ca-bundle\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.603001 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-public-tls-certs\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.603043 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sl5t\" (UniqueName: \"kubernetes.io/projected/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-kube-api-access-4sl5t\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.603101 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-logs\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.603129 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-scripts\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.603173 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-config-data\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.603522 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-logs\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.611119 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-scripts\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.611256 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-public-tls-certs\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.611392 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-combined-ca-bundle\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.611699 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-internal-tls-certs\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.617760 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-config-data\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.619388 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sl5t\" (UniqueName: \"kubernetes.io/projected/77f99599-4efe-41a7-8781-dcfd3fd4e8f4-kube-api-access-4sl5t\") pod \"placement-668cb558f4-t5z5m\" (UID: \"77f99599-4efe-41a7-8781-dcfd3fd4e8f4\") " pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:13:59 crc kubenswrapper[4810]: I1201 16:13:59.746000 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:14:00 crc kubenswrapper[4810]: I1201 16:14:00.202459 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-668cb558f4-t5z5m"] Dec 01 16:14:00 crc kubenswrapper[4810]: W1201 16:14:00.203264 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77f99599_4efe_41a7_8781_dcfd3fd4e8f4.slice/crio-d4318ed4952779bf1aa3e4e7a108a4759ba30cf4e4b0029cad32df28dfd10de8 WatchSource:0}: Error finding container d4318ed4952779bf1aa3e4e7a108a4759ba30cf4e4b0029cad32df28dfd10de8: Status 404 returned error can't find the container with id d4318ed4952779bf1aa3e4e7a108a4759ba30cf4e4b0029cad32df28dfd10de8 Dec 01 16:14:00 crc kubenswrapper[4810]: I1201 16:14:00.294635 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-668cb558f4-t5z5m" event={"ID":"77f99599-4efe-41a7-8781-dcfd3fd4e8f4","Type":"ContainerStarted","Data":"d4318ed4952779bf1aa3e4e7a108a4759ba30cf4e4b0029cad32df28dfd10de8"} Dec 01 16:14:01 crc kubenswrapper[4810]: I1201 16:14:01.308200 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-668cb558f4-t5z5m" event={"ID":"77f99599-4efe-41a7-8781-dcfd3fd4e8f4","Type":"ContainerStarted","Data":"845e1b1c1e13ec240811dafcdca8cda239d454c08ae63816e437ba2f2d772c7d"} Dec 01 16:14:01 crc kubenswrapper[4810]: I1201 16:14:01.309038 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:14:01 crc kubenswrapper[4810]: I1201 16:14:01.309054 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-668cb558f4-t5z5m" event={"ID":"77f99599-4efe-41a7-8781-dcfd3fd4e8f4","Type":"ContainerStarted","Data":"0c157288af7b0b507646bb21022d8afa613f4bc0191faecc932698f41f60c195"} Dec 01 16:14:01 crc kubenswrapper[4810]: I1201 16:14:01.327571 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-668cb558f4-t5z5m" podStartSLOduration=2.327549658 podStartE2EDuration="2.327549658s" podCreationTimestamp="2025-12-01 16:13:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:14:01.327102165 +0000 UTC m=+6007.090611778" watchObservedRunningTime="2025-12-01 16:14:01.327549658 +0000 UTC m=+6007.091059261" Dec 01 16:14:02 crc kubenswrapper[4810]: I1201 16:14:02.315742 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:14:02 crc kubenswrapper[4810]: I1201 16:14:02.972321 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:14:02 crc kubenswrapper[4810]: I1201 16:14:02.972398 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:14:02 crc kubenswrapper[4810]: I1201 16:14:02.972499 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 16:14:02 crc kubenswrapper[4810]: I1201 16:14:02.973379 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:14:02 crc kubenswrapper[4810]: I1201 16:14:02.973455 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" gracePeriod=600 Dec 01 16:14:03 crc kubenswrapper[4810]: E1201 16:14:03.102096 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:14:03 crc kubenswrapper[4810]: I1201 16:14:03.326118 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" exitCode=0 Dec 01 16:14:03 crc kubenswrapper[4810]: I1201 16:14:03.326196 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d"} Dec 01 16:14:03 crc kubenswrapper[4810]: I1201 16:14:03.326275 4810 scope.go:117] "RemoveContainer" containerID="a05157f175de8551ab2001ecba2c36ed1be28837c6a63ea1ad99ac3229a74f07" Dec 01 16:14:03 crc kubenswrapper[4810]: I1201 16:14:03.326929 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:14:03 crc kubenswrapper[4810]: E1201 16:14:03.327218 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:14:04 crc kubenswrapper[4810]: I1201 16:14:04.969638 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.049603 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9d778847c-h4rtf"] Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.049857 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9d778847c-h4rtf" podUID="ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c" containerName="dnsmasq-dns" containerID="cri-o://ff9363d79679c72ac1691d510f17ab6f5e9d55f30422534462c18f9889dd12f7" gracePeriod=10 Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.360189 4810 generic.go:334] "Generic (PLEG): container finished" podID="ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c" containerID="ff9363d79679c72ac1691d510f17ab6f5e9d55f30422534462c18f9889dd12f7" exitCode=0 Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.360228 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9d778847c-h4rtf" event={"ID":"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c","Type":"ContainerDied","Data":"ff9363d79679c72ac1691d510f17ab6f5e9d55f30422534462c18f9889dd12f7"} Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.533839 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.613487 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-dns-svc\") pod \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.613613 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-ovsdbserver-sb\") pod \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.613631 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-ovsdbserver-nb\") pod \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.613655 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cqzk\" (UniqueName: \"kubernetes.io/projected/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-kube-api-access-7cqzk\") pod \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.613777 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-config\") pod \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\" (UID: \"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c\") " Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.619671 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-kube-api-access-7cqzk" (OuterVolumeSpecName: "kube-api-access-7cqzk") pod "ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c" (UID: "ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c"). InnerVolumeSpecName "kube-api-access-7cqzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.663287 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c" (UID: "ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.665717 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c" (UID: "ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.665879 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-config" (OuterVolumeSpecName: "config") pod "ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c" (UID: "ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.672190 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c" (UID: "ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.715864 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.715900 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.715911 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.715921 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cqzk\" (UniqueName: \"kubernetes.io/projected/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-kube-api-access-7cqzk\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:05 crc kubenswrapper[4810]: I1201 16:14:05.715930 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:06 crc kubenswrapper[4810]: I1201 16:14:06.370932 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9d778847c-h4rtf" event={"ID":"ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c","Type":"ContainerDied","Data":"437bcdee2132857316e182d4df9ac56aa2dfffa66b661eebacbee09d7091b81d"} Dec 01 16:14:06 crc kubenswrapper[4810]: I1201 16:14:06.371088 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9d778847c-h4rtf" Dec 01 16:14:06 crc kubenswrapper[4810]: I1201 16:14:06.371279 4810 scope.go:117] "RemoveContainer" containerID="ff9363d79679c72ac1691d510f17ab6f5e9d55f30422534462c18f9889dd12f7" Dec 01 16:14:06 crc kubenswrapper[4810]: I1201 16:14:06.395296 4810 scope.go:117] "RemoveContainer" containerID="89e28ad83ce9149c604aea27f51aaffb5aa5ae0fea9e7d56a806354d291b4ab2" Dec 01 16:14:06 crc kubenswrapper[4810]: I1201 16:14:06.412320 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9d778847c-h4rtf"] Dec 01 16:14:06 crc kubenswrapper[4810]: I1201 16:14:06.421963 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9d778847c-h4rtf"] Dec 01 16:14:06 crc kubenswrapper[4810]: I1201 16:14:06.504722 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c" path="/var/lib/kubelet/pods/ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c/volumes" Dec 01 16:14:09 crc kubenswrapper[4810]: I1201 16:14:09.699391 4810 scope.go:117] "RemoveContainer" containerID="3c31fb968a8d869c9a460270298ca303afadc6ce67273d2992398035c3848aa8" Dec 01 16:14:09 crc kubenswrapper[4810]: I1201 16:14:09.719433 4810 scope.go:117] "RemoveContainer" containerID="e888a8bce60f9a4907d307433c7ad9b53378dd0ff8bb280328d9ef5bc7d41566" Dec 01 16:14:09 crc kubenswrapper[4810]: I1201 16:14:09.760585 4810 scope.go:117] "RemoveContainer" containerID="03b4a9f23ef2c19d78d55ba428b184c75e1f07608755280953ab0aa6589596a5" Dec 01 16:14:14 crc kubenswrapper[4810]: I1201 16:14:14.500955 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:14:14 crc kubenswrapper[4810]: E1201 16:14:14.502094 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:14:26 crc kubenswrapper[4810]: I1201 16:14:26.491267 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:14:26 crc kubenswrapper[4810]: E1201 16:14:26.492006 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:14:30 crc kubenswrapper[4810]: I1201 16:14:30.887604 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:14:31 crc kubenswrapper[4810]: I1201 16:14:31.897037 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-668cb558f4-t5z5m" Dec 01 16:14:41 crc kubenswrapper[4810]: I1201 16:14:41.490776 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:14:41 crc kubenswrapper[4810]: E1201 16:14:41.491640 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:14:51 crc kubenswrapper[4810]: I1201 16:14:51.945609 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-nfcz6"] Dec 01 16:14:51 crc kubenswrapper[4810]: E1201 16:14:51.948308 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c" containerName="dnsmasq-dns" Dec 01 16:14:51 crc kubenswrapper[4810]: I1201 16:14:51.948335 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c" containerName="dnsmasq-dns" Dec 01 16:14:51 crc kubenswrapper[4810]: E1201 16:14:51.948392 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c" containerName="init" Dec 01 16:14:51 crc kubenswrapper[4810]: I1201 16:14:51.948400 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c" containerName="init" Dec 01 16:14:51 crc kubenswrapper[4810]: I1201 16:14:51.949188 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba2acf0a-12c2-4ab0-9f5f-31f80a281b1c" containerName="dnsmasq-dns" Dec 01 16:14:51 crc kubenswrapper[4810]: I1201 16:14:51.950647 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nfcz6" Dec 01 16:14:51 crc kubenswrapper[4810]: I1201 16:14:51.965428 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-nfcz6"] Dec 01 16:14:51 crc kubenswrapper[4810]: I1201 16:14:51.993103 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j86fs\" (UniqueName: \"kubernetes.io/projected/25ad5508-0e9b-4c6e-86ad-472613af32be-kube-api-access-j86fs\") pod \"nova-api-db-create-nfcz6\" (UID: \"25ad5508-0e9b-4c6e-86ad-472613af32be\") " pod="openstack/nova-api-db-create-nfcz6" Dec 01 16:14:51 crc kubenswrapper[4810]: I1201 16:14:51.993269 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25ad5508-0e9b-4c6e-86ad-472613af32be-operator-scripts\") pod \"nova-api-db-create-nfcz6\" (UID: \"25ad5508-0e9b-4c6e-86ad-472613af32be\") " pod="openstack/nova-api-db-create-nfcz6" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.033275 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-942ws"] Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.034739 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-942ws" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.044608 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-a524-account-create-update-4cr2g"] Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.047630 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a524-account-create-update-4cr2g" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.054428 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-942ws"] Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.054618 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.066098 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-a524-account-create-update-4cr2g"] Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.100122 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f61fca3-87ad-4722-9d8d-f212af501835-operator-scripts\") pod \"nova-cell0-db-create-942ws\" (UID: \"0f61fca3-87ad-4722-9d8d-f212af501835\") " pod="openstack/nova-cell0-db-create-942ws" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.100378 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j86fs\" (UniqueName: \"kubernetes.io/projected/25ad5508-0e9b-4c6e-86ad-472613af32be-kube-api-access-j86fs\") pod \"nova-api-db-create-nfcz6\" (UID: \"25ad5508-0e9b-4c6e-86ad-472613af32be\") " pod="openstack/nova-api-db-create-nfcz6" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.100578 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxhln\" (UniqueName: \"kubernetes.io/projected/0f61fca3-87ad-4722-9d8d-f212af501835-kube-api-access-sxhln\") pod \"nova-cell0-db-create-942ws\" (UID: \"0f61fca3-87ad-4722-9d8d-f212af501835\") " pod="openstack/nova-cell0-db-create-942ws" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.100877 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25ad5508-0e9b-4c6e-86ad-472613af32be-operator-scripts\") pod \"nova-api-db-create-nfcz6\" (UID: \"25ad5508-0e9b-4c6e-86ad-472613af32be\") " pod="openstack/nova-api-db-create-nfcz6" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.100956 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55gsz\" (UniqueName: \"kubernetes.io/projected/ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4-kube-api-access-55gsz\") pod \"nova-api-a524-account-create-update-4cr2g\" (UID: \"ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4\") " pod="openstack/nova-api-a524-account-create-update-4cr2g" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.101000 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4-operator-scripts\") pod \"nova-api-a524-account-create-update-4cr2g\" (UID: \"ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4\") " pod="openstack/nova-api-a524-account-create-update-4cr2g" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.101979 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25ad5508-0e9b-4c6e-86ad-472613af32be-operator-scripts\") pod \"nova-api-db-create-nfcz6\" (UID: \"25ad5508-0e9b-4c6e-86ad-472613af32be\") " pod="openstack/nova-api-db-create-nfcz6" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.130256 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j86fs\" (UniqueName: \"kubernetes.io/projected/25ad5508-0e9b-4c6e-86ad-472613af32be-kube-api-access-j86fs\") pod \"nova-api-db-create-nfcz6\" (UID: \"25ad5508-0e9b-4c6e-86ad-472613af32be\") " pod="openstack/nova-api-db-create-nfcz6" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.203789 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f61fca3-87ad-4722-9d8d-f212af501835-operator-scripts\") pod \"nova-cell0-db-create-942ws\" (UID: \"0f61fca3-87ad-4722-9d8d-f212af501835\") " pod="openstack/nova-cell0-db-create-942ws" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.203947 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxhln\" (UniqueName: \"kubernetes.io/projected/0f61fca3-87ad-4722-9d8d-f212af501835-kube-api-access-sxhln\") pod \"nova-cell0-db-create-942ws\" (UID: \"0f61fca3-87ad-4722-9d8d-f212af501835\") " pod="openstack/nova-cell0-db-create-942ws" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.204071 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55gsz\" (UniqueName: \"kubernetes.io/projected/ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4-kube-api-access-55gsz\") pod \"nova-api-a524-account-create-update-4cr2g\" (UID: \"ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4\") " pod="openstack/nova-api-a524-account-create-update-4cr2g" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.204116 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4-operator-scripts\") pod \"nova-api-a524-account-create-update-4cr2g\" (UID: \"ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4\") " pod="openstack/nova-api-a524-account-create-update-4cr2g" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.204917 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f61fca3-87ad-4722-9d8d-f212af501835-operator-scripts\") pod \"nova-cell0-db-create-942ws\" (UID: \"0f61fca3-87ad-4722-9d8d-f212af501835\") " pod="openstack/nova-cell0-db-create-942ws" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.204977 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4-operator-scripts\") pod \"nova-api-a524-account-create-update-4cr2g\" (UID: \"ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4\") " pod="openstack/nova-api-a524-account-create-update-4cr2g" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.218417 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-kdnv8"] Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.219848 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kdnv8" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.224592 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxhln\" (UniqueName: \"kubernetes.io/projected/0f61fca3-87ad-4722-9d8d-f212af501835-kube-api-access-sxhln\") pod \"nova-cell0-db-create-942ws\" (UID: \"0f61fca3-87ad-4722-9d8d-f212af501835\") " pod="openstack/nova-cell0-db-create-942ws" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.230306 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-kdnv8"] Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.233775 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55gsz\" (UniqueName: \"kubernetes.io/projected/ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4-kube-api-access-55gsz\") pod \"nova-api-a524-account-create-update-4cr2g\" (UID: \"ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4\") " pod="openstack/nova-api-a524-account-create-update-4cr2g" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.241899 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-378c-account-create-update-rbsww"] Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.250176 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-378c-account-create-update-rbsww" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.254280 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.262770 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-378c-account-create-update-rbsww"] Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.284156 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nfcz6" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.307626 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmw8l\" (UniqueName: \"kubernetes.io/projected/b2d19437-5742-4516-8c26-52267458dcec-kube-api-access-gmw8l\") pod \"nova-cell0-378c-account-create-update-rbsww\" (UID: \"b2d19437-5742-4516-8c26-52267458dcec\") " pod="openstack/nova-cell0-378c-account-create-update-rbsww" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.307788 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6e67c08-6bc5-4c10-ac10-80bb918b95d5-operator-scripts\") pod \"nova-cell1-db-create-kdnv8\" (UID: \"e6e67c08-6bc5-4c10-ac10-80bb918b95d5\") " pod="openstack/nova-cell1-db-create-kdnv8" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.307871 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5vg7\" (UniqueName: \"kubernetes.io/projected/e6e67c08-6bc5-4c10-ac10-80bb918b95d5-kube-api-access-j5vg7\") pod \"nova-cell1-db-create-kdnv8\" (UID: \"e6e67c08-6bc5-4c10-ac10-80bb918b95d5\") " pod="openstack/nova-cell1-db-create-kdnv8" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.307898 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2d19437-5742-4516-8c26-52267458dcec-operator-scripts\") pod \"nova-cell0-378c-account-create-update-rbsww\" (UID: \"b2d19437-5742-4516-8c26-52267458dcec\") " pod="openstack/nova-cell0-378c-account-create-update-rbsww" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.366631 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-942ws" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.379954 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a524-account-create-update-4cr2g" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.409917 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmw8l\" (UniqueName: \"kubernetes.io/projected/b2d19437-5742-4516-8c26-52267458dcec-kube-api-access-gmw8l\") pod \"nova-cell0-378c-account-create-update-rbsww\" (UID: \"b2d19437-5742-4516-8c26-52267458dcec\") " pod="openstack/nova-cell0-378c-account-create-update-rbsww" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.410042 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6e67c08-6bc5-4c10-ac10-80bb918b95d5-operator-scripts\") pod \"nova-cell1-db-create-kdnv8\" (UID: \"e6e67c08-6bc5-4c10-ac10-80bb918b95d5\") " pod="openstack/nova-cell1-db-create-kdnv8" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.410127 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5vg7\" (UniqueName: \"kubernetes.io/projected/e6e67c08-6bc5-4c10-ac10-80bb918b95d5-kube-api-access-j5vg7\") pod \"nova-cell1-db-create-kdnv8\" (UID: \"e6e67c08-6bc5-4c10-ac10-80bb918b95d5\") " pod="openstack/nova-cell1-db-create-kdnv8" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.410145 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2d19437-5742-4516-8c26-52267458dcec-operator-scripts\") pod \"nova-cell0-378c-account-create-update-rbsww\" (UID: \"b2d19437-5742-4516-8c26-52267458dcec\") " pod="openstack/nova-cell0-378c-account-create-update-rbsww" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.411131 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6e67c08-6bc5-4c10-ac10-80bb918b95d5-operator-scripts\") pod \"nova-cell1-db-create-kdnv8\" (UID: \"e6e67c08-6bc5-4c10-ac10-80bb918b95d5\") " pod="openstack/nova-cell1-db-create-kdnv8" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.412282 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2d19437-5742-4516-8c26-52267458dcec-operator-scripts\") pod \"nova-cell0-378c-account-create-update-rbsww\" (UID: \"b2d19437-5742-4516-8c26-52267458dcec\") " pod="openstack/nova-cell0-378c-account-create-update-rbsww" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.435327 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmw8l\" (UniqueName: \"kubernetes.io/projected/b2d19437-5742-4516-8c26-52267458dcec-kube-api-access-gmw8l\") pod \"nova-cell0-378c-account-create-update-rbsww\" (UID: \"b2d19437-5742-4516-8c26-52267458dcec\") " pod="openstack/nova-cell0-378c-account-create-update-rbsww" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.438141 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5vg7\" (UniqueName: \"kubernetes.io/projected/e6e67c08-6bc5-4c10-ac10-80bb918b95d5-kube-api-access-j5vg7\") pod \"nova-cell1-db-create-kdnv8\" (UID: \"e6e67c08-6bc5-4c10-ac10-80bb918b95d5\") " pod="openstack/nova-cell1-db-create-kdnv8" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.449281 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-0fac-account-create-update-57jps"] Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.450776 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0fac-account-create-update-57jps" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.453558 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.470821 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-0fac-account-create-update-57jps"] Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.512413 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d789e4e-7e8f-48c6-94e2-cf1d081d0018-operator-scripts\") pod \"nova-cell1-0fac-account-create-update-57jps\" (UID: \"1d789e4e-7e8f-48c6-94e2-cf1d081d0018\") " pod="openstack/nova-cell1-0fac-account-create-update-57jps" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.512511 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q84ws\" (UniqueName: \"kubernetes.io/projected/1d789e4e-7e8f-48c6-94e2-cf1d081d0018-kube-api-access-q84ws\") pod \"nova-cell1-0fac-account-create-update-57jps\" (UID: \"1d789e4e-7e8f-48c6-94e2-cf1d081d0018\") " pod="openstack/nova-cell1-0fac-account-create-update-57jps" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.614559 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d789e4e-7e8f-48c6-94e2-cf1d081d0018-operator-scripts\") pod \"nova-cell1-0fac-account-create-update-57jps\" (UID: \"1d789e4e-7e8f-48c6-94e2-cf1d081d0018\") " pod="openstack/nova-cell1-0fac-account-create-update-57jps" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.614621 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q84ws\" (UniqueName: \"kubernetes.io/projected/1d789e4e-7e8f-48c6-94e2-cf1d081d0018-kube-api-access-q84ws\") pod \"nova-cell1-0fac-account-create-update-57jps\" (UID: \"1d789e4e-7e8f-48c6-94e2-cf1d081d0018\") " pod="openstack/nova-cell1-0fac-account-create-update-57jps" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.615564 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d789e4e-7e8f-48c6-94e2-cf1d081d0018-operator-scripts\") pod \"nova-cell1-0fac-account-create-update-57jps\" (UID: \"1d789e4e-7e8f-48c6-94e2-cf1d081d0018\") " pod="openstack/nova-cell1-0fac-account-create-update-57jps" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.631303 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q84ws\" (UniqueName: \"kubernetes.io/projected/1d789e4e-7e8f-48c6-94e2-cf1d081d0018-kube-api-access-q84ws\") pod \"nova-cell1-0fac-account-create-update-57jps\" (UID: \"1d789e4e-7e8f-48c6-94e2-cf1d081d0018\") " pod="openstack/nova-cell1-0fac-account-create-update-57jps" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.666583 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kdnv8" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.677695 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-378c-account-create-update-rbsww" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.754916 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-nfcz6"] Dec 01 16:14:52 crc kubenswrapper[4810]: W1201 16:14:52.768708 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25ad5508_0e9b_4c6e_86ad_472613af32be.slice/crio-37ddb8c10f53cc7d4b486f5fed3918af81caf924c841897f2236850c4310d3a8 WatchSource:0}: Error finding container 37ddb8c10f53cc7d4b486f5fed3918af81caf924c841897f2236850c4310d3a8: Status 404 returned error can't find the container with id 37ddb8c10f53cc7d4b486f5fed3918af81caf924c841897f2236850c4310d3a8 Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.784313 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0fac-account-create-update-57jps" Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.895913 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-a524-account-create-update-4cr2g"] Dec 01 16:14:52 crc kubenswrapper[4810]: W1201 16:14:52.903908 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae98057f_ede8_47b0_b0d8_fea7d5d2e6f4.slice/crio-4578b5c5269348af6782980ef19961a3e8ea685171f34f7867214bdd08e4d1ba WatchSource:0}: Error finding container 4578b5c5269348af6782980ef19961a3e8ea685171f34f7867214bdd08e4d1ba: Status 404 returned error can't find the container with id 4578b5c5269348af6782980ef19961a3e8ea685171f34f7867214bdd08e4d1ba Dec 01 16:14:52 crc kubenswrapper[4810]: I1201 16:14:52.919960 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-942ws"] Dec 01 16:14:52 crc kubenswrapper[4810]: W1201 16:14:52.927090 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f61fca3_87ad_4722_9d8d_f212af501835.slice/crio-a7514d247480ee649b2a98b0fd4e2fe8841dbfd8a28d8551e888319e1c780e0d WatchSource:0}: Error finding container a7514d247480ee649b2a98b0fd4e2fe8841dbfd8a28d8551e888319e1c780e0d: Status 404 returned error can't find the container with id a7514d247480ee649b2a98b0fd4e2fe8841dbfd8a28d8551e888319e1c780e0d Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.189541 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-kdnv8"] Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.196353 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-378c-account-create-update-rbsww"] Dec 01 16:14:53 crc kubenswrapper[4810]: W1201 16:14:53.249164 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2d19437_5742_4516_8c26_52267458dcec.slice/crio-01dafa4c56de45fc78a077fbead7d1c8357cf96ee9276beb7167f206c892302d WatchSource:0}: Error finding container 01dafa4c56de45fc78a077fbead7d1c8357cf96ee9276beb7167f206c892302d: Status 404 returned error can't find the container with id 01dafa4c56de45fc78a077fbead7d1c8357cf96ee9276beb7167f206c892302d Dec 01 16:14:53 crc kubenswrapper[4810]: W1201 16:14:53.249374 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6e67c08_6bc5_4c10_ac10_80bb918b95d5.slice/crio-82281cb109a5d1dc6b49d06b1dac667d58ce374862c687e225c112c859626fc1 WatchSource:0}: Error finding container 82281cb109a5d1dc6b49d06b1dac667d58ce374862c687e225c112c859626fc1: Status 404 returned error can't find the container with id 82281cb109a5d1dc6b49d06b1dac667d58ce374862c687e225c112c859626fc1 Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.328142 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-0fac-account-create-update-57jps"] Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.775605 4810 generic.go:334] "Generic (PLEG): container finished" podID="0f61fca3-87ad-4722-9d8d-f212af501835" containerID="93ab28f60c9265602b534910e42c28f3e5aad0b872db38a20f5be6c8fbf56c26" exitCode=0 Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.775756 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-942ws" event={"ID":"0f61fca3-87ad-4722-9d8d-f212af501835","Type":"ContainerDied","Data":"93ab28f60c9265602b534910e42c28f3e5aad0b872db38a20f5be6c8fbf56c26"} Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.775942 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-942ws" event={"ID":"0f61fca3-87ad-4722-9d8d-f212af501835","Type":"ContainerStarted","Data":"a7514d247480ee649b2a98b0fd4e2fe8841dbfd8a28d8551e888319e1c780e0d"} Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.778198 4810 generic.go:334] "Generic (PLEG): container finished" podID="ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4" containerID="2330e916d744addde52cc40fdaefdfbe1ae4811254a7cf013517f884ba2016b6" exitCode=0 Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.778233 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a524-account-create-update-4cr2g" event={"ID":"ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4","Type":"ContainerDied","Data":"2330e916d744addde52cc40fdaefdfbe1ae4811254a7cf013517f884ba2016b6"} Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.778277 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a524-account-create-update-4cr2g" event={"ID":"ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4","Type":"ContainerStarted","Data":"4578b5c5269348af6782980ef19961a3e8ea685171f34f7867214bdd08e4d1ba"} Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.780847 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0fac-account-create-update-57jps" event={"ID":"1d789e4e-7e8f-48c6-94e2-cf1d081d0018","Type":"ContainerStarted","Data":"51816684532c34a75051e9eaf7cc7f9a5a123b3b8686b2d8945d109e810b1eab"} Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.780882 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0fac-account-create-update-57jps" event={"ID":"1d789e4e-7e8f-48c6-94e2-cf1d081d0018","Type":"ContainerStarted","Data":"3f4d50c633901da574117258b78e87cff186902e41dd9f3b28425eaddcda901c"} Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.783208 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kdnv8" event={"ID":"e6e67c08-6bc5-4c10-ac10-80bb918b95d5","Type":"ContainerStarted","Data":"1e1050cbe92252ae560a925d4edb7788b2d75040a21bbffd2bd8ad3eb7fcd5e8"} Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.783248 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kdnv8" event={"ID":"e6e67c08-6bc5-4c10-ac10-80bb918b95d5","Type":"ContainerStarted","Data":"82281cb109a5d1dc6b49d06b1dac667d58ce374862c687e225c112c859626fc1"} Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.787177 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-378c-account-create-update-rbsww" event={"ID":"b2d19437-5742-4516-8c26-52267458dcec","Type":"ContainerStarted","Data":"685a66adc9ea7bc3944304fbd1630b0589690e6c8f287c51a0f91b25365ce8d4"} Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.787208 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-378c-account-create-update-rbsww" event={"ID":"b2d19437-5742-4516-8c26-52267458dcec","Type":"ContainerStarted","Data":"01dafa4c56de45fc78a077fbead7d1c8357cf96ee9276beb7167f206c892302d"} Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.789694 4810 generic.go:334] "Generic (PLEG): container finished" podID="25ad5508-0e9b-4c6e-86ad-472613af32be" containerID="e485960863af96767deee39cafeca708cce075ba3a5668045121fb9388bc3c7e" exitCode=0 Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.789785 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nfcz6" event={"ID":"25ad5508-0e9b-4c6e-86ad-472613af32be","Type":"ContainerDied","Data":"e485960863af96767deee39cafeca708cce075ba3a5668045121fb9388bc3c7e"} Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.789813 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nfcz6" event={"ID":"25ad5508-0e9b-4c6e-86ad-472613af32be","Type":"ContainerStarted","Data":"37ddb8c10f53cc7d4b486f5fed3918af81caf924c841897f2236850c4310d3a8"} Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.811349 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-378c-account-create-update-rbsww" podStartSLOduration=1.811323578 podStartE2EDuration="1.811323578s" podCreationTimestamp="2025-12-01 16:14:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:14:53.809406477 +0000 UTC m=+6059.572916080" watchObservedRunningTime="2025-12-01 16:14:53.811323578 +0000 UTC m=+6059.574833181" Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.869290 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-kdnv8" podStartSLOduration=1.8692646320000001 podStartE2EDuration="1.869264632s" podCreationTimestamp="2025-12-01 16:14:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:14:53.837294644 +0000 UTC m=+6059.600804247" watchObservedRunningTime="2025-12-01 16:14:53.869264632 +0000 UTC m=+6059.632774255" Dec 01 16:14:53 crc kubenswrapper[4810]: I1201 16:14:53.896330 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-0fac-account-create-update-57jps" podStartSLOduration=1.896305215 podStartE2EDuration="1.896305215s" podCreationTimestamp="2025-12-01 16:14:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:14:53.883099166 +0000 UTC m=+6059.646608769" watchObservedRunningTime="2025-12-01 16:14:53.896305215 +0000 UTC m=+6059.659814818" Dec 01 16:14:54 crc kubenswrapper[4810]: I1201 16:14:54.802397 4810 generic.go:334] "Generic (PLEG): container finished" podID="e6e67c08-6bc5-4c10-ac10-80bb918b95d5" containerID="1e1050cbe92252ae560a925d4edb7788b2d75040a21bbffd2bd8ad3eb7fcd5e8" exitCode=0 Dec 01 16:14:54 crc kubenswrapper[4810]: I1201 16:14:54.802486 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kdnv8" event={"ID":"e6e67c08-6bc5-4c10-ac10-80bb918b95d5","Type":"ContainerDied","Data":"1e1050cbe92252ae560a925d4edb7788b2d75040a21bbffd2bd8ad3eb7fcd5e8"} Dec 01 16:14:54 crc kubenswrapper[4810]: I1201 16:14:54.805088 4810 generic.go:334] "Generic (PLEG): container finished" podID="b2d19437-5742-4516-8c26-52267458dcec" containerID="685a66adc9ea7bc3944304fbd1630b0589690e6c8f287c51a0f91b25365ce8d4" exitCode=0 Dec 01 16:14:54 crc kubenswrapper[4810]: I1201 16:14:54.805203 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-378c-account-create-update-rbsww" event={"ID":"b2d19437-5742-4516-8c26-52267458dcec","Type":"ContainerDied","Data":"685a66adc9ea7bc3944304fbd1630b0589690e6c8f287c51a0f91b25365ce8d4"} Dec 01 16:14:54 crc kubenswrapper[4810]: I1201 16:14:54.807797 4810 generic.go:334] "Generic (PLEG): container finished" podID="1d789e4e-7e8f-48c6-94e2-cf1d081d0018" containerID="51816684532c34a75051e9eaf7cc7f9a5a123b3b8686b2d8945d109e810b1eab" exitCode=0 Dec 01 16:14:54 crc kubenswrapper[4810]: I1201 16:14:54.807886 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0fac-account-create-update-57jps" event={"ID":"1d789e4e-7e8f-48c6-94e2-cf1d081d0018","Type":"ContainerDied","Data":"51816684532c34a75051e9eaf7cc7f9a5a123b3b8686b2d8945d109e810b1eab"} Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.167995 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a524-account-create-update-4cr2g" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.261455 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nfcz6" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.266105 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-942ws" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.270123 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55gsz\" (UniqueName: \"kubernetes.io/projected/ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4-kube-api-access-55gsz\") pod \"ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4\" (UID: \"ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4\") " Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.270256 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4-operator-scripts\") pod \"ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4\" (UID: \"ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4\") " Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.271197 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4" (UID: "ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.284833 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4-kube-api-access-55gsz" (OuterVolumeSpecName: "kube-api-access-55gsz") pod "ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4" (UID: "ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4"). InnerVolumeSpecName "kube-api-access-55gsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.371161 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25ad5508-0e9b-4c6e-86ad-472613af32be-operator-scripts\") pod \"25ad5508-0e9b-4c6e-86ad-472613af32be\" (UID: \"25ad5508-0e9b-4c6e-86ad-472613af32be\") " Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.371214 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxhln\" (UniqueName: \"kubernetes.io/projected/0f61fca3-87ad-4722-9d8d-f212af501835-kube-api-access-sxhln\") pod \"0f61fca3-87ad-4722-9d8d-f212af501835\" (UID: \"0f61fca3-87ad-4722-9d8d-f212af501835\") " Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.371344 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j86fs\" (UniqueName: \"kubernetes.io/projected/25ad5508-0e9b-4c6e-86ad-472613af32be-kube-api-access-j86fs\") pod \"25ad5508-0e9b-4c6e-86ad-472613af32be\" (UID: \"25ad5508-0e9b-4c6e-86ad-472613af32be\") " Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.371365 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f61fca3-87ad-4722-9d8d-f212af501835-operator-scripts\") pod \"0f61fca3-87ad-4722-9d8d-f212af501835\" (UID: \"0f61fca3-87ad-4722-9d8d-f212af501835\") " Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.371709 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25ad5508-0e9b-4c6e-86ad-472613af32be-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "25ad5508-0e9b-4c6e-86ad-472613af32be" (UID: "25ad5508-0e9b-4c6e-86ad-472613af32be"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.371795 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55gsz\" (UniqueName: \"kubernetes.io/projected/ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4-kube-api-access-55gsz\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.371812 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.372144 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f61fca3-87ad-4722-9d8d-f212af501835-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0f61fca3-87ad-4722-9d8d-f212af501835" (UID: "0f61fca3-87ad-4722-9d8d-f212af501835"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.374688 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25ad5508-0e9b-4c6e-86ad-472613af32be-kube-api-access-j86fs" (OuterVolumeSpecName: "kube-api-access-j86fs") pod "25ad5508-0e9b-4c6e-86ad-472613af32be" (UID: "25ad5508-0e9b-4c6e-86ad-472613af32be"). InnerVolumeSpecName "kube-api-access-j86fs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.375718 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f61fca3-87ad-4722-9d8d-f212af501835-kube-api-access-sxhln" (OuterVolumeSpecName: "kube-api-access-sxhln") pod "0f61fca3-87ad-4722-9d8d-f212af501835" (UID: "0f61fca3-87ad-4722-9d8d-f212af501835"). InnerVolumeSpecName "kube-api-access-sxhln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.472725 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j86fs\" (UniqueName: \"kubernetes.io/projected/25ad5508-0e9b-4c6e-86ad-472613af32be-kube-api-access-j86fs\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.472761 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f61fca3-87ad-4722-9d8d-f212af501835-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.472773 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25ad5508-0e9b-4c6e-86ad-472613af32be-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.472782 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxhln\" (UniqueName: \"kubernetes.io/projected/0f61fca3-87ad-4722-9d8d-f212af501835-kube-api-access-sxhln\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.491158 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:14:55 crc kubenswrapper[4810]: E1201 16:14:55.491419 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.819795 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a524-account-create-update-4cr2g" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.819810 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a524-account-create-update-4cr2g" event={"ID":"ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4","Type":"ContainerDied","Data":"4578b5c5269348af6782980ef19961a3e8ea685171f34f7867214bdd08e4d1ba"} Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.820203 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4578b5c5269348af6782980ef19961a3e8ea685171f34f7867214bdd08e4d1ba" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.821412 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nfcz6" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.821435 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nfcz6" event={"ID":"25ad5508-0e9b-4c6e-86ad-472613af32be","Type":"ContainerDied","Data":"37ddb8c10f53cc7d4b486f5fed3918af81caf924c841897f2236850c4310d3a8"} Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.821497 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37ddb8c10f53cc7d4b486f5fed3918af81caf924c841897f2236850c4310d3a8" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.822891 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-942ws" event={"ID":"0f61fca3-87ad-4722-9d8d-f212af501835","Type":"ContainerDied","Data":"a7514d247480ee649b2a98b0fd4e2fe8841dbfd8a28d8551e888319e1c780e0d"} Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.822933 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-942ws" Dec 01 16:14:55 crc kubenswrapper[4810]: I1201 16:14:55.822934 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7514d247480ee649b2a98b0fd4e2fe8841dbfd8a28d8551e888319e1c780e0d" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.032421 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-378c-account-create-update-rbsww" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.084482 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmw8l\" (UniqueName: \"kubernetes.io/projected/b2d19437-5742-4516-8c26-52267458dcec-kube-api-access-gmw8l\") pod \"b2d19437-5742-4516-8c26-52267458dcec\" (UID: \"b2d19437-5742-4516-8c26-52267458dcec\") " Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.084684 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2d19437-5742-4516-8c26-52267458dcec-operator-scripts\") pod \"b2d19437-5742-4516-8c26-52267458dcec\" (UID: \"b2d19437-5742-4516-8c26-52267458dcec\") " Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.085859 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2d19437-5742-4516-8c26-52267458dcec-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b2d19437-5742-4516-8c26-52267458dcec" (UID: "b2d19437-5742-4516-8c26-52267458dcec"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.089931 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2d19437-5742-4516-8c26-52267458dcec-kube-api-access-gmw8l" (OuterVolumeSpecName: "kube-api-access-gmw8l") pod "b2d19437-5742-4516-8c26-52267458dcec" (UID: "b2d19437-5742-4516-8c26-52267458dcec"). InnerVolumeSpecName "kube-api-access-gmw8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.187034 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2d19437-5742-4516-8c26-52267458dcec-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.187070 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmw8l\" (UniqueName: \"kubernetes.io/projected/b2d19437-5742-4516-8c26-52267458dcec-kube-api-access-gmw8l\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.281048 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0fac-account-create-update-57jps" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.294396 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kdnv8" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.390957 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d789e4e-7e8f-48c6-94e2-cf1d081d0018-operator-scripts\") pod \"1d789e4e-7e8f-48c6-94e2-cf1d081d0018\" (UID: \"1d789e4e-7e8f-48c6-94e2-cf1d081d0018\") " Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.391415 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q84ws\" (UniqueName: \"kubernetes.io/projected/1d789e4e-7e8f-48c6-94e2-cf1d081d0018-kube-api-access-q84ws\") pod \"1d789e4e-7e8f-48c6-94e2-cf1d081d0018\" (UID: \"1d789e4e-7e8f-48c6-94e2-cf1d081d0018\") " Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.391465 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6e67c08-6bc5-4c10-ac10-80bb918b95d5-operator-scripts\") pod \"e6e67c08-6bc5-4c10-ac10-80bb918b95d5\" (UID: \"e6e67c08-6bc5-4c10-ac10-80bb918b95d5\") " Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.391536 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5vg7\" (UniqueName: \"kubernetes.io/projected/e6e67c08-6bc5-4c10-ac10-80bb918b95d5-kube-api-access-j5vg7\") pod \"e6e67c08-6bc5-4c10-ac10-80bb918b95d5\" (UID: \"e6e67c08-6bc5-4c10-ac10-80bb918b95d5\") " Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.391630 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d789e4e-7e8f-48c6-94e2-cf1d081d0018-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1d789e4e-7e8f-48c6-94e2-cf1d081d0018" (UID: "1d789e4e-7e8f-48c6-94e2-cf1d081d0018"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.392051 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6e67c08-6bc5-4c10-ac10-80bb918b95d5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e6e67c08-6bc5-4c10-ac10-80bb918b95d5" (UID: "e6e67c08-6bc5-4c10-ac10-80bb918b95d5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.392897 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e6e67c08-6bc5-4c10-ac10-80bb918b95d5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.392926 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d789e4e-7e8f-48c6-94e2-cf1d081d0018-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.394636 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d789e4e-7e8f-48c6-94e2-cf1d081d0018-kube-api-access-q84ws" (OuterVolumeSpecName: "kube-api-access-q84ws") pod "1d789e4e-7e8f-48c6-94e2-cf1d081d0018" (UID: "1d789e4e-7e8f-48c6-94e2-cf1d081d0018"). InnerVolumeSpecName "kube-api-access-q84ws". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.395344 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6e67c08-6bc5-4c10-ac10-80bb918b95d5-kube-api-access-j5vg7" (OuterVolumeSpecName: "kube-api-access-j5vg7") pod "e6e67c08-6bc5-4c10-ac10-80bb918b95d5" (UID: "e6e67c08-6bc5-4c10-ac10-80bb918b95d5"). InnerVolumeSpecName "kube-api-access-j5vg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.494674 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q84ws\" (UniqueName: \"kubernetes.io/projected/1d789e4e-7e8f-48c6-94e2-cf1d081d0018-kube-api-access-q84ws\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.494728 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5vg7\" (UniqueName: \"kubernetes.io/projected/e6e67c08-6bc5-4c10-ac10-80bb918b95d5-kube-api-access-j5vg7\") on node \"crc\" DevicePath \"\"" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.845523 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0fac-account-create-update-57jps" event={"ID":"1d789e4e-7e8f-48c6-94e2-cf1d081d0018","Type":"ContainerDied","Data":"3f4d50c633901da574117258b78e87cff186902e41dd9f3b28425eaddcda901c"} Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.845591 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f4d50c633901da574117258b78e87cff186902e41dd9f3b28425eaddcda901c" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.845562 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0fac-account-create-update-57jps" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.849370 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-kdnv8" event={"ID":"e6e67c08-6bc5-4c10-ac10-80bb918b95d5","Type":"ContainerDied","Data":"82281cb109a5d1dc6b49d06b1dac667d58ce374862c687e225c112c859626fc1"} Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.849425 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82281cb109a5d1dc6b49d06b1dac667d58ce374862c687e225c112c859626fc1" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.849537 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-kdnv8" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.854137 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-378c-account-create-update-rbsww" event={"ID":"b2d19437-5742-4516-8c26-52267458dcec","Type":"ContainerDied","Data":"01dafa4c56de45fc78a077fbead7d1c8357cf96ee9276beb7167f206c892302d"} Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.854185 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01dafa4c56de45fc78a077fbead7d1c8357cf96ee9276beb7167f206c892302d" Dec 01 16:14:56 crc kubenswrapper[4810]: I1201 16:14:56.854273 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-378c-account-create-update-rbsww" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.516339 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-82k7s"] Dec 01 16:14:57 crc kubenswrapper[4810]: E1201 16:14:57.518026 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d789e4e-7e8f-48c6-94e2-cf1d081d0018" containerName="mariadb-account-create-update" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.518051 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d789e4e-7e8f-48c6-94e2-cf1d081d0018" containerName="mariadb-account-create-update" Dec 01 16:14:57 crc kubenswrapper[4810]: E1201 16:14:57.518070 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4" containerName="mariadb-account-create-update" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.518078 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4" containerName="mariadb-account-create-update" Dec 01 16:14:57 crc kubenswrapper[4810]: E1201 16:14:57.518106 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6e67c08-6bc5-4c10-ac10-80bb918b95d5" containerName="mariadb-database-create" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.518113 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6e67c08-6bc5-4c10-ac10-80bb918b95d5" containerName="mariadb-database-create" Dec 01 16:14:57 crc kubenswrapper[4810]: E1201 16:14:57.518126 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f61fca3-87ad-4722-9d8d-f212af501835" containerName="mariadb-database-create" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.518131 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f61fca3-87ad-4722-9d8d-f212af501835" containerName="mariadb-database-create" Dec 01 16:14:57 crc kubenswrapper[4810]: E1201 16:14:57.518144 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25ad5508-0e9b-4c6e-86ad-472613af32be" containerName="mariadb-database-create" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.518151 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="25ad5508-0e9b-4c6e-86ad-472613af32be" containerName="mariadb-database-create" Dec 01 16:14:57 crc kubenswrapper[4810]: E1201 16:14:57.518161 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2d19437-5742-4516-8c26-52267458dcec" containerName="mariadb-account-create-update" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.518166 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2d19437-5742-4516-8c26-52267458dcec" containerName="mariadb-account-create-update" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.518541 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d789e4e-7e8f-48c6-94e2-cf1d081d0018" containerName="mariadb-account-create-update" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.518557 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6e67c08-6bc5-4c10-ac10-80bb918b95d5" containerName="mariadb-database-create" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.518575 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="25ad5508-0e9b-4c6e-86ad-472613af32be" containerName="mariadb-database-create" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.518588 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f61fca3-87ad-4722-9d8d-f212af501835" containerName="mariadb-database-create" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.518609 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2d19437-5742-4516-8c26-52267458dcec" containerName="mariadb-account-create-update" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.518620 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4" containerName="mariadb-account-create-update" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.519315 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-82k7s" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.524890 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.525186 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.525458 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-4xg2h" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.529603 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-82k7s"] Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.614204 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed991281-68e4-4e60-ab15-889f970b898d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-82k7s\" (UID: \"ed991281-68e4-4e60-ab15-889f970b898d\") " pod="openstack/nova-cell0-conductor-db-sync-82k7s" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.614416 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed991281-68e4-4e60-ab15-889f970b898d-scripts\") pod \"nova-cell0-conductor-db-sync-82k7s\" (UID: \"ed991281-68e4-4e60-ab15-889f970b898d\") " pod="openstack/nova-cell0-conductor-db-sync-82k7s" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.614610 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-926m5\" (UniqueName: \"kubernetes.io/projected/ed991281-68e4-4e60-ab15-889f970b898d-kube-api-access-926m5\") pod \"nova-cell0-conductor-db-sync-82k7s\" (UID: \"ed991281-68e4-4e60-ab15-889f970b898d\") " pod="openstack/nova-cell0-conductor-db-sync-82k7s" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.614697 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed991281-68e4-4e60-ab15-889f970b898d-config-data\") pod \"nova-cell0-conductor-db-sync-82k7s\" (UID: \"ed991281-68e4-4e60-ab15-889f970b898d\") " pod="openstack/nova-cell0-conductor-db-sync-82k7s" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.716716 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed991281-68e4-4e60-ab15-889f970b898d-scripts\") pod \"nova-cell0-conductor-db-sync-82k7s\" (UID: \"ed991281-68e4-4e60-ab15-889f970b898d\") " pod="openstack/nova-cell0-conductor-db-sync-82k7s" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.717105 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-926m5\" (UniqueName: \"kubernetes.io/projected/ed991281-68e4-4e60-ab15-889f970b898d-kube-api-access-926m5\") pod \"nova-cell0-conductor-db-sync-82k7s\" (UID: \"ed991281-68e4-4e60-ab15-889f970b898d\") " pod="openstack/nova-cell0-conductor-db-sync-82k7s" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.717238 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed991281-68e4-4e60-ab15-889f970b898d-config-data\") pod \"nova-cell0-conductor-db-sync-82k7s\" (UID: \"ed991281-68e4-4e60-ab15-889f970b898d\") " pod="openstack/nova-cell0-conductor-db-sync-82k7s" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.717361 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed991281-68e4-4e60-ab15-889f970b898d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-82k7s\" (UID: \"ed991281-68e4-4e60-ab15-889f970b898d\") " pod="openstack/nova-cell0-conductor-db-sync-82k7s" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.732119 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed991281-68e4-4e60-ab15-889f970b898d-scripts\") pod \"nova-cell0-conductor-db-sync-82k7s\" (UID: \"ed991281-68e4-4e60-ab15-889f970b898d\") " pod="openstack/nova-cell0-conductor-db-sync-82k7s" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.732373 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed991281-68e4-4e60-ab15-889f970b898d-config-data\") pod \"nova-cell0-conductor-db-sync-82k7s\" (UID: \"ed991281-68e4-4e60-ab15-889f970b898d\") " pod="openstack/nova-cell0-conductor-db-sync-82k7s" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.734084 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-926m5\" (UniqueName: \"kubernetes.io/projected/ed991281-68e4-4e60-ab15-889f970b898d-kube-api-access-926m5\") pod \"nova-cell0-conductor-db-sync-82k7s\" (UID: \"ed991281-68e4-4e60-ab15-889f970b898d\") " pod="openstack/nova-cell0-conductor-db-sync-82k7s" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.734614 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed991281-68e4-4e60-ab15-889f970b898d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-82k7s\" (UID: \"ed991281-68e4-4e60-ab15-889f970b898d\") " pod="openstack/nova-cell0-conductor-db-sync-82k7s" Dec 01 16:14:57 crc kubenswrapper[4810]: I1201 16:14:57.840001 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-82k7s" Dec 01 16:14:58 crc kubenswrapper[4810]: I1201 16:14:58.289942 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-82k7s"] Dec 01 16:14:58 crc kubenswrapper[4810]: W1201 16:14:58.293961 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded991281_68e4_4e60_ab15_889f970b898d.slice/crio-ce3180a7ba551eae7135fb1917c0dee5806fdadf99e685734880dd51ba99457d WatchSource:0}: Error finding container ce3180a7ba551eae7135fb1917c0dee5806fdadf99e685734880dd51ba99457d: Status 404 returned error can't find the container with id ce3180a7ba551eae7135fb1917c0dee5806fdadf99e685734880dd51ba99457d Dec 01 16:14:58 crc kubenswrapper[4810]: I1201 16:14:58.872065 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-82k7s" event={"ID":"ed991281-68e4-4e60-ab15-889f970b898d","Type":"ContainerStarted","Data":"ad04d91114a4cde163687103220e8844b8b24c733d5779dd77f33abbcedb2121"} Dec 01 16:14:58 crc kubenswrapper[4810]: I1201 16:14:58.872306 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-82k7s" event={"ID":"ed991281-68e4-4e60-ab15-889f970b898d","Type":"ContainerStarted","Data":"ce3180a7ba551eae7135fb1917c0dee5806fdadf99e685734880dd51ba99457d"} Dec 01 16:14:58 crc kubenswrapper[4810]: I1201 16:14:58.889345 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-82k7s" podStartSLOduration=1.889324641 podStartE2EDuration="1.889324641s" podCreationTimestamp="2025-12-01 16:14:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:14:58.886888484 +0000 UTC m=+6064.650398097" watchObservedRunningTime="2025-12-01 16:14:58.889324641 +0000 UTC m=+6064.652834244" Dec 01 16:15:00 crc kubenswrapper[4810]: I1201 16:15:00.143126 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q"] Dec 01 16:15:00 crc kubenswrapper[4810]: I1201 16:15:00.144985 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q" Dec 01 16:15:00 crc kubenswrapper[4810]: I1201 16:15:00.147643 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 16:15:00 crc kubenswrapper[4810]: I1201 16:15:00.148339 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 16:15:00 crc kubenswrapper[4810]: I1201 16:15:00.156451 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q"] Dec 01 16:15:00 crc kubenswrapper[4810]: I1201 16:15:00.157708 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k498s\" (UniqueName: \"kubernetes.io/projected/8ef74083-dfac-48d7-9f66-77dd31086f68-kube-api-access-k498s\") pod \"collect-profiles-29410095-b7b6q\" (UID: \"8ef74083-dfac-48d7-9f66-77dd31086f68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q" Dec 01 16:15:00 crc kubenswrapper[4810]: I1201 16:15:00.157898 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ef74083-dfac-48d7-9f66-77dd31086f68-config-volume\") pod \"collect-profiles-29410095-b7b6q\" (UID: \"8ef74083-dfac-48d7-9f66-77dd31086f68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q" Dec 01 16:15:00 crc kubenswrapper[4810]: I1201 16:15:00.157972 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ef74083-dfac-48d7-9f66-77dd31086f68-secret-volume\") pod \"collect-profiles-29410095-b7b6q\" (UID: \"8ef74083-dfac-48d7-9f66-77dd31086f68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q" Dec 01 16:15:00 crc kubenswrapper[4810]: I1201 16:15:00.259307 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ef74083-dfac-48d7-9f66-77dd31086f68-config-volume\") pod \"collect-profiles-29410095-b7b6q\" (UID: \"8ef74083-dfac-48d7-9f66-77dd31086f68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q" Dec 01 16:15:00 crc kubenswrapper[4810]: I1201 16:15:00.259372 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ef74083-dfac-48d7-9f66-77dd31086f68-secret-volume\") pod \"collect-profiles-29410095-b7b6q\" (UID: \"8ef74083-dfac-48d7-9f66-77dd31086f68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q" Dec 01 16:15:00 crc kubenswrapper[4810]: I1201 16:15:00.259446 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k498s\" (UniqueName: \"kubernetes.io/projected/8ef74083-dfac-48d7-9f66-77dd31086f68-kube-api-access-k498s\") pod \"collect-profiles-29410095-b7b6q\" (UID: \"8ef74083-dfac-48d7-9f66-77dd31086f68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q" Dec 01 16:15:00 crc kubenswrapper[4810]: I1201 16:15:00.260523 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ef74083-dfac-48d7-9f66-77dd31086f68-config-volume\") pod \"collect-profiles-29410095-b7b6q\" (UID: \"8ef74083-dfac-48d7-9f66-77dd31086f68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q" Dec 01 16:15:00 crc kubenswrapper[4810]: I1201 16:15:00.265261 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ef74083-dfac-48d7-9f66-77dd31086f68-secret-volume\") pod \"collect-profiles-29410095-b7b6q\" (UID: \"8ef74083-dfac-48d7-9f66-77dd31086f68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q" Dec 01 16:15:00 crc kubenswrapper[4810]: I1201 16:15:00.274233 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k498s\" (UniqueName: \"kubernetes.io/projected/8ef74083-dfac-48d7-9f66-77dd31086f68-kube-api-access-k498s\") pod \"collect-profiles-29410095-b7b6q\" (UID: \"8ef74083-dfac-48d7-9f66-77dd31086f68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q" Dec 01 16:15:00 crc kubenswrapper[4810]: I1201 16:15:00.482454 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q" Dec 01 16:15:00 crc kubenswrapper[4810]: I1201 16:15:00.935039 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q"] Dec 01 16:15:00 crc kubenswrapper[4810]: W1201 16:15:00.939646 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ef74083_dfac_48d7_9f66_77dd31086f68.slice/crio-8c0b057b919814343d60b1203f215b941db445db0c027b1c083d24ea48cee09a WatchSource:0}: Error finding container 8c0b057b919814343d60b1203f215b941db445db0c027b1c083d24ea48cee09a: Status 404 returned error can't find the container with id 8c0b057b919814343d60b1203f215b941db445db0c027b1c083d24ea48cee09a Dec 01 16:15:01 crc kubenswrapper[4810]: I1201 16:15:01.897434 4810 generic.go:334] "Generic (PLEG): container finished" podID="8ef74083-dfac-48d7-9f66-77dd31086f68" containerID="07f9b8592d9274a5c91042bff038a1f8b35cf054f99282d72115a68796255d82" exitCode=0 Dec 01 16:15:01 crc kubenswrapper[4810]: I1201 16:15:01.897521 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q" event={"ID":"8ef74083-dfac-48d7-9f66-77dd31086f68","Type":"ContainerDied","Data":"07f9b8592d9274a5c91042bff038a1f8b35cf054f99282d72115a68796255d82"} Dec 01 16:15:01 crc kubenswrapper[4810]: I1201 16:15:01.898743 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q" event={"ID":"8ef74083-dfac-48d7-9f66-77dd31086f68","Type":"ContainerStarted","Data":"8c0b057b919814343d60b1203f215b941db445db0c027b1c083d24ea48cee09a"} Dec 01 16:15:03 crc kubenswrapper[4810]: I1201 16:15:03.218288 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q" Dec 01 16:15:03 crc kubenswrapper[4810]: I1201 16:15:03.322509 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k498s\" (UniqueName: \"kubernetes.io/projected/8ef74083-dfac-48d7-9f66-77dd31086f68-kube-api-access-k498s\") pod \"8ef74083-dfac-48d7-9f66-77dd31086f68\" (UID: \"8ef74083-dfac-48d7-9f66-77dd31086f68\") " Dec 01 16:15:03 crc kubenswrapper[4810]: I1201 16:15:03.322592 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ef74083-dfac-48d7-9f66-77dd31086f68-secret-volume\") pod \"8ef74083-dfac-48d7-9f66-77dd31086f68\" (UID: \"8ef74083-dfac-48d7-9f66-77dd31086f68\") " Dec 01 16:15:03 crc kubenswrapper[4810]: I1201 16:15:03.322779 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ef74083-dfac-48d7-9f66-77dd31086f68-config-volume\") pod \"8ef74083-dfac-48d7-9f66-77dd31086f68\" (UID: \"8ef74083-dfac-48d7-9f66-77dd31086f68\") " Dec 01 16:15:03 crc kubenswrapper[4810]: I1201 16:15:03.323951 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ef74083-dfac-48d7-9f66-77dd31086f68-config-volume" (OuterVolumeSpecName: "config-volume") pod "8ef74083-dfac-48d7-9f66-77dd31086f68" (UID: "8ef74083-dfac-48d7-9f66-77dd31086f68"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:15:03 crc kubenswrapper[4810]: I1201 16:15:03.329630 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ef74083-dfac-48d7-9f66-77dd31086f68-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8ef74083-dfac-48d7-9f66-77dd31086f68" (UID: "8ef74083-dfac-48d7-9f66-77dd31086f68"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:03 crc kubenswrapper[4810]: I1201 16:15:03.329681 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ef74083-dfac-48d7-9f66-77dd31086f68-kube-api-access-k498s" (OuterVolumeSpecName: "kube-api-access-k498s") pod "8ef74083-dfac-48d7-9f66-77dd31086f68" (UID: "8ef74083-dfac-48d7-9f66-77dd31086f68"). InnerVolumeSpecName "kube-api-access-k498s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:03 crc kubenswrapper[4810]: I1201 16:15:03.425361 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ef74083-dfac-48d7-9f66-77dd31086f68-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:03 crc kubenswrapper[4810]: I1201 16:15:03.425398 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k498s\" (UniqueName: \"kubernetes.io/projected/8ef74083-dfac-48d7-9f66-77dd31086f68-kube-api-access-k498s\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:03 crc kubenswrapper[4810]: I1201 16:15:03.425410 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ef74083-dfac-48d7-9f66-77dd31086f68-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:03 crc kubenswrapper[4810]: I1201 16:15:03.936347 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q" Dec 01 16:15:03 crc kubenswrapper[4810]: I1201 16:15:03.936354 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q" event={"ID":"8ef74083-dfac-48d7-9f66-77dd31086f68","Type":"ContainerDied","Data":"8c0b057b919814343d60b1203f215b941db445db0c027b1c083d24ea48cee09a"} Dec 01 16:15:03 crc kubenswrapper[4810]: I1201 16:15:03.936865 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c0b057b919814343d60b1203f215b941db445db0c027b1c083d24ea48cee09a" Dec 01 16:15:03 crc kubenswrapper[4810]: I1201 16:15:03.938516 4810 generic.go:334] "Generic (PLEG): container finished" podID="ed991281-68e4-4e60-ab15-889f970b898d" containerID="ad04d91114a4cde163687103220e8844b8b24c733d5779dd77f33abbcedb2121" exitCode=0 Dec 01 16:15:03 crc kubenswrapper[4810]: I1201 16:15:03.938568 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-82k7s" event={"ID":"ed991281-68e4-4e60-ab15-889f970b898d","Type":"ContainerDied","Data":"ad04d91114a4cde163687103220e8844b8b24c733d5779dd77f33abbcedb2121"} Dec 01 16:15:04 crc kubenswrapper[4810]: I1201 16:15:04.290389 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn"] Dec 01 16:15:04 crc kubenswrapper[4810]: I1201 16:15:04.298584 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410050-g29pn"] Dec 01 16:15:04 crc kubenswrapper[4810]: I1201 16:15:04.503234 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70be3d5e-ec64-4b22-8cc7-c071e13ed43c" path="/var/lib/kubelet/pods/70be3d5e-ec64-4b22-8cc7-c071e13ed43c/volumes" Dec 01 16:15:05 crc kubenswrapper[4810]: I1201 16:15:05.251556 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-82k7s" Dec 01 16:15:05 crc kubenswrapper[4810]: I1201 16:15:05.358372 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed991281-68e4-4e60-ab15-889f970b898d-scripts\") pod \"ed991281-68e4-4e60-ab15-889f970b898d\" (UID: \"ed991281-68e4-4e60-ab15-889f970b898d\") " Dec 01 16:15:05 crc kubenswrapper[4810]: I1201 16:15:05.358493 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed991281-68e4-4e60-ab15-889f970b898d-combined-ca-bundle\") pod \"ed991281-68e4-4e60-ab15-889f970b898d\" (UID: \"ed991281-68e4-4e60-ab15-889f970b898d\") " Dec 01 16:15:05 crc kubenswrapper[4810]: I1201 16:15:05.358711 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-926m5\" (UniqueName: \"kubernetes.io/projected/ed991281-68e4-4e60-ab15-889f970b898d-kube-api-access-926m5\") pod \"ed991281-68e4-4e60-ab15-889f970b898d\" (UID: \"ed991281-68e4-4e60-ab15-889f970b898d\") " Dec 01 16:15:05 crc kubenswrapper[4810]: I1201 16:15:05.358804 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed991281-68e4-4e60-ab15-889f970b898d-config-data\") pod \"ed991281-68e4-4e60-ab15-889f970b898d\" (UID: \"ed991281-68e4-4e60-ab15-889f970b898d\") " Dec 01 16:15:05 crc kubenswrapper[4810]: I1201 16:15:05.363662 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed991281-68e4-4e60-ab15-889f970b898d-kube-api-access-926m5" (OuterVolumeSpecName: "kube-api-access-926m5") pod "ed991281-68e4-4e60-ab15-889f970b898d" (UID: "ed991281-68e4-4e60-ab15-889f970b898d"). InnerVolumeSpecName "kube-api-access-926m5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:05 crc kubenswrapper[4810]: I1201 16:15:05.363887 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed991281-68e4-4e60-ab15-889f970b898d-scripts" (OuterVolumeSpecName: "scripts") pod "ed991281-68e4-4e60-ab15-889f970b898d" (UID: "ed991281-68e4-4e60-ab15-889f970b898d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:05 crc kubenswrapper[4810]: I1201 16:15:05.384109 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed991281-68e4-4e60-ab15-889f970b898d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed991281-68e4-4e60-ab15-889f970b898d" (UID: "ed991281-68e4-4e60-ab15-889f970b898d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:05 crc kubenswrapper[4810]: I1201 16:15:05.384608 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed991281-68e4-4e60-ab15-889f970b898d-config-data" (OuterVolumeSpecName: "config-data") pod "ed991281-68e4-4e60-ab15-889f970b898d" (UID: "ed991281-68e4-4e60-ab15-889f970b898d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:05 crc kubenswrapper[4810]: I1201 16:15:05.461834 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed991281-68e4-4e60-ab15-889f970b898d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:05 crc kubenswrapper[4810]: I1201 16:15:05.461879 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-926m5\" (UniqueName: \"kubernetes.io/projected/ed991281-68e4-4e60-ab15-889f970b898d-kube-api-access-926m5\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:05 crc kubenswrapper[4810]: I1201 16:15:05.461894 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed991281-68e4-4e60-ab15-889f970b898d-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:05 crc kubenswrapper[4810]: I1201 16:15:05.461907 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed991281-68e4-4e60-ab15-889f970b898d-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:05 crc kubenswrapper[4810]: I1201 16:15:05.960044 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-82k7s" event={"ID":"ed991281-68e4-4e60-ab15-889f970b898d","Type":"ContainerDied","Data":"ce3180a7ba551eae7135fb1917c0dee5806fdadf99e685734880dd51ba99457d"} Dec 01 16:15:05 crc kubenswrapper[4810]: I1201 16:15:05.960085 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce3180a7ba551eae7135fb1917c0dee5806fdadf99e685734880dd51ba99457d" Dec 01 16:15:05 crc kubenswrapper[4810]: I1201 16:15:05.960137 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-82k7s" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.041444 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 16:15:06 crc kubenswrapper[4810]: E1201 16:15:06.041934 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ef74083-dfac-48d7-9f66-77dd31086f68" containerName="collect-profiles" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.041956 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ef74083-dfac-48d7-9f66-77dd31086f68" containerName="collect-profiles" Dec 01 16:15:06 crc kubenswrapper[4810]: E1201 16:15:06.041990 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed991281-68e4-4e60-ab15-889f970b898d" containerName="nova-cell0-conductor-db-sync" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.041999 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed991281-68e4-4e60-ab15-889f970b898d" containerName="nova-cell0-conductor-db-sync" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.042283 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed991281-68e4-4e60-ab15-889f970b898d" containerName="nova-cell0-conductor-db-sync" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.042311 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ef74083-dfac-48d7-9f66-77dd31086f68" containerName="collect-profiles" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.043177 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.085407 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-4xg2h" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.086619 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/578dc44d-25f5-4ef1-bbe7-eb2328954767-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"578dc44d-25f5-4ef1-bbe7-eb2328954767\") " pod="openstack/nova-cell0-conductor-0" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.086690 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt2nv\" (UniqueName: \"kubernetes.io/projected/578dc44d-25f5-4ef1-bbe7-eb2328954767-kube-api-access-mt2nv\") pod \"nova-cell0-conductor-0\" (UID: \"578dc44d-25f5-4ef1-bbe7-eb2328954767\") " pod="openstack/nova-cell0-conductor-0" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.086864 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.086987 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/578dc44d-25f5-4ef1-bbe7-eb2328954767-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"578dc44d-25f5-4ef1-bbe7-eb2328954767\") " pod="openstack/nova-cell0-conductor-0" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.101697 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.188086 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/578dc44d-25f5-4ef1-bbe7-eb2328954767-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"578dc44d-25f5-4ef1-bbe7-eb2328954767\") " pod="openstack/nova-cell0-conductor-0" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.188655 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/578dc44d-25f5-4ef1-bbe7-eb2328954767-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"578dc44d-25f5-4ef1-bbe7-eb2328954767\") " pod="openstack/nova-cell0-conductor-0" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.188687 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt2nv\" (UniqueName: \"kubernetes.io/projected/578dc44d-25f5-4ef1-bbe7-eb2328954767-kube-api-access-mt2nv\") pod \"nova-cell0-conductor-0\" (UID: \"578dc44d-25f5-4ef1-bbe7-eb2328954767\") " pod="openstack/nova-cell0-conductor-0" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.191968 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/578dc44d-25f5-4ef1-bbe7-eb2328954767-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"578dc44d-25f5-4ef1-bbe7-eb2328954767\") " pod="openstack/nova-cell0-conductor-0" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.192114 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/578dc44d-25f5-4ef1-bbe7-eb2328954767-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"578dc44d-25f5-4ef1-bbe7-eb2328954767\") " pod="openstack/nova-cell0-conductor-0" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.204975 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt2nv\" (UniqueName: \"kubernetes.io/projected/578dc44d-25f5-4ef1-bbe7-eb2328954767-kube-api-access-mt2nv\") pod \"nova-cell0-conductor-0\" (UID: \"578dc44d-25f5-4ef1-bbe7-eb2328954767\") " pod="openstack/nova-cell0-conductor-0" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.406661 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.829012 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 16:15:06 crc kubenswrapper[4810]: I1201 16:15:06.969670 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"578dc44d-25f5-4ef1-bbe7-eb2328954767","Type":"ContainerStarted","Data":"a35bd522cf8087246aea95321c33263df1b1816e1052396c74e09e5eee743665"} Dec 01 16:15:07 crc kubenswrapper[4810]: I1201 16:15:07.491215 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:15:07 crc kubenswrapper[4810]: E1201 16:15:07.492985 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:15:07 crc kubenswrapper[4810]: I1201 16:15:07.979550 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"578dc44d-25f5-4ef1-bbe7-eb2328954767","Type":"ContainerStarted","Data":"2f4d53685fe68630eb2427db02a9e6c2864c578f0eee81039835517195889ab1"} Dec 01 16:15:07 crc kubenswrapper[4810]: I1201 16:15:07.980135 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 01 16:15:08 crc kubenswrapper[4810]: I1201 16:15:08.012727 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.012692692 podStartE2EDuration="2.012692692s" podCreationTimestamp="2025-12-01 16:15:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:15:07.996102192 +0000 UTC m=+6073.759611805" watchObservedRunningTime="2025-12-01 16:15:08.012692692 +0000 UTC m=+6073.776202315" Dec 01 16:15:09 crc kubenswrapper[4810]: I1201 16:15:09.907459 4810 scope.go:117] "RemoveContainer" containerID="2195a8be29b3f937a0e6244ce829f622b65caef3a438114604d3e7440f5401dc" Dec 01 16:15:16 crc kubenswrapper[4810]: I1201 16:15:16.432487 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 01 16:15:16 crc kubenswrapper[4810]: I1201 16:15:16.861789 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-llq2h"] Dec 01 16:15:16 crc kubenswrapper[4810]: I1201 16:15:16.863537 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-llq2h" Dec 01 16:15:16 crc kubenswrapper[4810]: I1201 16:15:16.865914 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 01 16:15:16 crc kubenswrapper[4810]: I1201 16:15:16.868880 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 01 16:15:16 crc kubenswrapper[4810]: I1201 16:15:16.887270 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-llq2h"] Dec 01 16:15:16 crc kubenswrapper[4810]: I1201 16:15:16.991636 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf7g2\" (UniqueName: \"kubernetes.io/projected/37eef714-f1dc-47b3-8489-5204d08789bb-kube-api-access-rf7g2\") pod \"nova-cell0-cell-mapping-llq2h\" (UID: \"37eef714-f1dc-47b3-8489-5204d08789bb\") " pod="openstack/nova-cell0-cell-mapping-llq2h" Dec 01 16:15:16 crc kubenswrapper[4810]: I1201 16:15:16.991683 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37eef714-f1dc-47b3-8489-5204d08789bb-scripts\") pod \"nova-cell0-cell-mapping-llq2h\" (UID: \"37eef714-f1dc-47b3-8489-5204d08789bb\") " pod="openstack/nova-cell0-cell-mapping-llq2h" Dec 01 16:15:16 crc kubenswrapper[4810]: I1201 16:15:16.991757 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37eef714-f1dc-47b3-8489-5204d08789bb-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-llq2h\" (UID: \"37eef714-f1dc-47b3-8489-5204d08789bb\") " pod="openstack/nova-cell0-cell-mapping-llq2h" Dec 01 16:15:16 crc kubenswrapper[4810]: I1201 16:15:16.991853 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37eef714-f1dc-47b3-8489-5204d08789bb-config-data\") pod \"nova-cell0-cell-mapping-llq2h\" (UID: \"37eef714-f1dc-47b3-8489-5204d08789bb\") " pod="openstack/nova-cell0-cell-mapping-llq2h" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.000261 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.002121 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.005308 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.036362 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.052390 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.054367 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.069813 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.093108 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf7g2\" (UniqueName: \"kubernetes.io/projected/37eef714-f1dc-47b3-8489-5204d08789bb-kube-api-access-rf7g2\") pod \"nova-cell0-cell-mapping-llq2h\" (UID: \"37eef714-f1dc-47b3-8489-5204d08789bb\") " pod="openstack/nova-cell0-cell-mapping-llq2h" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.093165 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37eef714-f1dc-47b3-8489-5204d08789bb-scripts\") pod \"nova-cell0-cell-mapping-llq2h\" (UID: \"37eef714-f1dc-47b3-8489-5204d08789bb\") " pod="openstack/nova-cell0-cell-mapping-llq2h" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.093210 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fcbd109-53af-4ba8-9418-10466117a3d3-logs\") pod \"nova-api-0\" (UID: \"3fcbd109-53af-4ba8-9418-10466117a3d3\") " pod="openstack/nova-api-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.093249 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fcbd109-53af-4ba8-9418-10466117a3d3-config-data\") pod \"nova-api-0\" (UID: \"3fcbd109-53af-4ba8-9418-10466117a3d3\") " pod="openstack/nova-api-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.093294 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fcbd109-53af-4ba8-9418-10466117a3d3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3fcbd109-53af-4ba8-9418-10466117a3d3\") " pod="openstack/nova-api-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.093322 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37eef714-f1dc-47b3-8489-5204d08789bb-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-llq2h\" (UID: \"37eef714-f1dc-47b3-8489-5204d08789bb\") " pod="openstack/nova-cell0-cell-mapping-llq2h" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.093425 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z4gj\" (UniqueName: \"kubernetes.io/projected/3fcbd109-53af-4ba8-9418-10466117a3d3-kube-api-access-4z4gj\") pod \"nova-api-0\" (UID: \"3fcbd109-53af-4ba8-9418-10466117a3d3\") " pod="openstack/nova-api-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.093524 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37eef714-f1dc-47b3-8489-5204d08789bb-config-data\") pod \"nova-cell0-cell-mapping-llq2h\" (UID: \"37eef714-f1dc-47b3-8489-5204d08789bb\") " pod="openstack/nova-cell0-cell-mapping-llq2h" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.105077 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.107460 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37eef714-f1dc-47b3-8489-5204d08789bb-scripts\") pod \"nova-cell0-cell-mapping-llq2h\" (UID: \"37eef714-f1dc-47b3-8489-5204d08789bb\") " pod="openstack/nova-cell0-cell-mapping-llq2h" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.112860 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37eef714-f1dc-47b3-8489-5204d08789bb-config-data\") pod \"nova-cell0-cell-mapping-llq2h\" (UID: \"37eef714-f1dc-47b3-8489-5204d08789bb\") " pod="openstack/nova-cell0-cell-mapping-llq2h" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.131244 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37eef714-f1dc-47b3-8489-5204d08789bb-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-llq2h\" (UID: \"37eef714-f1dc-47b3-8489-5204d08789bb\") " pod="openstack/nova-cell0-cell-mapping-llq2h" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.145076 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf7g2\" (UniqueName: \"kubernetes.io/projected/37eef714-f1dc-47b3-8489-5204d08789bb-kube-api-access-rf7g2\") pod \"nova-cell0-cell-mapping-llq2h\" (UID: \"37eef714-f1dc-47b3-8489-5204d08789bb\") " pod="openstack/nova-cell0-cell-mapping-llq2h" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.148395 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.150484 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.155823 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.191711 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.192747 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-llq2h" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.203139 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dc35829-d42b-4852-8bb7-f3ba752409da-logs\") pod \"nova-metadata-0\" (UID: \"2dc35829-d42b-4852-8bb7-f3ba752409da\") " pod="openstack/nova-metadata-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.203883 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fcbd109-53af-4ba8-9418-10466117a3d3-logs\") pod \"nova-api-0\" (UID: \"3fcbd109-53af-4ba8-9418-10466117a3d3\") " pod="openstack/nova-api-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.203465 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fcbd109-53af-4ba8-9418-10466117a3d3-logs\") pod \"nova-api-0\" (UID: \"3fcbd109-53af-4ba8-9418-10466117a3d3\") " pod="openstack/nova-api-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.204149 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fcbd109-53af-4ba8-9418-10466117a3d3-config-data\") pod \"nova-api-0\" (UID: \"3fcbd109-53af-4ba8-9418-10466117a3d3\") " pod="openstack/nova-api-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.204844 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fcbd109-53af-4ba8-9418-10466117a3d3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3fcbd109-53af-4ba8-9418-10466117a3d3\") " pod="openstack/nova-api-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.205067 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dc35829-d42b-4852-8bb7-f3ba752409da-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2dc35829-d42b-4852-8bb7-f3ba752409da\") " pod="openstack/nova-metadata-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.209068 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z4gj\" (UniqueName: \"kubernetes.io/projected/3fcbd109-53af-4ba8-9418-10466117a3d3-kube-api-access-4z4gj\") pod \"nova-api-0\" (UID: \"3fcbd109-53af-4ba8-9418-10466117a3d3\") " pod="openstack/nova-api-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.209178 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dc35829-d42b-4852-8bb7-f3ba752409da-config-data\") pod \"nova-metadata-0\" (UID: \"2dc35829-d42b-4852-8bb7-f3ba752409da\") " pod="openstack/nova-metadata-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.209322 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8rv6\" (UniqueName: \"kubernetes.io/projected/2dc35829-d42b-4852-8bb7-f3ba752409da-kube-api-access-g8rv6\") pod \"nova-metadata-0\" (UID: \"2dc35829-d42b-4852-8bb7-f3ba752409da\") " pod="openstack/nova-metadata-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.223195 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fcbd109-53af-4ba8-9418-10466117a3d3-config-data\") pod \"nova-api-0\" (UID: \"3fcbd109-53af-4ba8-9418-10466117a3d3\") " pod="openstack/nova-api-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.237002 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fcbd109-53af-4ba8-9418-10466117a3d3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3fcbd109-53af-4ba8-9418-10466117a3d3\") " pod="openstack/nova-api-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.240966 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z4gj\" (UniqueName: \"kubernetes.io/projected/3fcbd109-53af-4ba8-9418-10466117a3d3-kube-api-access-4z4gj\") pod \"nova-api-0\" (UID: \"3fcbd109-53af-4ba8-9418-10466117a3d3\") " pod="openstack/nova-api-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.256820 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-867b4b5c7c-k9wk9"] Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.258675 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.293514 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.298393 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.306758 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.312625 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.318669 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dc35829-d42b-4852-8bb7-f3ba752409da-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2dc35829-d42b-4852-8bb7-f3ba752409da\") " pod="openstack/nova-metadata-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.318753 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dc35829-d42b-4852-8bb7-f3ba752409da-config-data\") pod \"nova-metadata-0\" (UID: \"2dc35829-d42b-4852-8bb7-f3ba752409da\") " pod="openstack/nova-metadata-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.318798 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8rv6\" (UniqueName: \"kubernetes.io/projected/2dc35829-d42b-4852-8bb7-f3ba752409da-kube-api-access-g8rv6\") pod \"nova-metadata-0\" (UID: \"2dc35829-d42b-4852-8bb7-f3ba752409da\") " pod="openstack/nova-metadata-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.318849 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/febfc322-9d2b-470b-bfdf-f7ee31996fc8-config-data\") pod \"nova-scheduler-0\" (UID: \"febfc322-9d2b-470b-bfdf-f7ee31996fc8\") " pod="openstack/nova-scheduler-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.318882 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2qfj\" (UniqueName: \"kubernetes.io/projected/febfc322-9d2b-470b-bfdf-f7ee31996fc8-kube-api-access-s2qfj\") pod \"nova-scheduler-0\" (UID: \"febfc322-9d2b-470b-bfdf-f7ee31996fc8\") " pod="openstack/nova-scheduler-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.318917 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dc35829-d42b-4852-8bb7-f3ba752409da-logs\") pod \"nova-metadata-0\" (UID: \"2dc35829-d42b-4852-8bb7-f3ba752409da\") " pod="openstack/nova-metadata-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.318954 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/febfc322-9d2b-470b-bfdf-f7ee31996fc8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"febfc322-9d2b-470b-bfdf-f7ee31996fc8\") " pod="openstack/nova-scheduler-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.324539 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.334988 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dc35829-d42b-4852-8bb7-f3ba752409da-logs\") pod \"nova-metadata-0\" (UID: \"2dc35829-d42b-4852-8bb7-f3ba752409da\") " pod="openstack/nova-metadata-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.342284 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dc35829-d42b-4852-8bb7-f3ba752409da-config-data\") pod \"nova-metadata-0\" (UID: \"2dc35829-d42b-4852-8bb7-f3ba752409da\") " pod="openstack/nova-metadata-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.343603 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dc35829-d42b-4852-8bb7-f3ba752409da-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2dc35829-d42b-4852-8bb7-f3ba752409da\") " pod="openstack/nova-metadata-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.348853 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8rv6\" (UniqueName: \"kubernetes.io/projected/2dc35829-d42b-4852-8bb7-f3ba752409da-kube-api-access-g8rv6\") pod \"nova-metadata-0\" (UID: \"2dc35829-d42b-4852-8bb7-f3ba752409da\") " pod="openstack/nova-metadata-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.374276 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-867b4b5c7c-k9wk9"] Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.385279 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.422774 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-ovsdbserver-sb\") pod \"dnsmasq-dns-867b4b5c7c-k9wk9\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.422847 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03364899-f3a3-44cc-8514-78257dbd82d3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"03364899-f3a3-44cc-8514-78257dbd82d3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.422870 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ps7h\" (UniqueName: \"kubernetes.io/projected/59e48d6b-851f-4977-bb5e-9720f697ef6c-kube-api-access-8ps7h\") pod \"dnsmasq-dns-867b4b5c7c-k9wk9\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.422888 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03364899-f3a3-44cc-8514-78257dbd82d3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"03364899-f3a3-44cc-8514-78257dbd82d3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.422905 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-config\") pod \"dnsmasq-dns-867b4b5c7c-k9wk9\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.422926 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b8v5\" (UniqueName: \"kubernetes.io/projected/03364899-f3a3-44cc-8514-78257dbd82d3-kube-api-access-8b8v5\") pod \"nova-cell1-novncproxy-0\" (UID: \"03364899-f3a3-44cc-8514-78257dbd82d3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.422963 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/febfc322-9d2b-470b-bfdf-f7ee31996fc8-config-data\") pod \"nova-scheduler-0\" (UID: \"febfc322-9d2b-470b-bfdf-f7ee31996fc8\") " pod="openstack/nova-scheduler-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.422987 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2qfj\" (UniqueName: \"kubernetes.io/projected/febfc322-9d2b-470b-bfdf-f7ee31996fc8-kube-api-access-s2qfj\") pod \"nova-scheduler-0\" (UID: \"febfc322-9d2b-470b-bfdf-f7ee31996fc8\") " pod="openstack/nova-scheduler-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.423018 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/febfc322-9d2b-470b-bfdf-f7ee31996fc8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"febfc322-9d2b-470b-bfdf-f7ee31996fc8\") " pod="openstack/nova-scheduler-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.423038 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-dns-svc\") pod \"dnsmasq-dns-867b4b5c7c-k9wk9\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.423069 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-ovsdbserver-nb\") pod \"dnsmasq-dns-867b4b5c7c-k9wk9\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.427457 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/febfc322-9d2b-470b-bfdf-f7ee31996fc8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"febfc322-9d2b-470b-bfdf-f7ee31996fc8\") " pod="openstack/nova-scheduler-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.428056 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/febfc322-9d2b-470b-bfdf-f7ee31996fc8-config-data\") pod \"nova-scheduler-0\" (UID: \"febfc322-9d2b-470b-bfdf-f7ee31996fc8\") " pod="openstack/nova-scheduler-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.453500 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2qfj\" (UniqueName: \"kubernetes.io/projected/febfc322-9d2b-470b-bfdf-f7ee31996fc8-kube-api-access-s2qfj\") pod \"nova-scheduler-0\" (UID: \"febfc322-9d2b-470b-bfdf-f7ee31996fc8\") " pod="openstack/nova-scheduler-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.525705 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-ovsdbserver-nb\") pod \"dnsmasq-dns-867b4b5c7c-k9wk9\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.525801 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-ovsdbserver-sb\") pod \"dnsmasq-dns-867b4b5c7c-k9wk9\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.525884 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03364899-f3a3-44cc-8514-78257dbd82d3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"03364899-f3a3-44cc-8514-78257dbd82d3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.525919 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ps7h\" (UniqueName: \"kubernetes.io/projected/59e48d6b-851f-4977-bb5e-9720f697ef6c-kube-api-access-8ps7h\") pod \"dnsmasq-dns-867b4b5c7c-k9wk9\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.525949 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03364899-f3a3-44cc-8514-78257dbd82d3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"03364899-f3a3-44cc-8514-78257dbd82d3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.525974 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-config\") pod \"dnsmasq-dns-867b4b5c7c-k9wk9\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.526003 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b8v5\" (UniqueName: \"kubernetes.io/projected/03364899-f3a3-44cc-8514-78257dbd82d3-kube-api-access-8b8v5\") pod \"nova-cell1-novncproxy-0\" (UID: \"03364899-f3a3-44cc-8514-78257dbd82d3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.526085 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-dns-svc\") pod \"dnsmasq-dns-867b4b5c7c-k9wk9\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.527926 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-ovsdbserver-nb\") pod \"dnsmasq-dns-867b4b5c7c-k9wk9\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.528020 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-config\") pod \"dnsmasq-dns-867b4b5c7c-k9wk9\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.535195 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-dns-svc\") pod \"dnsmasq-dns-867b4b5c7c-k9wk9\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.537836 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-ovsdbserver-sb\") pod \"dnsmasq-dns-867b4b5c7c-k9wk9\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.545982 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03364899-f3a3-44cc-8514-78257dbd82d3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"03364899-f3a3-44cc-8514-78257dbd82d3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.559177 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b8v5\" (UniqueName: \"kubernetes.io/projected/03364899-f3a3-44cc-8514-78257dbd82d3-kube-api-access-8b8v5\") pod \"nova-cell1-novncproxy-0\" (UID: \"03364899-f3a3-44cc-8514-78257dbd82d3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.559278 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03364899-f3a3-44cc-8514-78257dbd82d3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"03364899-f3a3-44cc-8514-78257dbd82d3\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.560547 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ps7h\" (UniqueName: \"kubernetes.io/projected/59e48d6b-851f-4977-bb5e-9720f697ef6c-kube-api-access-8ps7h\") pod \"dnsmasq-dns-867b4b5c7c-k9wk9\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.694958 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.713193 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.722865 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.935438 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-llq2h"] Dec 01 16:15:17 crc kubenswrapper[4810]: I1201 16:15:17.991448 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.076138 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3fcbd109-53af-4ba8-9418-10466117a3d3","Type":"ContainerStarted","Data":"c55d8fe051997af708c2fe07e693fbcbf6942b6f29f514b3a654b4110929fc79"} Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.109558 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2m2vb"] Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.110771 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2m2vb" Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.112070 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-llq2h" event={"ID":"37eef714-f1dc-47b3-8489-5204d08789bb","Type":"ContainerStarted","Data":"fb75d8b4d0244d16fe2b761e5030de745df5e1c509429fd4b9ba875e3d0b2576"} Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.125040 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.125183 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.127390 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.183309 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2m2vb"] Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.250868 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f99947-fa7b-43af-9d9d-640e6102771d-config-data\") pod \"nova-cell1-conductor-db-sync-2m2vb\" (UID: \"95f99947-fa7b-43af-9d9d-640e6102771d\") " pod="openstack/nova-cell1-conductor-db-sync-2m2vb" Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.250939 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h8sw\" (UniqueName: \"kubernetes.io/projected/95f99947-fa7b-43af-9d9d-640e6102771d-kube-api-access-8h8sw\") pod \"nova-cell1-conductor-db-sync-2m2vb\" (UID: \"95f99947-fa7b-43af-9d9d-640e6102771d\") " pod="openstack/nova-cell1-conductor-db-sync-2m2vb" Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.250966 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f99947-fa7b-43af-9d9d-640e6102771d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-2m2vb\" (UID: \"95f99947-fa7b-43af-9d9d-640e6102771d\") " pod="openstack/nova-cell1-conductor-db-sync-2m2vb" Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.250983 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95f99947-fa7b-43af-9d9d-640e6102771d-scripts\") pod \"nova-cell1-conductor-db-sync-2m2vb\" (UID: \"95f99947-fa7b-43af-9d9d-640e6102771d\") " pod="openstack/nova-cell1-conductor-db-sync-2m2vb" Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.284643 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 16:15:18 crc kubenswrapper[4810]: W1201 16:15:18.291224 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03364899_f3a3_44cc_8514_78257dbd82d3.slice/crio-77c6b2d3a5eb73b86bfef2426f975afc2e31253ad88229bfebd8591eb242aba7 WatchSource:0}: Error finding container 77c6b2d3a5eb73b86bfef2426f975afc2e31253ad88229bfebd8591eb242aba7: Status 404 returned error can't find the container with id 77c6b2d3a5eb73b86bfef2426f975afc2e31253ad88229bfebd8591eb242aba7 Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.305183 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.353025 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f99947-fa7b-43af-9d9d-640e6102771d-config-data\") pod \"nova-cell1-conductor-db-sync-2m2vb\" (UID: \"95f99947-fa7b-43af-9d9d-640e6102771d\") " pod="openstack/nova-cell1-conductor-db-sync-2m2vb" Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.353079 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h8sw\" (UniqueName: \"kubernetes.io/projected/95f99947-fa7b-43af-9d9d-640e6102771d-kube-api-access-8h8sw\") pod \"nova-cell1-conductor-db-sync-2m2vb\" (UID: \"95f99947-fa7b-43af-9d9d-640e6102771d\") " pod="openstack/nova-cell1-conductor-db-sync-2m2vb" Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.353112 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f99947-fa7b-43af-9d9d-640e6102771d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-2m2vb\" (UID: \"95f99947-fa7b-43af-9d9d-640e6102771d\") " pod="openstack/nova-cell1-conductor-db-sync-2m2vb" Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.353130 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95f99947-fa7b-43af-9d9d-640e6102771d-scripts\") pod \"nova-cell1-conductor-db-sync-2m2vb\" (UID: \"95f99947-fa7b-43af-9d9d-640e6102771d\") " pod="openstack/nova-cell1-conductor-db-sync-2m2vb" Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.361439 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f99947-fa7b-43af-9d9d-640e6102771d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-2m2vb\" (UID: \"95f99947-fa7b-43af-9d9d-640e6102771d\") " pod="openstack/nova-cell1-conductor-db-sync-2m2vb" Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.364212 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95f99947-fa7b-43af-9d9d-640e6102771d-scripts\") pod \"nova-cell1-conductor-db-sync-2m2vb\" (UID: \"95f99947-fa7b-43af-9d9d-640e6102771d\") " pod="openstack/nova-cell1-conductor-db-sync-2m2vb" Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.365010 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f99947-fa7b-43af-9d9d-640e6102771d-config-data\") pod \"nova-cell1-conductor-db-sync-2m2vb\" (UID: \"95f99947-fa7b-43af-9d9d-640e6102771d\") " pod="openstack/nova-cell1-conductor-db-sync-2m2vb" Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.379425 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h8sw\" (UniqueName: \"kubernetes.io/projected/95f99947-fa7b-43af-9d9d-640e6102771d-kube-api-access-8h8sw\") pod \"nova-cell1-conductor-db-sync-2m2vb\" (UID: \"95f99947-fa7b-43af-9d9d-640e6102771d\") " pod="openstack/nova-cell1-conductor-db-sync-2m2vb" Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.407057 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-867b4b5c7c-k9wk9"] Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.497787 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:15:18 crc kubenswrapper[4810]: E1201 16:15:18.497992 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:15:18 crc kubenswrapper[4810]: I1201 16:15:18.530402 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2m2vb" Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.130276 4810 generic.go:334] "Generic (PLEG): container finished" podID="59e48d6b-851f-4977-bb5e-9720f697ef6c" containerID="78572dc9cd3125f374819f01e0e59b8ed100f18300c7c8d7e9edb5257ad15a66" exitCode=0 Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.130382 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" event={"ID":"59e48d6b-851f-4977-bb5e-9720f697ef6c","Type":"ContainerDied","Data":"78572dc9cd3125f374819f01e0e59b8ed100f18300c7c8d7e9edb5257ad15a66"} Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.130928 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" event={"ID":"59e48d6b-851f-4977-bb5e-9720f697ef6c","Type":"ContainerStarted","Data":"89cbcd209b6be7d6eb189ddc52c94915d59aef477f10dd3ff4ba0e200e1a2486"} Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.135065 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2dc35829-d42b-4852-8bb7-f3ba752409da","Type":"ContainerStarted","Data":"607000e7ede7cba1ab8433636ec44e3895c8381aed743313c1cc08662957e3ef"} Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.135106 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2dc35829-d42b-4852-8bb7-f3ba752409da","Type":"ContainerStarted","Data":"14f3e3008cb05ebaae3fe14ef15377e71981ff3bc97ab4d07a9b1f1ece2875c4"} Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.135116 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2dc35829-d42b-4852-8bb7-f3ba752409da","Type":"ContainerStarted","Data":"5f366d8cd6edfa337e4d2830652cfbf57be9bab9bc89820849b681401d2c9c12"} Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.145066 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"febfc322-9d2b-470b-bfdf-f7ee31996fc8","Type":"ContainerStarted","Data":"6a2eec563994f549af7272b80ded697cd4379e34fc30dd094901d08ad0f9408c"} Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.145395 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"febfc322-9d2b-470b-bfdf-f7ee31996fc8","Type":"ContainerStarted","Data":"5157e0ad3bc0bb968412975155d21da950ea5f0a80421aea49fc9f7ffcdab2f2"} Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.148231 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3fcbd109-53af-4ba8-9418-10466117a3d3","Type":"ContainerStarted","Data":"a48b965be1fb027015443f27baee0ee5689d05f71ed3a69d979efc1f793ddf01"} Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.148267 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3fcbd109-53af-4ba8-9418-10466117a3d3","Type":"ContainerStarted","Data":"bc231bdd562c07a555c11a5594ba7d693ad7210f6b005e2f7d9065668eb5380a"} Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.167852 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-llq2h" event={"ID":"37eef714-f1dc-47b3-8489-5204d08789bb","Type":"ContainerStarted","Data":"ff0b3f547bd0fe4827488b9725fc4957403a6bc1f94a84d700e80db21c809c1e"} Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.188212 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"03364899-f3a3-44cc-8514-78257dbd82d3","Type":"ContainerStarted","Data":"2c40e25e48567b25c4495e3bf43d30a0cbbab7e1fa3b6c755dbdd915487db172"} Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.188257 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"03364899-f3a3-44cc-8514-78257dbd82d3","Type":"ContainerStarted","Data":"77c6b2d3a5eb73b86bfef2426f975afc2e31253ad88229bfebd8591eb242aba7"} Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.202307 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2m2vb"] Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.228882 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.228857607 podStartE2EDuration="2.228857607s" podCreationTimestamp="2025-12-01 16:15:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:15:19.188521052 +0000 UTC m=+6084.952030655" watchObservedRunningTime="2025-12-01 16:15:19.228857607 +0000 UTC m=+6084.992367210" Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.232352 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.232334591 podStartE2EDuration="2.232334591s" podCreationTimestamp="2025-12-01 16:15:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:15:19.210033455 +0000 UTC m=+6084.973543058" watchObservedRunningTime="2025-12-01 16:15:19.232334591 +0000 UTC m=+6084.995844204" Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.270561 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.270539578 podStartE2EDuration="3.270539578s" podCreationTimestamp="2025-12-01 16:15:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:15:19.235505117 +0000 UTC m=+6084.999014720" watchObservedRunningTime="2025-12-01 16:15:19.270539578 +0000 UTC m=+6085.034049191" Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.318105 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-llq2h" podStartSLOduration=3.318086779 podStartE2EDuration="3.318086779s" podCreationTimestamp="2025-12-01 16:15:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:15:19.273942991 +0000 UTC m=+6085.037452594" watchObservedRunningTime="2025-12-01 16:15:19.318086779 +0000 UTC m=+6085.081596372" Dec 01 16:15:19 crc kubenswrapper[4810]: I1201 16:15:19.352826 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.352802411 podStartE2EDuration="2.352802411s" podCreationTimestamp="2025-12-01 16:15:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:15:19.291406614 +0000 UTC m=+6085.054916227" watchObservedRunningTime="2025-12-01 16:15:19.352802411 +0000 UTC m=+6085.116312014" Dec 01 16:15:20 crc kubenswrapper[4810]: I1201 16:15:20.200758 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" event={"ID":"59e48d6b-851f-4977-bb5e-9720f697ef6c","Type":"ContainerStarted","Data":"a1543b61db7c6141aeb0e45fa2812f91cf44f55889a9e387463962a25656553d"} Dec 01 16:15:20 crc kubenswrapper[4810]: I1201 16:15:20.202370 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:20 crc kubenswrapper[4810]: I1201 16:15:20.203640 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2m2vb" event={"ID":"95f99947-fa7b-43af-9d9d-640e6102771d","Type":"ContainerStarted","Data":"2110fa6f8420d02721840cf754957f5be6b1e0af350c6d0523a83fddf56d0682"} Dec 01 16:15:20 crc kubenswrapper[4810]: I1201 16:15:20.203684 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2m2vb" event={"ID":"95f99947-fa7b-43af-9d9d-640e6102771d","Type":"ContainerStarted","Data":"4cb01e23607913b147af8e7d68c72334509fd659914529f7244ff6842043a457"} Dec 01 16:15:20 crc kubenswrapper[4810]: I1201 16:15:20.240630 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" podStartSLOduration=3.240612543 podStartE2EDuration="3.240612543s" podCreationTimestamp="2025-12-01 16:15:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:15:20.235991337 +0000 UTC m=+6085.999500940" watchObservedRunningTime="2025-12-01 16:15:20.240612543 +0000 UTC m=+6086.004122146" Dec 01 16:15:20 crc kubenswrapper[4810]: I1201 16:15:20.265994 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-2m2vb" podStartSLOduration=2.265974581 podStartE2EDuration="2.265974581s" podCreationTimestamp="2025-12-01 16:15:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:15:20.25780461 +0000 UTC m=+6086.021314213" watchObservedRunningTime="2025-12-01 16:15:20.265974581 +0000 UTC m=+6086.029484174" Dec 01 16:15:21 crc kubenswrapper[4810]: I1201 16:15:21.922244 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 16:15:21 crc kubenswrapper[4810]: I1201 16:15:21.922494 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="03364899-f3a3-44cc-8514-78257dbd82d3" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2c40e25e48567b25c4495e3bf43d30a0cbbab7e1fa3b6c755dbdd915487db172" gracePeriod=30 Dec 01 16:15:21 crc kubenswrapper[4810]: I1201 16:15:21.951795 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:21 crc kubenswrapper[4810]: I1201 16:15:21.952006 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2dc35829-d42b-4852-8bb7-f3ba752409da" containerName="nova-metadata-log" containerID="cri-o://14f3e3008cb05ebaae3fe14ef15377e71981ff3bc97ab4d07a9b1f1ece2875c4" gracePeriod=30 Dec 01 16:15:21 crc kubenswrapper[4810]: I1201 16:15:21.952292 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2dc35829-d42b-4852-8bb7-f3ba752409da" containerName="nova-metadata-metadata" containerID="cri-o://607000e7ede7cba1ab8433636ec44e3895c8381aed743313c1cc08662957e3ef" gracePeriod=30 Dec 01 16:15:22 crc kubenswrapper[4810]: I1201 16:15:22.227345 4810 generic.go:334] "Generic (PLEG): container finished" podID="2dc35829-d42b-4852-8bb7-f3ba752409da" containerID="14f3e3008cb05ebaae3fe14ef15377e71981ff3bc97ab4d07a9b1f1ece2875c4" exitCode=143 Dec 01 16:15:22 crc kubenswrapper[4810]: I1201 16:15:22.227441 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2dc35829-d42b-4852-8bb7-f3ba752409da","Type":"ContainerDied","Data":"14f3e3008cb05ebaae3fe14ef15377e71981ff3bc97ab4d07a9b1f1ece2875c4"} Dec 01 16:15:22 crc kubenswrapper[4810]: I1201 16:15:22.386573 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 16:15:22 crc kubenswrapper[4810]: I1201 16:15:22.386649 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 16:15:22 crc kubenswrapper[4810]: I1201 16:15:22.695802 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 16:15:22 crc kubenswrapper[4810]: I1201 16:15:22.723711 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.092812 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.099046 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.236599 4810 generic.go:334] "Generic (PLEG): container finished" podID="03364899-f3a3-44cc-8514-78257dbd82d3" containerID="2c40e25e48567b25c4495e3bf43d30a0cbbab7e1fa3b6c755dbdd915487db172" exitCode=0 Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.236649 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.236666 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"03364899-f3a3-44cc-8514-78257dbd82d3","Type":"ContainerDied","Data":"2c40e25e48567b25c4495e3bf43d30a0cbbab7e1fa3b6c755dbdd915487db172"} Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.237056 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"03364899-f3a3-44cc-8514-78257dbd82d3","Type":"ContainerDied","Data":"77c6b2d3a5eb73b86bfef2426f975afc2e31253ad88229bfebd8591eb242aba7"} Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.237107 4810 scope.go:117] "RemoveContainer" containerID="2c40e25e48567b25c4495e3bf43d30a0cbbab7e1fa3b6c755dbdd915487db172" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.239012 4810 generic.go:334] "Generic (PLEG): container finished" podID="95f99947-fa7b-43af-9d9d-640e6102771d" containerID="2110fa6f8420d02721840cf754957f5be6b1e0af350c6d0523a83fddf56d0682" exitCode=0 Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.239077 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2m2vb" event={"ID":"95f99947-fa7b-43af-9d9d-640e6102771d","Type":"ContainerDied","Data":"2110fa6f8420d02721840cf754957f5be6b1e0af350c6d0523a83fddf56d0682"} Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.243257 4810 generic.go:334] "Generic (PLEG): container finished" podID="2dc35829-d42b-4852-8bb7-f3ba752409da" containerID="607000e7ede7cba1ab8433636ec44e3895c8381aed743313c1cc08662957e3ef" exitCode=0 Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.243302 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2dc35829-d42b-4852-8bb7-f3ba752409da","Type":"ContainerDied","Data":"607000e7ede7cba1ab8433636ec44e3895c8381aed743313c1cc08662957e3ef"} Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.243348 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2dc35829-d42b-4852-8bb7-f3ba752409da","Type":"ContainerDied","Data":"5f366d8cd6edfa337e4d2830652cfbf57be9bab9bc89820849b681401d2c9c12"} Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.243309 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.287196 4810 scope.go:117] "RemoveContainer" containerID="2c40e25e48567b25c4495e3bf43d30a0cbbab7e1fa3b6c755dbdd915487db172" Dec 01 16:15:23 crc kubenswrapper[4810]: E1201 16:15:23.287680 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c40e25e48567b25c4495e3bf43d30a0cbbab7e1fa3b6c755dbdd915487db172\": container with ID starting with 2c40e25e48567b25c4495e3bf43d30a0cbbab7e1fa3b6c755dbdd915487db172 not found: ID does not exist" containerID="2c40e25e48567b25c4495e3bf43d30a0cbbab7e1fa3b6c755dbdd915487db172" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.287725 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c40e25e48567b25c4495e3bf43d30a0cbbab7e1fa3b6c755dbdd915487db172"} err="failed to get container status \"2c40e25e48567b25c4495e3bf43d30a0cbbab7e1fa3b6c755dbdd915487db172\": rpc error: code = NotFound desc = could not find container \"2c40e25e48567b25c4495e3bf43d30a0cbbab7e1fa3b6c755dbdd915487db172\": container with ID starting with 2c40e25e48567b25c4495e3bf43d30a0cbbab7e1fa3b6c755dbdd915487db172 not found: ID does not exist" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.287792 4810 scope.go:117] "RemoveContainer" containerID="607000e7ede7cba1ab8433636ec44e3895c8381aed743313c1cc08662957e3ef" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.292253 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dc35829-d42b-4852-8bb7-f3ba752409da-combined-ca-bundle\") pod \"2dc35829-d42b-4852-8bb7-f3ba752409da\" (UID: \"2dc35829-d42b-4852-8bb7-f3ba752409da\") " Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.292314 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dc35829-d42b-4852-8bb7-f3ba752409da-logs\") pod \"2dc35829-d42b-4852-8bb7-f3ba752409da\" (UID: \"2dc35829-d42b-4852-8bb7-f3ba752409da\") " Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.292352 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8b8v5\" (UniqueName: \"kubernetes.io/projected/03364899-f3a3-44cc-8514-78257dbd82d3-kube-api-access-8b8v5\") pod \"03364899-f3a3-44cc-8514-78257dbd82d3\" (UID: \"03364899-f3a3-44cc-8514-78257dbd82d3\") " Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.292437 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03364899-f3a3-44cc-8514-78257dbd82d3-config-data\") pod \"03364899-f3a3-44cc-8514-78257dbd82d3\" (UID: \"03364899-f3a3-44cc-8514-78257dbd82d3\") " Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.292551 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dc35829-d42b-4852-8bb7-f3ba752409da-config-data\") pod \"2dc35829-d42b-4852-8bb7-f3ba752409da\" (UID: \"2dc35829-d42b-4852-8bb7-f3ba752409da\") " Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.292631 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03364899-f3a3-44cc-8514-78257dbd82d3-combined-ca-bundle\") pod \"03364899-f3a3-44cc-8514-78257dbd82d3\" (UID: \"03364899-f3a3-44cc-8514-78257dbd82d3\") " Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.292709 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8rv6\" (UniqueName: \"kubernetes.io/projected/2dc35829-d42b-4852-8bb7-f3ba752409da-kube-api-access-g8rv6\") pod \"2dc35829-d42b-4852-8bb7-f3ba752409da\" (UID: \"2dc35829-d42b-4852-8bb7-f3ba752409da\") " Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.293853 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dc35829-d42b-4852-8bb7-f3ba752409da-logs" (OuterVolumeSpecName: "logs") pod "2dc35829-d42b-4852-8bb7-f3ba752409da" (UID: "2dc35829-d42b-4852-8bb7-f3ba752409da"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.301680 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dc35829-d42b-4852-8bb7-f3ba752409da-kube-api-access-g8rv6" (OuterVolumeSpecName: "kube-api-access-g8rv6") pod "2dc35829-d42b-4852-8bb7-f3ba752409da" (UID: "2dc35829-d42b-4852-8bb7-f3ba752409da"). InnerVolumeSpecName "kube-api-access-g8rv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.307959 4810 scope.go:117] "RemoveContainer" containerID="14f3e3008cb05ebaae3fe14ef15377e71981ff3bc97ab4d07a9b1f1ece2875c4" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.310713 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03364899-f3a3-44cc-8514-78257dbd82d3-kube-api-access-8b8v5" (OuterVolumeSpecName: "kube-api-access-8b8v5") pod "03364899-f3a3-44cc-8514-78257dbd82d3" (UID: "03364899-f3a3-44cc-8514-78257dbd82d3"). InnerVolumeSpecName "kube-api-access-8b8v5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.322452 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03364899-f3a3-44cc-8514-78257dbd82d3-config-data" (OuterVolumeSpecName: "config-data") pod "03364899-f3a3-44cc-8514-78257dbd82d3" (UID: "03364899-f3a3-44cc-8514-78257dbd82d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.324661 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03364899-f3a3-44cc-8514-78257dbd82d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03364899-f3a3-44cc-8514-78257dbd82d3" (UID: "03364899-f3a3-44cc-8514-78257dbd82d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.337742 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dc35829-d42b-4852-8bb7-f3ba752409da-config-data" (OuterVolumeSpecName: "config-data") pod "2dc35829-d42b-4852-8bb7-f3ba752409da" (UID: "2dc35829-d42b-4852-8bb7-f3ba752409da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.353625 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2dc35829-d42b-4852-8bb7-f3ba752409da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2dc35829-d42b-4852-8bb7-f3ba752409da" (UID: "2dc35829-d42b-4852-8bb7-f3ba752409da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.394301 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03364899-f3a3-44cc-8514-78257dbd82d3-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.395108 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dc35829-d42b-4852-8bb7-f3ba752409da-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.395178 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03364899-f3a3-44cc-8514-78257dbd82d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.395235 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8rv6\" (UniqueName: \"kubernetes.io/projected/2dc35829-d42b-4852-8bb7-f3ba752409da-kube-api-access-g8rv6\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.395348 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dc35829-d42b-4852-8bb7-f3ba752409da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.395403 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dc35829-d42b-4852-8bb7-f3ba752409da-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.395459 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8b8v5\" (UniqueName: \"kubernetes.io/projected/03364899-f3a3-44cc-8514-78257dbd82d3-kube-api-access-8b8v5\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.427143 4810 scope.go:117] "RemoveContainer" containerID="607000e7ede7cba1ab8433636ec44e3895c8381aed743313c1cc08662957e3ef" Dec 01 16:15:23 crc kubenswrapper[4810]: E1201 16:15:23.427551 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"607000e7ede7cba1ab8433636ec44e3895c8381aed743313c1cc08662957e3ef\": container with ID starting with 607000e7ede7cba1ab8433636ec44e3895c8381aed743313c1cc08662957e3ef not found: ID does not exist" containerID="607000e7ede7cba1ab8433636ec44e3895c8381aed743313c1cc08662957e3ef" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.427587 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"607000e7ede7cba1ab8433636ec44e3895c8381aed743313c1cc08662957e3ef"} err="failed to get container status \"607000e7ede7cba1ab8433636ec44e3895c8381aed743313c1cc08662957e3ef\": rpc error: code = NotFound desc = could not find container \"607000e7ede7cba1ab8433636ec44e3895c8381aed743313c1cc08662957e3ef\": container with ID starting with 607000e7ede7cba1ab8433636ec44e3895c8381aed743313c1cc08662957e3ef not found: ID does not exist" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.427607 4810 scope.go:117] "RemoveContainer" containerID="14f3e3008cb05ebaae3fe14ef15377e71981ff3bc97ab4d07a9b1f1ece2875c4" Dec 01 16:15:23 crc kubenswrapper[4810]: E1201 16:15:23.428136 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14f3e3008cb05ebaae3fe14ef15377e71981ff3bc97ab4d07a9b1f1ece2875c4\": container with ID starting with 14f3e3008cb05ebaae3fe14ef15377e71981ff3bc97ab4d07a9b1f1ece2875c4 not found: ID does not exist" containerID="14f3e3008cb05ebaae3fe14ef15377e71981ff3bc97ab4d07a9b1f1ece2875c4" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.428158 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14f3e3008cb05ebaae3fe14ef15377e71981ff3bc97ab4d07a9b1f1ece2875c4"} err="failed to get container status \"14f3e3008cb05ebaae3fe14ef15377e71981ff3bc97ab4d07a9b1f1ece2875c4\": rpc error: code = NotFound desc = could not find container \"14f3e3008cb05ebaae3fe14ef15377e71981ff3bc97ab4d07a9b1f1ece2875c4\": container with ID starting with 14f3e3008cb05ebaae3fe14ef15377e71981ff3bc97ab4d07a9b1f1ece2875c4 not found: ID does not exist" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.580633 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.626296 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.637771 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.658303 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.668019 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 16:15:23 crc kubenswrapper[4810]: E1201 16:15:23.668669 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dc35829-d42b-4852-8bb7-f3ba752409da" containerName="nova-metadata-log" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.668691 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dc35829-d42b-4852-8bb7-f3ba752409da" containerName="nova-metadata-log" Dec 01 16:15:23 crc kubenswrapper[4810]: E1201 16:15:23.668702 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dc35829-d42b-4852-8bb7-f3ba752409da" containerName="nova-metadata-metadata" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.668709 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dc35829-d42b-4852-8bb7-f3ba752409da" containerName="nova-metadata-metadata" Dec 01 16:15:23 crc kubenswrapper[4810]: E1201 16:15:23.668727 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03364899-f3a3-44cc-8514-78257dbd82d3" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.668733 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="03364899-f3a3-44cc-8514-78257dbd82d3" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.668902 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="03364899-f3a3-44cc-8514-78257dbd82d3" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.668928 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dc35829-d42b-4852-8bb7-f3ba752409da" containerName="nova-metadata-log" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.668943 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dc35829-d42b-4852-8bb7-f3ba752409da" containerName="nova-metadata-metadata" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.669630 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.671766 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.671938 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.672125 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.685663 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.695712 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.697667 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.700050 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.700050 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.720156 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.801668 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf4e50ff-bd60-4fef-a1dd-b539720329ba-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"cf4e50ff-bd60-4fef-a1dd-b539720329ba\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.801714 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/087764ab-72da-4656-93c4-cd86bcd386a9-logs\") pod \"nova-metadata-0\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " pod="openstack/nova-metadata-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.801740 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/087764ab-72da-4656-93c4-cd86bcd386a9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " pod="openstack/nova-metadata-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.801765 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087764ab-72da-4656-93c4-cd86bcd386a9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " pod="openstack/nova-metadata-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.801849 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf4e50ff-bd60-4fef-a1dd-b539720329ba-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cf4e50ff-bd60-4fef-a1dd-b539720329ba\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.801866 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf4e50ff-bd60-4fef-a1dd-b539720329ba-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cf4e50ff-bd60-4fef-a1dd-b539720329ba\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.801895 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pphdt\" (UniqueName: \"kubernetes.io/projected/087764ab-72da-4656-93c4-cd86bcd386a9-kube-api-access-pphdt\") pod \"nova-metadata-0\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " pod="openstack/nova-metadata-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.801915 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wvmc\" (UniqueName: \"kubernetes.io/projected/cf4e50ff-bd60-4fef-a1dd-b539720329ba-kube-api-access-4wvmc\") pod \"nova-cell1-novncproxy-0\" (UID: \"cf4e50ff-bd60-4fef-a1dd-b539720329ba\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.801934 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf4e50ff-bd60-4fef-a1dd-b539720329ba-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"cf4e50ff-bd60-4fef-a1dd-b539720329ba\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.801955 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/087764ab-72da-4656-93c4-cd86bcd386a9-config-data\") pod \"nova-metadata-0\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " pod="openstack/nova-metadata-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.903358 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf4e50ff-bd60-4fef-a1dd-b539720329ba-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cf4e50ff-bd60-4fef-a1dd-b539720329ba\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.903409 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf4e50ff-bd60-4fef-a1dd-b539720329ba-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cf4e50ff-bd60-4fef-a1dd-b539720329ba\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.903445 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pphdt\" (UniqueName: \"kubernetes.io/projected/087764ab-72da-4656-93c4-cd86bcd386a9-kube-api-access-pphdt\") pod \"nova-metadata-0\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " pod="openstack/nova-metadata-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.903501 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wvmc\" (UniqueName: \"kubernetes.io/projected/cf4e50ff-bd60-4fef-a1dd-b539720329ba-kube-api-access-4wvmc\") pod \"nova-cell1-novncproxy-0\" (UID: \"cf4e50ff-bd60-4fef-a1dd-b539720329ba\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.903521 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf4e50ff-bd60-4fef-a1dd-b539720329ba-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"cf4e50ff-bd60-4fef-a1dd-b539720329ba\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.903542 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/087764ab-72da-4656-93c4-cd86bcd386a9-config-data\") pod \"nova-metadata-0\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " pod="openstack/nova-metadata-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.903601 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf4e50ff-bd60-4fef-a1dd-b539720329ba-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"cf4e50ff-bd60-4fef-a1dd-b539720329ba\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.903616 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/087764ab-72da-4656-93c4-cd86bcd386a9-logs\") pod \"nova-metadata-0\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " pod="openstack/nova-metadata-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.903635 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/087764ab-72da-4656-93c4-cd86bcd386a9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " pod="openstack/nova-metadata-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.903660 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087764ab-72da-4656-93c4-cd86bcd386a9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " pod="openstack/nova-metadata-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.906375 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/087764ab-72da-4656-93c4-cd86bcd386a9-logs\") pod \"nova-metadata-0\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " pod="openstack/nova-metadata-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.908516 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087764ab-72da-4656-93c4-cd86bcd386a9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " pod="openstack/nova-metadata-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.909513 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf4e50ff-bd60-4fef-a1dd-b539720329ba-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"cf4e50ff-bd60-4fef-a1dd-b539720329ba\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.909513 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf4e50ff-bd60-4fef-a1dd-b539720329ba-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"cf4e50ff-bd60-4fef-a1dd-b539720329ba\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.909902 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/087764ab-72da-4656-93c4-cd86bcd386a9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " pod="openstack/nova-metadata-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.910812 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/087764ab-72da-4656-93c4-cd86bcd386a9-config-data\") pod \"nova-metadata-0\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " pod="openstack/nova-metadata-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.918949 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf4e50ff-bd60-4fef-a1dd-b539720329ba-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cf4e50ff-bd60-4fef-a1dd-b539720329ba\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.919525 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf4e50ff-bd60-4fef-a1dd-b539720329ba-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cf4e50ff-bd60-4fef-a1dd-b539720329ba\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.922364 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wvmc\" (UniqueName: \"kubernetes.io/projected/cf4e50ff-bd60-4fef-a1dd-b539720329ba-kube-api-access-4wvmc\") pod \"nova-cell1-novncproxy-0\" (UID: \"cf4e50ff-bd60-4fef-a1dd-b539720329ba\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.923291 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pphdt\" (UniqueName: \"kubernetes.io/projected/087764ab-72da-4656-93c4-cd86bcd386a9-kube-api-access-pphdt\") pod \"nova-metadata-0\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " pod="openstack/nova-metadata-0" Dec 01 16:15:23 crc kubenswrapper[4810]: I1201 16:15:23.997222 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:24 crc kubenswrapper[4810]: I1201 16:15:24.023991 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 16:15:24 crc kubenswrapper[4810]: I1201 16:15:24.509406 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03364899-f3a3-44cc-8514-78257dbd82d3" path="/var/lib/kubelet/pods/03364899-f3a3-44cc-8514-78257dbd82d3/volumes" Dec 01 16:15:24 crc kubenswrapper[4810]: I1201 16:15:24.510531 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dc35829-d42b-4852-8bb7-f3ba752409da" path="/var/lib/kubelet/pods/2dc35829-d42b-4852-8bb7-f3ba752409da/volumes" Dec 01 16:15:24 crc kubenswrapper[4810]: I1201 16:15:24.517545 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 16:15:24 crc kubenswrapper[4810]: I1201 16:15:24.617286 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:24 crc kubenswrapper[4810]: W1201 16:15:24.620708 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod087764ab_72da_4656_93c4_cd86bcd386a9.slice/crio-b964bc747c23b56d3c70cff8d77a4ad50542ef70fa66848ced8d43e086985727 WatchSource:0}: Error finding container b964bc747c23b56d3c70cff8d77a4ad50542ef70fa66848ced8d43e086985727: Status 404 returned error can't find the container with id b964bc747c23b56d3c70cff8d77a4ad50542ef70fa66848ced8d43e086985727 Dec 01 16:15:24 crc kubenswrapper[4810]: I1201 16:15:24.899275 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2m2vb" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.027401 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8h8sw\" (UniqueName: \"kubernetes.io/projected/95f99947-fa7b-43af-9d9d-640e6102771d-kube-api-access-8h8sw\") pod \"95f99947-fa7b-43af-9d9d-640e6102771d\" (UID: \"95f99947-fa7b-43af-9d9d-640e6102771d\") " Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.027495 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95f99947-fa7b-43af-9d9d-640e6102771d-scripts\") pod \"95f99947-fa7b-43af-9d9d-640e6102771d\" (UID: \"95f99947-fa7b-43af-9d9d-640e6102771d\") " Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.027543 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f99947-fa7b-43af-9d9d-640e6102771d-config-data\") pod \"95f99947-fa7b-43af-9d9d-640e6102771d\" (UID: \"95f99947-fa7b-43af-9d9d-640e6102771d\") " Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.027617 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f99947-fa7b-43af-9d9d-640e6102771d-combined-ca-bundle\") pod \"95f99947-fa7b-43af-9d9d-640e6102771d\" (UID: \"95f99947-fa7b-43af-9d9d-640e6102771d\") " Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.031007 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95f99947-fa7b-43af-9d9d-640e6102771d-kube-api-access-8h8sw" (OuterVolumeSpecName: "kube-api-access-8h8sw") pod "95f99947-fa7b-43af-9d9d-640e6102771d" (UID: "95f99947-fa7b-43af-9d9d-640e6102771d"). InnerVolumeSpecName "kube-api-access-8h8sw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.031328 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f99947-fa7b-43af-9d9d-640e6102771d-scripts" (OuterVolumeSpecName: "scripts") pod "95f99947-fa7b-43af-9d9d-640e6102771d" (UID: "95f99947-fa7b-43af-9d9d-640e6102771d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.052894 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f99947-fa7b-43af-9d9d-640e6102771d-config-data" (OuterVolumeSpecName: "config-data") pod "95f99947-fa7b-43af-9d9d-640e6102771d" (UID: "95f99947-fa7b-43af-9d9d-640e6102771d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.056200 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95f99947-fa7b-43af-9d9d-640e6102771d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95f99947-fa7b-43af-9d9d-640e6102771d" (UID: "95f99947-fa7b-43af-9d9d-640e6102771d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.129981 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8h8sw\" (UniqueName: \"kubernetes.io/projected/95f99947-fa7b-43af-9d9d-640e6102771d-kube-api-access-8h8sw\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.130198 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95f99947-fa7b-43af-9d9d-640e6102771d-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.130284 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95f99947-fa7b-43af-9d9d-640e6102771d-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.130431 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95f99947-fa7b-43af-9d9d-640e6102771d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.278933 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"087764ab-72da-4656-93c4-cd86bcd386a9","Type":"ContainerStarted","Data":"cc1fc6af3b6ee08876afd501c686a2ab238d77d7488ca447ee3307f947912200"} Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.279212 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"087764ab-72da-4656-93c4-cd86bcd386a9","Type":"ContainerStarted","Data":"ee034e98de7c9caf7a0d2d2e952f7ef74534066953ec461ad0f3fd2359a9a47c"} Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.279329 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"087764ab-72da-4656-93c4-cd86bcd386a9","Type":"ContainerStarted","Data":"b964bc747c23b56d3c70cff8d77a4ad50542ef70fa66848ced8d43e086985727"} Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.284406 4810 generic.go:334] "Generic (PLEG): container finished" podID="37eef714-f1dc-47b3-8489-5204d08789bb" containerID="ff0b3f547bd0fe4827488b9725fc4957403a6bc1f94a84d700e80db21c809c1e" exitCode=0 Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.284505 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-llq2h" event={"ID":"37eef714-f1dc-47b3-8489-5204d08789bb","Type":"ContainerDied","Data":"ff0b3f547bd0fe4827488b9725fc4957403a6bc1f94a84d700e80db21c809c1e"} Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.290952 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-2m2vb" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.290999 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-2m2vb" event={"ID":"95f99947-fa7b-43af-9d9d-640e6102771d","Type":"ContainerDied","Data":"4cb01e23607913b147af8e7d68c72334509fd659914529f7244ff6842043a457"} Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.291038 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4cb01e23607913b147af8e7d68c72334509fd659914529f7244ff6842043a457" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.293819 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cf4e50ff-bd60-4fef-a1dd-b539720329ba","Type":"ContainerStarted","Data":"e4ba51ec2d35024f73ec2e11c7ef89e62f3894a51592ca5331bfa66a90e22fc6"} Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.293865 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cf4e50ff-bd60-4fef-a1dd-b539720329ba","Type":"ContainerStarted","Data":"b6bc4e82f0230dc1dc01653df8ad81f012567b59e717f73481684f84df0cff28"} Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.300683 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.300666887 podStartE2EDuration="2.300666887s" podCreationTimestamp="2025-12-01 16:15:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:15:25.296448992 +0000 UTC m=+6091.059958595" watchObservedRunningTime="2025-12-01 16:15:25.300666887 +0000 UTC m=+6091.064176490" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.323359 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.323343362 podStartE2EDuration="2.323343362s" podCreationTimestamp="2025-12-01 16:15:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:15:25.320494465 +0000 UTC m=+6091.084004068" watchObservedRunningTime="2025-12-01 16:15:25.323343362 +0000 UTC m=+6091.086852965" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.341429 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 16:15:25 crc kubenswrapper[4810]: E1201 16:15:25.343632 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95f99947-fa7b-43af-9d9d-640e6102771d" containerName="nova-cell1-conductor-db-sync" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.343665 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="95f99947-fa7b-43af-9d9d-640e6102771d" containerName="nova-cell1-conductor-db-sync" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.343945 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="95f99947-fa7b-43af-9d9d-640e6102771d" containerName="nova-cell1-conductor-db-sync" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.345250 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.347806 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.367531 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.437062 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9cqm\" (UniqueName: \"kubernetes.io/projected/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3-kube-api-access-d9cqm\") pod \"nova-cell1-conductor-0\" (UID: \"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3\") " pod="openstack/nova-cell1-conductor-0" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.437116 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3\") " pod="openstack/nova-cell1-conductor-0" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.437150 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3\") " pod="openstack/nova-cell1-conductor-0" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.538630 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9cqm\" (UniqueName: \"kubernetes.io/projected/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3-kube-api-access-d9cqm\") pod \"nova-cell1-conductor-0\" (UID: \"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3\") " pod="openstack/nova-cell1-conductor-0" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.538698 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3\") " pod="openstack/nova-cell1-conductor-0" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.538769 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3\") " pod="openstack/nova-cell1-conductor-0" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.548373 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3\") " pod="openstack/nova-cell1-conductor-0" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.551230 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3\") " pod="openstack/nova-cell1-conductor-0" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.588075 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9cqm\" (UniqueName: \"kubernetes.io/projected/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3-kube-api-access-d9cqm\") pod \"nova-cell1-conductor-0\" (UID: \"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3\") " pod="openstack/nova-cell1-conductor-0" Dec 01 16:15:25 crc kubenswrapper[4810]: I1201 16:15:25.701840 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 16:15:26 crc kubenswrapper[4810]: I1201 16:15:26.170230 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 16:15:26 crc kubenswrapper[4810]: W1201 16:15:26.174955 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca965e9d_7a9d_4351_bfa1_36d77a9af2a3.slice/crio-85e8754a6969f138363e81f01c3e488808ce08dacd0739daf10c8f079869e0ba WatchSource:0}: Error finding container 85e8754a6969f138363e81f01c3e488808ce08dacd0739daf10c8f079869e0ba: Status 404 returned error can't find the container with id 85e8754a6969f138363e81f01c3e488808ce08dacd0739daf10c8f079869e0ba Dec 01 16:15:26 crc kubenswrapper[4810]: I1201 16:15:26.303311 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3","Type":"ContainerStarted","Data":"85e8754a6969f138363e81f01c3e488808ce08dacd0739daf10c8f079869e0ba"} Dec 01 16:15:26 crc kubenswrapper[4810]: I1201 16:15:26.642839 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-llq2h" Dec 01 16:15:26 crc kubenswrapper[4810]: I1201 16:15:26.758672 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37eef714-f1dc-47b3-8489-5204d08789bb-scripts\") pod \"37eef714-f1dc-47b3-8489-5204d08789bb\" (UID: \"37eef714-f1dc-47b3-8489-5204d08789bb\") " Dec 01 16:15:26 crc kubenswrapper[4810]: I1201 16:15:26.758769 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf7g2\" (UniqueName: \"kubernetes.io/projected/37eef714-f1dc-47b3-8489-5204d08789bb-kube-api-access-rf7g2\") pod \"37eef714-f1dc-47b3-8489-5204d08789bb\" (UID: \"37eef714-f1dc-47b3-8489-5204d08789bb\") " Dec 01 16:15:26 crc kubenswrapper[4810]: I1201 16:15:26.758882 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37eef714-f1dc-47b3-8489-5204d08789bb-combined-ca-bundle\") pod \"37eef714-f1dc-47b3-8489-5204d08789bb\" (UID: \"37eef714-f1dc-47b3-8489-5204d08789bb\") " Dec 01 16:15:26 crc kubenswrapper[4810]: I1201 16:15:26.758972 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37eef714-f1dc-47b3-8489-5204d08789bb-config-data\") pod \"37eef714-f1dc-47b3-8489-5204d08789bb\" (UID: \"37eef714-f1dc-47b3-8489-5204d08789bb\") " Dec 01 16:15:26 crc kubenswrapper[4810]: I1201 16:15:26.763406 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37eef714-f1dc-47b3-8489-5204d08789bb-scripts" (OuterVolumeSpecName: "scripts") pod "37eef714-f1dc-47b3-8489-5204d08789bb" (UID: "37eef714-f1dc-47b3-8489-5204d08789bb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:26 crc kubenswrapper[4810]: I1201 16:15:26.764817 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37eef714-f1dc-47b3-8489-5204d08789bb-kube-api-access-rf7g2" (OuterVolumeSpecName: "kube-api-access-rf7g2") pod "37eef714-f1dc-47b3-8489-5204d08789bb" (UID: "37eef714-f1dc-47b3-8489-5204d08789bb"). InnerVolumeSpecName "kube-api-access-rf7g2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:26 crc kubenswrapper[4810]: I1201 16:15:26.793199 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37eef714-f1dc-47b3-8489-5204d08789bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37eef714-f1dc-47b3-8489-5204d08789bb" (UID: "37eef714-f1dc-47b3-8489-5204d08789bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:26 crc kubenswrapper[4810]: I1201 16:15:26.806634 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37eef714-f1dc-47b3-8489-5204d08789bb-config-data" (OuterVolumeSpecName: "config-data") pod "37eef714-f1dc-47b3-8489-5204d08789bb" (UID: "37eef714-f1dc-47b3-8489-5204d08789bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:26 crc kubenswrapper[4810]: I1201 16:15:26.861190 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37eef714-f1dc-47b3-8489-5204d08789bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:26 crc kubenswrapper[4810]: I1201 16:15:26.861238 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37eef714-f1dc-47b3-8489-5204d08789bb-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:26 crc kubenswrapper[4810]: I1201 16:15:26.861247 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37eef714-f1dc-47b3-8489-5204d08789bb-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:26 crc kubenswrapper[4810]: I1201 16:15:26.861256 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf7g2\" (UniqueName: \"kubernetes.io/projected/37eef714-f1dc-47b3-8489-5204d08789bb-kube-api-access-rf7g2\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:27 crc kubenswrapper[4810]: I1201 16:15:27.314160 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-llq2h" event={"ID":"37eef714-f1dc-47b3-8489-5204d08789bb","Type":"ContainerDied","Data":"fb75d8b4d0244d16fe2b761e5030de745df5e1c509429fd4b9ba875e3d0b2576"} Dec 01 16:15:27 crc kubenswrapper[4810]: I1201 16:15:27.314210 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb75d8b4d0244d16fe2b761e5030de745df5e1c509429fd4b9ba875e3d0b2576" Dec 01 16:15:27 crc kubenswrapper[4810]: I1201 16:15:27.314288 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-llq2h" Dec 01 16:15:27 crc kubenswrapper[4810]: I1201 16:15:27.325319 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 16:15:27 crc kubenswrapper[4810]: I1201 16:15:27.325382 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 16:15:27 crc kubenswrapper[4810]: I1201 16:15:27.325538 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3","Type":"ContainerStarted","Data":"13f99bc5259f34633b3dc899ea67dcc6b4654616360f83abebaf1504118d2bb4"} Dec 01 16:15:27 crc kubenswrapper[4810]: I1201 16:15:27.326366 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 01 16:15:27 crc kubenswrapper[4810]: I1201 16:15:27.355055 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.355035027 podStartE2EDuration="2.355035027s" podCreationTimestamp="2025-12-01 16:15:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:15:27.346732972 +0000 UTC m=+6093.110242575" watchObservedRunningTime="2025-12-01 16:15:27.355035027 +0000 UTC m=+6093.118544630" Dec 01 16:15:27 crc kubenswrapper[4810]: I1201 16:15:27.541004 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 16:15:27 crc kubenswrapper[4810]: I1201 16:15:27.581162 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 16:15:27 crc kubenswrapper[4810]: I1201 16:15:27.581411 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="febfc322-9d2b-470b-bfdf-f7ee31996fc8" containerName="nova-scheduler-scheduler" containerID="cri-o://6a2eec563994f549af7272b80ded697cd4379e34fc30dd094901d08ad0f9408c" gracePeriod=30 Dec 01 16:15:27 crc kubenswrapper[4810]: I1201 16:15:27.590335 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:27 crc kubenswrapper[4810]: I1201 16:15:27.590529 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="087764ab-72da-4656-93c4-cd86bcd386a9" containerName="nova-metadata-log" containerID="cri-o://ee034e98de7c9caf7a0d2d2e952f7ef74534066953ec461ad0f3fd2359a9a47c" gracePeriod=30 Dec 01 16:15:27 crc kubenswrapper[4810]: I1201 16:15:27.590673 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="087764ab-72da-4656-93c4-cd86bcd386a9" containerName="nova-metadata-metadata" containerID="cri-o://cc1fc6af3b6ee08876afd501c686a2ab238d77d7488ca447ee3307f947912200" gracePeriod=30 Dec 01 16:15:27 crc kubenswrapper[4810]: I1201 16:15:27.714718 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:15:27 crc kubenswrapper[4810]: I1201 16:15:27.795762 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86b95c6b9-2jj57"] Dec 01 16:15:27 crc kubenswrapper[4810]: I1201 16:15:27.796308 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" podUID="bbd15d1f-1b0d-452d-8110-537140ccd88c" containerName="dnsmasq-dns" containerID="cri-o://429b2ac6c979ceff96755077a8398310e6de3258b8bbbb97f9f5436478aa9d9c" gracePeriod=10 Dec 01 16:15:28 crc kubenswrapper[4810]: E1201 16:15:28.165334 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbd15d1f_1b0d_452d_8110_537140ccd88c.slice/crio-429b2ac6c979ceff96755077a8398310e6de3258b8bbbb97f9f5436478aa9d9c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbd15d1f_1b0d_452d_8110_537140ccd88c.slice/crio-conmon-429b2ac6c979ceff96755077a8398310e6de3258b8bbbb97f9f5436478aa9d9c.scope\": RecentStats: unable to find data in memory cache]" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.300835 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.342352 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.343849 4810 generic.go:334] "Generic (PLEG): container finished" podID="087764ab-72da-4656-93c4-cd86bcd386a9" containerID="cc1fc6af3b6ee08876afd501c686a2ab238d77d7488ca447ee3307f947912200" exitCode=0 Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.343879 4810 generic.go:334] "Generic (PLEG): container finished" podID="087764ab-72da-4656-93c4-cd86bcd386a9" containerID="ee034e98de7c9caf7a0d2d2e952f7ef74534066953ec461ad0f3fd2359a9a47c" exitCode=143 Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.343915 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.343926 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"087764ab-72da-4656-93c4-cd86bcd386a9","Type":"ContainerDied","Data":"cc1fc6af3b6ee08876afd501c686a2ab238d77d7488ca447ee3307f947912200"} Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.343952 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"087764ab-72da-4656-93c4-cd86bcd386a9","Type":"ContainerDied","Data":"ee034e98de7c9caf7a0d2d2e952f7ef74534066953ec461ad0f3fd2359a9a47c"} Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.343964 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"087764ab-72da-4656-93c4-cd86bcd386a9","Type":"ContainerDied","Data":"b964bc747c23b56d3c70cff8d77a4ad50542ef70fa66848ced8d43e086985727"} Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.343982 4810 scope.go:117] "RemoveContainer" containerID="cc1fc6af3b6ee08876afd501c686a2ab238d77d7488ca447ee3307f947912200" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.354708 4810 generic.go:334] "Generic (PLEG): container finished" podID="bbd15d1f-1b0d-452d-8110-537140ccd88c" containerID="429b2ac6c979ceff96755077a8398310e6de3258b8bbbb97f9f5436478aa9d9c" exitCode=0 Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.355066 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.355130 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3fcbd109-53af-4ba8-9418-10466117a3d3" containerName="nova-api-log" containerID="cri-o://bc231bdd562c07a555c11a5594ba7d693ad7210f6b005e2f7d9065668eb5380a" gracePeriod=30 Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.355315 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3fcbd109-53af-4ba8-9418-10466117a3d3" containerName="nova-api-api" containerID="cri-o://a48b965be1fb027015443f27baee0ee5689d05f71ed3a69d979efc1f793ddf01" gracePeriod=30 Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.355603 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86b95c6b9-2jj57" event={"ID":"bbd15d1f-1b0d-452d-8110-537140ccd88c","Type":"ContainerDied","Data":"429b2ac6c979ceff96755077a8398310e6de3258b8bbbb97f9f5436478aa9d9c"} Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.383132 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3fcbd109-53af-4ba8-9418-10466117a3d3" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.77:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.383195 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3fcbd109-53af-4ba8-9418-10466117a3d3" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.77:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.401862 4810 scope.go:117] "RemoveContainer" containerID="ee034e98de7c9caf7a0d2d2e952f7ef74534066953ec461ad0f3fd2359a9a47c" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.411335 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/087764ab-72da-4656-93c4-cd86bcd386a9-nova-metadata-tls-certs\") pod \"087764ab-72da-4656-93c4-cd86bcd386a9\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.411406 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pphdt\" (UniqueName: \"kubernetes.io/projected/087764ab-72da-4656-93c4-cd86bcd386a9-kube-api-access-pphdt\") pod \"087764ab-72da-4656-93c4-cd86bcd386a9\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.411532 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-ovsdbserver-nb\") pod \"bbd15d1f-1b0d-452d-8110-537140ccd88c\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.411555 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-ovsdbserver-sb\") pod \"bbd15d1f-1b0d-452d-8110-537140ccd88c\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.411644 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087764ab-72da-4656-93c4-cd86bcd386a9-combined-ca-bundle\") pod \"087764ab-72da-4656-93c4-cd86bcd386a9\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.411669 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-config\") pod \"bbd15d1f-1b0d-452d-8110-537140ccd88c\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.411730 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhmr9\" (UniqueName: \"kubernetes.io/projected/bbd15d1f-1b0d-452d-8110-537140ccd88c-kube-api-access-lhmr9\") pod \"bbd15d1f-1b0d-452d-8110-537140ccd88c\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.411768 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-dns-svc\") pod \"bbd15d1f-1b0d-452d-8110-537140ccd88c\" (UID: \"bbd15d1f-1b0d-452d-8110-537140ccd88c\") " Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.411816 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/087764ab-72da-4656-93c4-cd86bcd386a9-config-data\") pod \"087764ab-72da-4656-93c4-cd86bcd386a9\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.411849 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/087764ab-72da-4656-93c4-cd86bcd386a9-logs\") pod \"087764ab-72da-4656-93c4-cd86bcd386a9\" (UID: \"087764ab-72da-4656-93c4-cd86bcd386a9\") " Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.415673 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/087764ab-72da-4656-93c4-cd86bcd386a9-logs" (OuterVolumeSpecName: "logs") pod "087764ab-72da-4656-93c4-cd86bcd386a9" (UID: "087764ab-72da-4656-93c4-cd86bcd386a9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.426935 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/087764ab-72da-4656-93c4-cd86bcd386a9-kube-api-access-pphdt" (OuterVolumeSpecName: "kube-api-access-pphdt") pod "087764ab-72da-4656-93c4-cd86bcd386a9" (UID: "087764ab-72da-4656-93c4-cd86bcd386a9"). InnerVolumeSpecName "kube-api-access-pphdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.433712 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbd15d1f-1b0d-452d-8110-537140ccd88c-kube-api-access-lhmr9" (OuterVolumeSpecName: "kube-api-access-lhmr9") pod "bbd15d1f-1b0d-452d-8110-537140ccd88c" (UID: "bbd15d1f-1b0d-452d-8110-537140ccd88c"). InnerVolumeSpecName "kube-api-access-lhmr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.441216 4810 scope.go:117] "RemoveContainer" containerID="cc1fc6af3b6ee08876afd501c686a2ab238d77d7488ca447ee3307f947912200" Dec 01 16:15:28 crc kubenswrapper[4810]: E1201 16:15:28.444626 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc1fc6af3b6ee08876afd501c686a2ab238d77d7488ca447ee3307f947912200\": container with ID starting with cc1fc6af3b6ee08876afd501c686a2ab238d77d7488ca447ee3307f947912200 not found: ID does not exist" containerID="cc1fc6af3b6ee08876afd501c686a2ab238d77d7488ca447ee3307f947912200" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.444668 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc1fc6af3b6ee08876afd501c686a2ab238d77d7488ca447ee3307f947912200"} err="failed to get container status \"cc1fc6af3b6ee08876afd501c686a2ab238d77d7488ca447ee3307f947912200\": rpc error: code = NotFound desc = could not find container \"cc1fc6af3b6ee08876afd501c686a2ab238d77d7488ca447ee3307f947912200\": container with ID starting with cc1fc6af3b6ee08876afd501c686a2ab238d77d7488ca447ee3307f947912200 not found: ID does not exist" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.444690 4810 scope.go:117] "RemoveContainer" containerID="ee034e98de7c9caf7a0d2d2e952f7ef74534066953ec461ad0f3fd2359a9a47c" Dec 01 16:15:28 crc kubenswrapper[4810]: E1201 16:15:28.445049 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee034e98de7c9caf7a0d2d2e952f7ef74534066953ec461ad0f3fd2359a9a47c\": container with ID starting with ee034e98de7c9caf7a0d2d2e952f7ef74534066953ec461ad0f3fd2359a9a47c not found: ID does not exist" containerID="ee034e98de7c9caf7a0d2d2e952f7ef74534066953ec461ad0f3fd2359a9a47c" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.445078 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee034e98de7c9caf7a0d2d2e952f7ef74534066953ec461ad0f3fd2359a9a47c"} err="failed to get container status \"ee034e98de7c9caf7a0d2d2e952f7ef74534066953ec461ad0f3fd2359a9a47c\": rpc error: code = NotFound desc = could not find container \"ee034e98de7c9caf7a0d2d2e952f7ef74534066953ec461ad0f3fd2359a9a47c\": container with ID starting with ee034e98de7c9caf7a0d2d2e952f7ef74534066953ec461ad0f3fd2359a9a47c not found: ID does not exist" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.445094 4810 scope.go:117] "RemoveContainer" containerID="cc1fc6af3b6ee08876afd501c686a2ab238d77d7488ca447ee3307f947912200" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.446795 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc1fc6af3b6ee08876afd501c686a2ab238d77d7488ca447ee3307f947912200"} err="failed to get container status \"cc1fc6af3b6ee08876afd501c686a2ab238d77d7488ca447ee3307f947912200\": rpc error: code = NotFound desc = could not find container \"cc1fc6af3b6ee08876afd501c686a2ab238d77d7488ca447ee3307f947912200\": container with ID starting with cc1fc6af3b6ee08876afd501c686a2ab238d77d7488ca447ee3307f947912200 not found: ID does not exist" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.446824 4810 scope.go:117] "RemoveContainer" containerID="ee034e98de7c9caf7a0d2d2e952f7ef74534066953ec461ad0f3fd2359a9a47c" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.447046 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee034e98de7c9caf7a0d2d2e952f7ef74534066953ec461ad0f3fd2359a9a47c"} err="failed to get container status \"ee034e98de7c9caf7a0d2d2e952f7ef74534066953ec461ad0f3fd2359a9a47c\": rpc error: code = NotFound desc = could not find container \"ee034e98de7c9caf7a0d2d2e952f7ef74534066953ec461ad0f3fd2359a9a47c\": container with ID starting with ee034e98de7c9caf7a0d2d2e952f7ef74534066953ec461ad0f3fd2359a9a47c not found: ID does not exist" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.447073 4810 scope.go:117] "RemoveContainer" containerID="429b2ac6c979ceff96755077a8398310e6de3258b8bbbb97f9f5436478aa9d9c" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.467974 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/087764ab-72da-4656-93c4-cd86bcd386a9-config-data" (OuterVolumeSpecName: "config-data") pod "087764ab-72da-4656-93c4-cd86bcd386a9" (UID: "087764ab-72da-4656-93c4-cd86bcd386a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.477965 4810 scope.go:117] "RemoveContainer" containerID="8d2cd0ef7276f18b52476ac042cd52cb32082380af17e26b8fcc01b21a3008d5" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.506116 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/087764ab-72da-4656-93c4-cd86bcd386a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "087764ab-72da-4656-93c4-cd86bcd386a9" (UID: "087764ab-72da-4656-93c4-cd86bcd386a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.518150 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/087764ab-72da-4656-93c4-cd86bcd386a9-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.518189 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/087764ab-72da-4656-93c4-cd86bcd386a9-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.518204 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pphdt\" (UniqueName: \"kubernetes.io/projected/087764ab-72da-4656-93c4-cd86bcd386a9-kube-api-access-pphdt\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.518245 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087764ab-72da-4656-93c4-cd86bcd386a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.518365 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhmr9\" (UniqueName: \"kubernetes.io/projected/bbd15d1f-1b0d-452d-8110-537140ccd88c-kube-api-access-lhmr9\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.520783 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bbd15d1f-1b0d-452d-8110-537140ccd88c" (UID: "bbd15d1f-1b0d-452d-8110-537140ccd88c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.528632 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/087764ab-72da-4656-93c4-cd86bcd386a9-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "087764ab-72da-4656-93c4-cd86bcd386a9" (UID: "087764ab-72da-4656-93c4-cd86bcd386a9"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.537202 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-config" (OuterVolumeSpecName: "config") pod "bbd15d1f-1b0d-452d-8110-537140ccd88c" (UID: "bbd15d1f-1b0d-452d-8110-537140ccd88c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.538891 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bbd15d1f-1b0d-452d-8110-537140ccd88c" (UID: "bbd15d1f-1b0d-452d-8110-537140ccd88c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.565305 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bbd15d1f-1b0d-452d-8110-537140ccd88c" (UID: "bbd15d1f-1b0d-452d-8110-537140ccd88c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.620373 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.620409 4810 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/087764ab-72da-4656-93c4-cd86bcd386a9-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.620424 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.620435 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.620447 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbd15d1f-1b0d-452d-8110-537140ccd88c-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.712364 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86b95c6b9-2jj57"] Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.724639 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86b95c6b9-2jj57"] Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.733611 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.746825 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.778352 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:28 crc kubenswrapper[4810]: E1201 16:15:28.778991 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbd15d1f-1b0d-452d-8110-537140ccd88c" containerName="dnsmasq-dns" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.779082 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbd15d1f-1b0d-452d-8110-537140ccd88c" containerName="dnsmasq-dns" Dec 01 16:15:28 crc kubenswrapper[4810]: E1201 16:15:28.779146 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087764ab-72da-4656-93c4-cd86bcd386a9" containerName="nova-metadata-log" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.779212 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="087764ab-72da-4656-93c4-cd86bcd386a9" containerName="nova-metadata-log" Dec 01 16:15:28 crc kubenswrapper[4810]: E1201 16:15:28.779269 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087764ab-72da-4656-93c4-cd86bcd386a9" containerName="nova-metadata-metadata" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.779318 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="087764ab-72da-4656-93c4-cd86bcd386a9" containerName="nova-metadata-metadata" Dec 01 16:15:28 crc kubenswrapper[4810]: E1201 16:15:28.779384 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbd15d1f-1b0d-452d-8110-537140ccd88c" containerName="init" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.779436 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbd15d1f-1b0d-452d-8110-537140ccd88c" containerName="init" Dec 01 16:15:28 crc kubenswrapper[4810]: E1201 16:15:28.779502 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37eef714-f1dc-47b3-8489-5204d08789bb" containerName="nova-manage" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.779551 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="37eef714-f1dc-47b3-8489-5204d08789bb" containerName="nova-manage" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.779905 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="087764ab-72da-4656-93c4-cd86bcd386a9" containerName="nova-metadata-metadata" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.779985 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="087764ab-72da-4656-93c4-cd86bcd386a9" containerName="nova-metadata-log" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.780043 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbd15d1f-1b0d-452d-8110-537140ccd88c" containerName="dnsmasq-dns" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.780108 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="37eef714-f1dc-47b3-8489-5204d08789bb" containerName="nova-manage" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.781327 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.783749 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.791150 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.792334 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.823412 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be84761b-974e-451a-be7d-c4ba9a8ea300-config-data\") pod \"nova-metadata-0\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " pod="openstack/nova-metadata-0" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.823669 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be84761b-974e-451a-be7d-c4ba9a8ea300-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " pod="openstack/nova-metadata-0" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.823747 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/be84761b-974e-451a-be7d-c4ba9a8ea300-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " pod="openstack/nova-metadata-0" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.823830 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx54s\" (UniqueName: \"kubernetes.io/projected/be84761b-974e-451a-be7d-c4ba9a8ea300-kube-api-access-jx54s\") pod \"nova-metadata-0\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " pod="openstack/nova-metadata-0" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.823896 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be84761b-974e-451a-be7d-c4ba9a8ea300-logs\") pod \"nova-metadata-0\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " pod="openstack/nova-metadata-0" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.925696 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be84761b-974e-451a-be7d-c4ba9a8ea300-config-data\") pod \"nova-metadata-0\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " pod="openstack/nova-metadata-0" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.925857 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be84761b-974e-451a-be7d-c4ba9a8ea300-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " pod="openstack/nova-metadata-0" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.925917 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/be84761b-974e-451a-be7d-c4ba9a8ea300-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " pod="openstack/nova-metadata-0" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.925981 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx54s\" (UniqueName: \"kubernetes.io/projected/be84761b-974e-451a-be7d-c4ba9a8ea300-kube-api-access-jx54s\") pod \"nova-metadata-0\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " pod="openstack/nova-metadata-0" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.926015 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be84761b-974e-451a-be7d-c4ba9a8ea300-logs\") pod \"nova-metadata-0\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " pod="openstack/nova-metadata-0" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.926748 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be84761b-974e-451a-be7d-c4ba9a8ea300-logs\") pod \"nova-metadata-0\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " pod="openstack/nova-metadata-0" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.945241 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be84761b-974e-451a-be7d-c4ba9a8ea300-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " pod="openstack/nova-metadata-0" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.945918 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/be84761b-974e-451a-be7d-c4ba9a8ea300-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " pod="openstack/nova-metadata-0" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.947126 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be84761b-974e-451a-be7d-c4ba9a8ea300-config-data\") pod \"nova-metadata-0\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " pod="openstack/nova-metadata-0" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.960321 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx54s\" (UniqueName: \"kubernetes.io/projected/be84761b-974e-451a-be7d-c4ba9a8ea300-kube-api-access-jx54s\") pod \"nova-metadata-0\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " pod="openstack/nova-metadata-0" Dec 01 16:15:28 crc kubenswrapper[4810]: I1201 16:15:28.997770 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:29 crc kubenswrapper[4810]: I1201 16:15:29.115370 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 16:15:29 crc kubenswrapper[4810]: I1201 16:15:29.365263 4810 generic.go:334] "Generic (PLEG): container finished" podID="3fcbd109-53af-4ba8-9418-10466117a3d3" containerID="bc231bdd562c07a555c11a5594ba7d693ad7210f6b005e2f7d9065668eb5380a" exitCode=143 Dec 01 16:15:29 crc kubenswrapper[4810]: I1201 16:15:29.365713 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3fcbd109-53af-4ba8-9418-10466117a3d3","Type":"ContainerDied","Data":"bc231bdd562c07a555c11a5594ba7d693ad7210f6b005e2f7d9065668eb5380a"} Dec 01 16:15:29 crc kubenswrapper[4810]: I1201 16:15:29.606590 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:30 crc kubenswrapper[4810]: I1201 16:15:30.385285 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be84761b-974e-451a-be7d-c4ba9a8ea300","Type":"ContainerStarted","Data":"39455d721b30c8b092f65e569a04eb48ffb6a7963ed9eb3b789f8fb07ac618dc"} Dec 01 16:15:30 crc kubenswrapper[4810]: I1201 16:15:30.385721 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be84761b-974e-451a-be7d-c4ba9a8ea300","Type":"ContainerStarted","Data":"c422294469fa10663a6b0fc3c4839a5955f3b74bca3ceacf651466ce00c3af1c"} Dec 01 16:15:30 crc kubenswrapper[4810]: I1201 16:15:30.385760 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be84761b-974e-451a-be7d-c4ba9a8ea300","Type":"ContainerStarted","Data":"2ebe0d93273497c4b999dc795ec11d3dc1ee46ad426b4de3a1be4708010a1e1f"} Dec 01 16:15:30 crc kubenswrapper[4810]: I1201 16:15:30.406355 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.4063338 podStartE2EDuration="2.4063338s" podCreationTimestamp="2025-12-01 16:15:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:15:30.401094418 +0000 UTC m=+6096.164604041" watchObservedRunningTime="2025-12-01 16:15:30.4063338 +0000 UTC m=+6096.169843403" Dec 01 16:15:30 crc kubenswrapper[4810]: I1201 16:15:30.503418 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="087764ab-72da-4656-93c4-cd86bcd386a9" path="/var/lib/kubelet/pods/087764ab-72da-4656-93c4-cd86bcd386a9/volumes" Dec 01 16:15:30 crc kubenswrapper[4810]: I1201 16:15:30.504494 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbd15d1f-1b0d-452d-8110-537140ccd88c" path="/var/lib/kubelet/pods/bbd15d1f-1b0d-452d-8110-537140ccd88c/volumes" Dec 01 16:15:33 crc kubenswrapper[4810]: I1201 16:15:33.490741 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:15:33 crc kubenswrapper[4810]: E1201 16:15:33.491295 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:15:33 crc kubenswrapper[4810]: I1201 16:15:33.997937 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:34 crc kubenswrapper[4810]: I1201 16:15:34.015442 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:34 crc kubenswrapper[4810]: I1201 16:15:34.117375 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 16:15:34 crc kubenswrapper[4810]: I1201 16:15:34.117815 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 16:15:34 crc kubenswrapper[4810]: I1201 16:15:34.437007 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 01 16:15:35 crc kubenswrapper[4810]: I1201 16:15:35.731541 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.176844 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-7pr5v"] Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.178091 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7pr5v" Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.180559 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.180728 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.196085 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-7pr5v"] Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.261251 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rl8q\" (UniqueName: \"kubernetes.io/projected/36bd4279-0ed0-4266-953a-7535b1653ec2-kube-api-access-4rl8q\") pod \"nova-cell1-cell-mapping-7pr5v\" (UID: \"36bd4279-0ed0-4266-953a-7535b1653ec2\") " pod="openstack/nova-cell1-cell-mapping-7pr5v" Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.261305 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36bd4279-0ed0-4266-953a-7535b1653ec2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-7pr5v\" (UID: \"36bd4279-0ed0-4266-953a-7535b1653ec2\") " pod="openstack/nova-cell1-cell-mapping-7pr5v" Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.261342 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36bd4279-0ed0-4266-953a-7535b1653ec2-config-data\") pod \"nova-cell1-cell-mapping-7pr5v\" (UID: \"36bd4279-0ed0-4266-953a-7535b1653ec2\") " pod="openstack/nova-cell1-cell-mapping-7pr5v" Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.261458 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36bd4279-0ed0-4266-953a-7535b1653ec2-scripts\") pod \"nova-cell1-cell-mapping-7pr5v\" (UID: \"36bd4279-0ed0-4266-953a-7535b1653ec2\") " pod="openstack/nova-cell1-cell-mapping-7pr5v" Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.362656 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36bd4279-0ed0-4266-953a-7535b1653ec2-scripts\") pod \"nova-cell1-cell-mapping-7pr5v\" (UID: \"36bd4279-0ed0-4266-953a-7535b1653ec2\") " pod="openstack/nova-cell1-cell-mapping-7pr5v" Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.362750 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rl8q\" (UniqueName: \"kubernetes.io/projected/36bd4279-0ed0-4266-953a-7535b1653ec2-kube-api-access-4rl8q\") pod \"nova-cell1-cell-mapping-7pr5v\" (UID: \"36bd4279-0ed0-4266-953a-7535b1653ec2\") " pod="openstack/nova-cell1-cell-mapping-7pr5v" Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.362775 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36bd4279-0ed0-4266-953a-7535b1653ec2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-7pr5v\" (UID: \"36bd4279-0ed0-4266-953a-7535b1653ec2\") " pod="openstack/nova-cell1-cell-mapping-7pr5v" Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.362804 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36bd4279-0ed0-4266-953a-7535b1653ec2-config-data\") pod \"nova-cell1-cell-mapping-7pr5v\" (UID: \"36bd4279-0ed0-4266-953a-7535b1653ec2\") " pod="openstack/nova-cell1-cell-mapping-7pr5v" Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.368878 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36bd4279-0ed0-4266-953a-7535b1653ec2-scripts\") pod \"nova-cell1-cell-mapping-7pr5v\" (UID: \"36bd4279-0ed0-4266-953a-7535b1653ec2\") " pod="openstack/nova-cell1-cell-mapping-7pr5v" Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.368916 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36bd4279-0ed0-4266-953a-7535b1653ec2-config-data\") pod \"nova-cell1-cell-mapping-7pr5v\" (UID: \"36bd4279-0ed0-4266-953a-7535b1653ec2\") " pod="openstack/nova-cell1-cell-mapping-7pr5v" Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.377859 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36bd4279-0ed0-4266-953a-7535b1653ec2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-7pr5v\" (UID: \"36bd4279-0ed0-4266-953a-7535b1653ec2\") " pod="openstack/nova-cell1-cell-mapping-7pr5v" Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.379683 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rl8q\" (UniqueName: \"kubernetes.io/projected/36bd4279-0ed0-4266-953a-7535b1653ec2-kube-api-access-4rl8q\") pod \"nova-cell1-cell-mapping-7pr5v\" (UID: \"36bd4279-0ed0-4266-953a-7535b1653ec2\") " pod="openstack/nova-cell1-cell-mapping-7pr5v" Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.505571 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7pr5v" Dec 01 16:15:36 crc kubenswrapper[4810]: I1201 16:15:36.986001 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-7pr5v"] Dec 01 16:15:37 crc kubenswrapper[4810]: I1201 16:15:37.459732 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7pr5v" event={"ID":"36bd4279-0ed0-4266-953a-7535b1653ec2","Type":"ContainerStarted","Data":"550ed3cf881185aa310513f0a4de4346116cfc8e8b9498af1450fd43e74377fa"} Dec 01 16:15:37 crc kubenswrapper[4810]: I1201 16:15:37.459999 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7pr5v" event={"ID":"36bd4279-0ed0-4266-953a-7535b1653ec2","Type":"ContainerStarted","Data":"bddf2e9a88f058ffcbc9ee29e85039afcb1ca3c006d3796bd2bd4870c16cebec"} Dec 01 16:15:37 crc kubenswrapper[4810]: I1201 16:15:37.487181 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-7pr5v" podStartSLOduration=1.487159223 podStartE2EDuration="1.487159223s" podCreationTimestamp="2025-12-01 16:15:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:15:37.4741496 +0000 UTC m=+6103.237659203" watchObservedRunningTime="2025-12-01 16:15:37.487159223 +0000 UTC m=+6103.250668836" Dec 01 16:15:39 crc kubenswrapper[4810]: I1201 16:15:39.117017 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 16:15:39 crc kubenswrapper[4810]: I1201 16:15:39.117373 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 16:15:40 crc kubenswrapper[4810]: I1201 16:15:40.129712 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="be84761b-974e-451a-be7d-c4ba9a8ea300" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.86:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 16:15:40 crc kubenswrapper[4810]: I1201 16:15:40.130128 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="be84761b-974e-451a-be7d-c4ba9a8ea300" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.86:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.250237 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.389535 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fcbd109-53af-4ba8-9418-10466117a3d3-config-data\") pod \"3fcbd109-53af-4ba8-9418-10466117a3d3\" (UID: \"3fcbd109-53af-4ba8-9418-10466117a3d3\") " Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.389642 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fcbd109-53af-4ba8-9418-10466117a3d3-logs\") pod \"3fcbd109-53af-4ba8-9418-10466117a3d3\" (UID: \"3fcbd109-53af-4ba8-9418-10466117a3d3\") " Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.389677 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4z4gj\" (UniqueName: \"kubernetes.io/projected/3fcbd109-53af-4ba8-9418-10466117a3d3-kube-api-access-4z4gj\") pod \"3fcbd109-53af-4ba8-9418-10466117a3d3\" (UID: \"3fcbd109-53af-4ba8-9418-10466117a3d3\") " Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.389707 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fcbd109-53af-4ba8-9418-10466117a3d3-combined-ca-bundle\") pod \"3fcbd109-53af-4ba8-9418-10466117a3d3\" (UID: \"3fcbd109-53af-4ba8-9418-10466117a3d3\") " Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.390782 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fcbd109-53af-4ba8-9418-10466117a3d3-logs" (OuterVolumeSpecName: "logs") pod "3fcbd109-53af-4ba8-9418-10466117a3d3" (UID: "3fcbd109-53af-4ba8-9418-10466117a3d3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.404374 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fcbd109-53af-4ba8-9418-10466117a3d3-kube-api-access-4z4gj" (OuterVolumeSpecName: "kube-api-access-4z4gj") pod "3fcbd109-53af-4ba8-9418-10466117a3d3" (UID: "3fcbd109-53af-4ba8-9418-10466117a3d3"). InnerVolumeSpecName "kube-api-access-4z4gj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.421295 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fcbd109-53af-4ba8-9418-10466117a3d3-config-data" (OuterVolumeSpecName: "config-data") pod "3fcbd109-53af-4ba8-9418-10466117a3d3" (UID: "3fcbd109-53af-4ba8-9418-10466117a3d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.424734 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fcbd109-53af-4ba8-9418-10466117a3d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3fcbd109-53af-4ba8-9418-10466117a3d3" (UID: "3fcbd109-53af-4ba8-9418-10466117a3d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.491989 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fcbd109-53af-4ba8-9418-10466117a3d3-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.492293 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fcbd109-53af-4ba8-9418-10466117a3d3-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.492321 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4z4gj\" (UniqueName: \"kubernetes.io/projected/3fcbd109-53af-4ba8-9418-10466117a3d3-kube-api-access-4z4gj\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.492334 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fcbd109-53af-4ba8-9418-10466117a3d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.509570 4810 generic.go:334] "Generic (PLEG): container finished" podID="3fcbd109-53af-4ba8-9418-10466117a3d3" containerID="a48b965be1fb027015443f27baee0ee5689d05f71ed3a69d979efc1f793ddf01" exitCode=0 Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.509655 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.509659 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3fcbd109-53af-4ba8-9418-10466117a3d3","Type":"ContainerDied","Data":"a48b965be1fb027015443f27baee0ee5689d05f71ed3a69d979efc1f793ddf01"} Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.509791 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3fcbd109-53af-4ba8-9418-10466117a3d3","Type":"ContainerDied","Data":"c55d8fe051997af708c2fe07e693fbcbf6942b6f29f514b3a654b4110929fc79"} Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.509812 4810 scope.go:117] "RemoveContainer" containerID="a48b965be1fb027015443f27baee0ee5689d05f71ed3a69d979efc1f793ddf01" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.512055 4810 generic.go:334] "Generic (PLEG): container finished" podID="36bd4279-0ed0-4266-953a-7535b1653ec2" containerID="550ed3cf881185aa310513f0a4de4346116cfc8e8b9498af1450fd43e74377fa" exitCode=0 Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.512092 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7pr5v" event={"ID":"36bd4279-0ed0-4266-953a-7535b1653ec2","Type":"ContainerDied","Data":"550ed3cf881185aa310513f0a4de4346116cfc8e8b9498af1450fd43e74377fa"} Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.553221 4810 scope.go:117] "RemoveContainer" containerID="bc231bdd562c07a555c11a5594ba7d693ad7210f6b005e2f7d9065668eb5380a" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.560262 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.575845 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.596728 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 16:15:42 crc kubenswrapper[4810]: E1201 16:15:42.597309 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fcbd109-53af-4ba8-9418-10466117a3d3" containerName="nova-api-api" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.597332 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fcbd109-53af-4ba8-9418-10466117a3d3" containerName="nova-api-api" Dec 01 16:15:42 crc kubenswrapper[4810]: E1201 16:15:42.597347 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fcbd109-53af-4ba8-9418-10466117a3d3" containerName="nova-api-log" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.597354 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fcbd109-53af-4ba8-9418-10466117a3d3" containerName="nova-api-log" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.597640 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fcbd109-53af-4ba8-9418-10466117a3d3" containerName="nova-api-log" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.597662 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fcbd109-53af-4ba8-9418-10466117a3d3" containerName="nova-api-api" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.598763 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.602333 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.605449 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.606531 4810 scope.go:117] "RemoveContainer" containerID="a48b965be1fb027015443f27baee0ee5689d05f71ed3a69d979efc1f793ddf01" Dec 01 16:15:42 crc kubenswrapper[4810]: E1201 16:15:42.607161 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a48b965be1fb027015443f27baee0ee5689d05f71ed3a69d979efc1f793ddf01\": container with ID starting with a48b965be1fb027015443f27baee0ee5689d05f71ed3a69d979efc1f793ddf01 not found: ID does not exist" containerID="a48b965be1fb027015443f27baee0ee5689d05f71ed3a69d979efc1f793ddf01" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.607185 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a48b965be1fb027015443f27baee0ee5689d05f71ed3a69d979efc1f793ddf01"} err="failed to get container status \"a48b965be1fb027015443f27baee0ee5689d05f71ed3a69d979efc1f793ddf01\": rpc error: code = NotFound desc = could not find container \"a48b965be1fb027015443f27baee0ee5689d05f71ed3a69d979efc1f793ddf01\": container with ID starting with a48b965be1fb027015443f27baee0ee5689d05f71ed3a69d979efc1f793ddf01 not found: ID does not exist" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.607204 4810 scope.go:117] "RemoveContainer" containerID="bc231bdd562c07a555c11a5594ba7d693ad7210f6b005e2f7d9065668eb5380a" Dec 01 16:15:42 crc kubenswrapper[4810]: E1201 16:15:42.607474 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc231bdd562c07a555c11a5594ba7d693ad7210f6b005e2f7d9065668eb5380a\": container with ID starting with bc231bdd562c07a555c11a5594ba7d693ad7210f6b005e2f7d9065668eb5380a not found: ID does not exist" containerID="bc231bdd562c07a555c11a5594ba7d693ad7210f6b005e2f7d9065668eb5380a" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.607515 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc231bdd562c07a555c11a5594ba7d693ad7210f6b005e2f7d9065668eb5380a"} err="failed to get container status \"bc231bdd562c07a555c11a5594ba7d693ad7210f6b005e2f7d9065668eb5380a\": rpc error: code = NotFound desc = could not find container \"bc231bdd562c07a555c11a5594ba7d693ad7210f6b005e2f7d9065668eb5380a\": container with ID starting with bc231bdd562c07a555c11a5594ba7d693ad7210f6b005e2f7d9065668eb5380a not found: ID does not exist" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.695943 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f11fa28-8010-4c4f-aedb-b3fe144893f7-config-data\") pod \"nova-api-0\" (UID: \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\") " pod="openstack/nova-api-0" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.696028 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hv69\" (UniqueName: \"kubernetes.io/projected/7f11fa28-8010-4c4f-aedb-b3fe144893f7-kube-api-access-4hv69\") pod \"nova-api-0\" (UID: \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\") " pod="openstack/nova-api-0" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.696114 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f11fa28-8010-4c4f-aedb-b3fe144893f7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\") " pod="openstack/nova-api-0" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.696396 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f11fa28-8010-4c4f-aedb-b3fe144893f7-logs\") pod \"nova-api-0\" (UID: \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\") " pod="openstack/nova-api-0" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.798210 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f11fa28-8010-4c4f-aedb-b3fe144893f7-config-data\") pod \"nova-api-0\" (UID: \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\") " pod="openstack/nova-api-0" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.798279 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hv69\" (UniqueName: \"kubernetes.io/projected/7f11fa28-8010-4c4f-aedb-b3fe144893f7-kube-api-access-4hv69\") pod \"nova-api-0\" (UID: \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\") " pod="openstack/nova-api-0" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.798371 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f11fa28-8010-4c4f-aedb-b3fe144893f7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\") " pod="openstack/nova-api-0" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.798452 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f11fa28-8010-4c4f-aedb-b3fe144893f7-logs\") pod \"nova-api-0\" (UID: \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\") " pod="openstack/nova-api-0" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.799095 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f11fa28-8010-4c4f-aedb-b3fe144893f7-logs\") pod \"nova-api-0\" (UID: \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\") " pod="openstack/nova-api-0" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.803676 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f11fa28-8010-4c4f-aedb-b3fe144893f7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\") " pod="openstack/nova-api-0" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.804288 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f11fa28-8010-4c4f-aedb-b3fe144893f7-config-data\") pod \"nova-api-0\" (UID: \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\") " pod="openstack/nova-api-0" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.816109 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hv69\" (UniqueName: \"kubernetes.io/projected/7f11fa28-8010-4c4f-aedb-b3fe144893f7-kube-api-access-4hv69\") pod \"nova-api-0\" (UID: \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\") " pod="openstack/nova-api-0" Dec 01 16:15:42 crc kubenswrapper[4810]: I1201 16:15:42.915201 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 16:15:43 crc kubenswrapper[4810]: I1201 16:15:43.341101 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 16:15:43 crc kubenswrapper[4810]: I1201 16:15:43.520914 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7f11fa28-8010-4c4f-aedb-b3fe144893f7","Type":"ContainerStarted","Data":"7d9e69556111e97cf9fafb3be73f23b56bdf7d07cfe4300efe8b13734c104469"} Dec 01 16:15:43 crc kubenswrapper[4810]: I1201 16:15:43.520975 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7f11fa28-8010-4c4f-aedb-b3fe144893f7","Type":"ContainerStarted","Data":"5808be7d362578861dc16a4dddff1e2ea9b46d1705a5132533d18db329020200"} Dec 01 16:15:43 crc kubenswrapper[4810]: I1201 16:15:43.790331 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7pr5v" Dec 01 16:15:43 crc kubenswrapper[4810]: I1201 16:15:43.919999 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36bd4279-0ed0-4266-953a-7535b1653ec2-scripts\") pod \"36bd4279-0ed0-4266-953a-7535b1653ec2\" (UID: \"36bd4279-0ed0-4266-953a-7535b1653ec2\") " Dec 01 16:15:43 crc kubenswrapper[4810]: I1201 16:15:43.920071 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36bd4279-0ed0-4266-953a-7535b1653ec2-combined-ca-bundle\") pod \"36bd4279-0ed0-4266-953a-7535b1653ec2\" (UID: \"36bd4279-0ed0-4266-953a-7535b1653ec2\") " Dec 01 16:15:43 crc kubenswrapper[4810]: I1201 16:15:43.920145 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rl8q\" (UniqueName: \"kubernetes.io/projected/36bd4279-0ed0-4266-953a-7535b1653ec2-kube-api-access-4rl8q\") pod \"36bd4279-0ed0-4266-953a-7535b1653ec2\" (UID: \"36bd4279-0ed0-4266-953a-7535b1653ec2\") " Dec 01 16:15:43 crc kubenswrapper[4810]: I1201 16:15:43.920238 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36bd4279-0ed0-4266-953a-7535b1653ec2-config-data\") pod \"36bd4279-0ed0-4266-953a-7535b1653ec2\" (UID: \"36bd4279-0ed0-4266-953a-7535b1653ec2\") " Dec 01 16:15:43 crc kubenswrapper[4810]: I1201 16:15:43.924593 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36bd4279-0ed0-4266-953a-7535b1653ec2-scripts" (OuterVolumeSpecName: "scripts") pod "36bd4279-0ed0-4266-953a-7535b1653ec2" (UID: "36bd4279-0ed0-4266-953a-7535b1653ec2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:43 crc kubenswrapper[4810]: I1201 16:15:43.925020 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36bd4279-0ed0-4266-953a-7535b1653ec2-kube-api-access-4rl8q" (OuterVolumeSpecName: "kube-api-access-4rl8q") pod "36bd4279-0ed0-4266-953a-7535b1653ec2" (UID: "36bd4279-0ed0-4266-953a-7535b1653ec2"). InnerVolumeSpecName "kube-api-access-4rl8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:43 crc kubenswrapper[4810]: I1201 16:15:43.955666 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36bd4279-0ed0-4266-953a-7535b1653ec2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36bd4279-0ed0-4266-953a-7535b1653ec2" (UID: "36bd4279-0ed0-4266-953a-7535b1653ec2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:43 crc kubenswrapper[4810]: I1201 16:15:43.955691 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36bd4279-0ed0-4266-953a-7535b1653ec2-config-data" (OuterVolumeSpecName: "config-data") pod "36bd4279-0ed0-4266-953a-7535b1653ec2" (UID: "36bd4279-0ed0-4266-953a-7535b1653ec2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:44 crc kubenswrapper[4810]: I1201 16:15:44.022237 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36bd4279-0ed0-4266-953a-7535b1653ec2-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:44 crc kubenswrapper[4810]: I1201 16:15:44.022269 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36bd4279-0ed0-4266-953a-7535b1653ec2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:44 crc kubenswrapper[4810]: I1201 16:15:44.022280 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rl8q\" (UniqueName: \"kubernetes.io/projected/36bd4279-0ed0-4266-953a-7535b1653ec2-kube-api-access-4rl8q\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:44 crc kubenswrapper[4810]: I1201 16:15:44.022297 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36bd4279-0ed0-4266-953a-7535b1653ec2-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:44 crc kubenswrapper[4810]: I1201 16:15:44.505498 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fcbd109-53af-4ba8-9418-10466117a3d3" path="/var/lib/kubelet/pods/3fcbd109-53af-4ba8-9418-10466117a3d3/volumes" Dec 01 16:15:44 crc kubenswrapper[4810]: I1201 16:15:44.540368 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7f11fa28-8010-4c4f-aedb-b3fe144893f7","Type":"ContainerStarted","Data":"eec84cab948a09e69085c9a28702d090a4a9342a8536780a92526970be5b8ac4"} Dec 01 16:15:44 crc kubenswrapper[4810]: I1201 16:15:44.547494 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7pr5v" event={"ID":"36bd4279-0ed0-4266-953a-7535b1653ec2","Type":"ContainerDied","Data":"bddf2e9a88f058ffcbc9ee29e85039afcb1ca3c006d3796bd2bd4870c16cebec"} Dec 01 16:15:44 crc kubenswrapper[4810]: I1201 16:15:44.547540 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bddf2e9a88f058ffcbc9ee29e85039afcb1ca3c006d3796bd2bd4870c16cebec" Dec 01 16:15:44 crc kubenswrapper[4810]: I1201 16:15:44.547549 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7pr5v" Dec 01 16:15:44 crc kubenswrapper[4810]: I1201 16:15:44.570146 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.570130074 podStartE2EDuration="2.570130074s" podCreationTimestamp="2025-12-01 16:15:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:15:44.563994197 +0000 UTC m=+6110.327503860" watchObservedRunningTime="2025-12-01 16:15:44.570130074 +0000 UTC m=+6110.333639677" Dec 01 16:15:44 crc kubenswrapper[4810]: I1201 16:15:44.709713 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 16:15:44 crc kubenswrapper[4810]: I1201 16:15:44.788513 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:44 crc kubenswrapper[4810]: I1201 16:15:44.788804 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="be84761b-974e-451a-be7d-c4ba9a8ea300" containerName="nova-metadata-log" containerID="cri-o://c422294469fa10663a6b0fc3c4839a5955f3b74bca3ceacf651466ce00c3af1c" gracePeriod=30 Dec 01 16:15:44 crc kubenswrapper[4810]: I1201 16:15:44.788860 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="be84761b-974e-451a-be7d-c4ba9a8ea300" containerName="nova-metadata-metadata" containerID="cri-o://39455d721b30c8b092f65e569a04eb48ffb6a7963ed9eb3b789f8fb07ac618dc" gracePeriod=30 Dec 01 16:15:45 crc kubenswrapper[4810]: I1201 16:15:45.557218 4810 generic.go:334] "Generic (PLEG): container finished" podID="be84761b-974e-451a-be7d-c4ba9a8ea300" containerID="c422294469fa10663a6b0fc3c4839a5955f3b74bca3ceacf651466ce00c3af1c" exitCode=143 Dec 01 16:15:45 crc kubenswrapper[4810]: I1201 16:15:45.557309 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be84761b-974e-451a-be7d-c4ba9a8ea300","Type":"ContainerDied","Data":"c422294469fa10663a6b0fc3c4839a5955f3b74bca3ceacf651466ce00c3af1c"} Dec 01 16:15:46 crc kubenswrapper[4810]: I1201 16:15:46.491566 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:15:46 crc kubenswrapper[4810]: E1201 16:15:46.492049 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:15:46 crc kubenswrapper[4810]: I1201 16:15:46.565865 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7f11fa28-8010-4c4f-aedb-b3fe144893f7" containerName="nova-api-log" containerID="cri-o://7d9e69556111e97cf9fafb3be73f23b56bdf7d07cfe4300efe8b13734c104469" gracePeriod=30 Dec 01 16:15:46 crc kubenswrapper[4810]: I1201 16:15:46.565895 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7f11fa28-8010-4c4f-aedb-b3fe144893f7" containerName="nova-api-api" containerID="cri-o://eec84cab948a09e69085c9a28702d090a4a9342a8536780a92526970be5b8ac4" gracePeriod=30 Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.025835 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.182826 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f11fa28-8010-4c4f-aedb-b3fe144893f7-config-data\") pod \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\" (UID: \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\") " Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.182971 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hv69\" (UniqueName: \"kubernetes.io/projected/7f11fa28-8010-4c4f-aedb-b3fe144893f7-kube-api-access-4hv69\") pod \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\" (UID: \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\") " Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.183013 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f11fa28-8010-4c4f-aedb-b3fe144893f7-logs\") pod \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\" (UID: \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\") " Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.183091 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f11fa28-8010-4c4f-aedb-b3fe144893f7-combined-ca-bundle\") pod \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\" (UID: \"7f11fa28-8010-4c4f-aedb-b3fe144893f7\") " Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.183525 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f11fa28-8010-4c4f-aedb-b3fe144893f7-logs" (OuterVolumeSpecName: "logs") pod "7f11fa28-8010-4c4f-aedb-b3fe144893f7" (UID: "7f11fa28-8010-4c4f-aedb-b3fe144893f7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.184041 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f11fa28-8010-4c4f-aedb-b3fe144893f7-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.207029 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f11fa28-8010-4c4f-aedb-b3fe144893f7-kube-api-access-4hv69" (OuterVolumeSpecName: "kube-api-access-4hv69") pod "7f11fa28-8010-4c4f-aedb-b3fe144893f7" (UID: "7f11fa28-8010-4c4f-aedb-b3fe144893f7"). InnerVolumeSpecName "kube-api-access-4hv69". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.216615 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f11fa28-8010-4c4f-aedb-b3fe144893f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f11fa28-8010-4c4f-aedb-b3fe144893f7" (UID: "7f11fa28-8010-4c4f-aedb-b3fe144893f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.225583 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f11fa28-8010-4c4f-aedb-b3fe144893f7-config-data" (OuterVolumeSpecName: "config-data") pod "7f11fa28-8010-4c4f-aedb-b3fe144893f7" (UID: "7f11fa28-8010-4c4f-aedb-b3fe144893f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.285560 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hv69\" (UniqueName: \"kubernetes.io/projected/7f11fa28-8010-4c4f-aedb-b3fe144893f7-kube-api-access-4hv69\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.285600 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f11fa28-8010-4c4f-aedb-b3fe144893f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.285612 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f11fa28-8010-4c4f-aedb-b3fe144893f7-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.577569 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.577565 4810 generic.go:334] "Generic (PLEG): container finished" podID="7f11fa28-8010-4c4f-aedb-b3fe144893f7" containerID="eec84cab948a09e69085c9a28702d090a4a9342a8536780a92526970be5b8ac4" exitCode=0 Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.577719 4810 generic.go:334] "Generic (PLEG): container finished" podID="7f11fa28-8010-4c4f-aedb-b3fe144893f7" containerID="7d9e69556111e97cf9fafb3be73f23b56bdf7d07cfe4300efe8b13734c104469" exitCode=143 Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.578417 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7f11fa28-8010-4c4f-aedb-b3fe144893f7","Type":"ContainerDied","Data":"eec84cab948a09e69085c9a28702d090a4a9342a8536780a92526970be5b8ac4"} Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.578557 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7f11fa28-8010-4c4f-aedb-b3fe144893f7","Type":"ContainerDied","Data":"7d9e69556111e97cf9fafb3be73f23b56bdf7d07cfe4300efe8b13734c104469"} Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.578654 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7f11fa28-8010-4c4f-aedb-b3fe144893f7","Type":"ContainerDied","Data":"5808be7d362578861dc16a4dddff1e2ea9b46d1705a5132533d18db329020200"} Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.578734 4810 scope.go:117] "RemoveContainer" containerID="eec84cab948a09e69085c9a28702d090a4a9342a8536780a92526970be5b8ac4" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.613341 4810 scope.go:117] "RemoveContainer" containerID="7d9e69556111e97cf9fafb3be73f23b56bdf7d07cfe4300efe8b13734c104469" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.622990 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.636279 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.646694 4810 scope.go:117] "RemoveContainer" containerID="eec84cab948a09e69085c9a28702d090a4a9342a8536780a92526970be5b8ac4" Dec 01 16:15:47 crc kubenswrapper[4810]: E1201 16:15:47.647214 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eec84cab948a09e69085c9a28702d090a4a9342a8536780a92526970be5b8ac4\": container with ID starting with eec84cab948a09e69085c9a28702d090a4a9342a8536780a92526970be5b8ac4 not found: ID does not exist" containerID="eec84cab948a09e69085c9a28702d090a4a9342a8536780a92526970be5b8ac4" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.647257 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eec84cab948a09e69085c9a28702d090a4a9342a8536780a92526970be5b8ac4"} err="failed to get container status \"eec84cab948a09e69085c9a28702d090a4a9342a8536780a92526970be5b8ac4\": rpc error: code = NotFound desc = could not find container \"eec84cab948a09e69085c9a28702d090a4a9342a8536780a92526970be5b8ac4\": container with ID starting with eec84cab948a09e69085c9a28702d090a4a9342a8536780a92526970be5b8ac4 not found: ID does not exist" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.647284 4810 scope.go:117] "RemoveContainer" containerID="7d9e69556111e97cf9fafb3be73f23b56bdf7d07cfe4300efe8b13734c104469" Dec 01 16:15:47 crc kubenswrapper[4810]: E1201 16:15:47.647565 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d9e69556111e97cf9fafb3be73f23b56bdf7d07cfe4300efe8b13734c104469\": container with ID starting with 7d9e69556111e97cf9fafb3be73f23b56bdf7d07cfe4300efe8b13734c104469 not found: ID does not exist" containerID="7d9e69556111e97cf9fafb3be73f23b56bdf7d07cfe4300efe8b13734c104469" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.647597 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d9e69556111e97cf9fafb3be73f23b56bdf7d07cfe4300efe8b13734c104469"} err="failed to get container status \"7d9e69556111e97cf9fafb3be73f23b56bdf7d07cfe4300efe8b13734c104469\": rpc error: code = NotFound desc = could not find container \"7d9e69556111e97cf9fafb3be73f23b56bdf7d07cfe4300efe8b13734c104469\": container with ID starting with 7d9e69556111e97cf9fafb3be73f23b56bdf7d07cfe4300efe8b13734c104469 not found: ID does not exist" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.647618 4810 scope.go:117] "RemoveContainer" containerID="eec84cab948a09e69085c9a28702d090a4a9342a8536780a92526970be5b8ac4" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.648561 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eec84cab948a09e69085c9a28702d090a4a9342a8536780a92526970be5b8ac4"} err="failed to get container status \"eec84cab948a09e69085c9a28702d090a4a9342a8536780a92526970be5b8ac4\": rpc error: code = NotFound desc = could not find container \"eec84cab948a09e69085c9a28702d090a4a9342a8536780a92526970be5b8ac4\": container with ID starting with eec84cab948a09e69085c9a28702d090a4a9342a8536780a92526970be5b8ac4 not found: ID does not exist" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.648594 4810 scope.go:117] "RemoveContainer" containerID="7d9e69556111e97cf9fafb3be73f23b56bdf7d07cfe4300efe8b13734c104469" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.653592 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 16:15:47 crc kubenswrapper[4810]: E1201 16:15:47.654110 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f11fa28-8010-4c4f-aedb-b3fe144893f7" containerName="nova-api-api" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.654132 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f11fa28-8010-4c4f-aedb-b3fe144893f7" containerName="nova-api-api" Dec 01 16:15:47 crc kubenswrapper[4810]: E1201 16:15:47.654167 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36bd4279-0ed0-4266-953a-7535b1653ec2" containerName="nova-manage" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.654176 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="36bd4279-0ed0-4266-953a-7535b1653ec2" containerName="nova-manage" Dec 01 16:15:47 crc kubenswrapper[4810]: E1201 16:15:47.654190 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f11fa28-8010-4c4f-aedb-b3fe144893f7" containerName="nova-api-log" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.654199 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f11fa28-8010-4c4f-aedb-b3fe144893f7" containerName="nova-api-log" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.654441 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f11fa28-8010-4c4f-aedb-b3fe144893f7" containerName="nova-api-api" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.654485 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f11fa28-8010-4c4f-aedb-b3fe144893f7" containerName="nova-api-log" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.654508 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="36bd4279-0ed0-4266-953a-7535b1653ec2" containerName="nova-manage" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.655781 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.668174 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.668912 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d9e69556111e97cf9fafb3be73f23b56bdf7d07cfe4300efe8b13734c104469"} err="failed to get container status \"7d9e69556111e97cf9fafb3be73f23b56bdf7d07cfe4300efe8b13734c104469\": rpc error: code = NotFound desc = could not find container \"7d9e69556111e97cf9fafb3be73f23b56bdf7d07cfe4300efe8b13734c104469\": container with ID starting with 7d9e69556111e97cf9fafb3be73f23b56bdf7d07cfe4300efe8b13734c104469 not found: ID does not exist" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.669244 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.793021 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b833c4d1-603f-476a-928a-2cb312cdebcb-logs\") pod \"nova-api-0\" (UID: \"b833c4d1-603f-476a-928a-2cb312cdebcb\") " pod="openstack/nova-api-0" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.793168 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs5vb\" (UniqueName: \"kubernetes.io/projected/b833c4d1-603f-476a-928a-2cb312cdebcb-kube-api-access-zs5vb\") pod \"nova-api-0\" (UID: \"b833c4d1-603f-476a-928a-2cb312cdebcb\") " pod="openstack/nova-api-0" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.793209 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b833c4d1-603f-476a-928a-2cb312cdebcb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b833c4d1-603f-476a-928a-2cb312cdebcb\") " pod="openstack/nova-api-0" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.793240 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b833c4d1-603f-476a-928a-2cb312cdebcb-config-data\") pod \"nova-api-0\" (UID: \"b833c4d1-603f-476a-928a-2cb312cdebcb\") " pod="openstack/nova-api-0" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.895228 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b833c4d1-603f-476a-928a-2cb312cdebcb-logs\") pod \"nova-api-0\" (UID: \"b833c4d1-603f-476a-928a-2cb312cdebcb\") " pod="openstack/nova-api-0" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.895339 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs5vb\" (UniqueName: \"kubernetes.io/projected/b833c4d1-603f-476a-928a-2cb312cdebcb-kube-api-access-zs5vb\") pod \"nova-api-0\" (UID: \"b833c4d1-603f-476a-928a-2cb312cdebcb\") " pod="openstack/nova-api-0" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.895368 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b833c4d1-603f-476a-928a-2cb312cdebcb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b833c4d1-603f-476a-928a-2cb312cdebcb\") " pod="openstack/nova-api-0" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.895394 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b833c4d1-603f-476a-928a-2cb312cdebcb-config-data\") pod \"nova-api-0\" (UID: \"b833c4d1-603f-476a-928a-2cb312cdebcb\") " pod="openstack/nova-api-0" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.895811 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b833c4d1-603f-476a-928a-2cb312cdebcb-logs\") pod \"nova-api-0\" (UID: \"b833c4d1-603f-476a-928a-2cb312cdebcb\") " pod="openstack/nova-api-0" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.900652 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b833c4d1-603f-476a-928a-2cb312cdebcb-config-data\") pod \"nova-api-0\" (UID: \"b833c4d1-603f-476a-928a-2cb312cdebcb\") " pod="openstack/nova-api-0" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.906628 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b833c4d1-603f-476a-928a-2cb312cdebcb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b833c4d1-603f-476a-928a-2cb312cdebcb\") " pod="openstack/nova-api-0" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.911151 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs5vb\" (UniqueName: \"kubernetes.io/projected/b833c4d1-603f-476a-928a-2cb312cdebcb-kube-api-access-zs5vb\") pod \"nova-api-0\" (UID: \"b833c4d1-603f-476a-928a-2cb312cdebcb\") " pod="openstack/nova-api-0" Dec 01 16:15:47 crc kubenswrapper[4810]: I1201 16:15:47.981392 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 16:15:48 crc kubenswrapper[4810]: I1201 16:15:48.474075 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 16:15:48 crc kubenswrapper[4810]: W1201 16:15:48.475998 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb833c4d1_603f_476a_928a_2cb312cdebcb.slice/crio-105f56e0ed466b2191f96211c4d59a734cf87103c7b9f9f7404c43e98cfb36a7 WatchSource:0}: Error finding container 105f56e0ed466b2191f96211c4d59a734cf87103c7b9f9f7404c43e98cfb36a7: Status 404 returned error can't find the container with id 105f56e0ed466b2191f96211c4d59a734cf87103c7b9f9f7404c43e98cfb36a7 Dec 01 16:15:48 crc kubenswrapper[4810]: I1201 16:15:48.506698 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f11fa28-8010-4c4f-aedb-b3fe144893f7" path="/var/lib/kubelet/pods/7f11fa28-8010-4c4f-aedb-b3fe144893f7/volumes" Dec 01 16:15:48 crc kubenswrapper[4810]: I1201 16:15:48.587859 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b833c4d1-603f-476a-928a-2cb312cdebcb","Type":"ContainerStarted","Data":"105f56e0ed466b2191f96211c4d59a734cf87103c7b9f9f7404c43e98cfb36a7"} Dec 01 16:15:49 crc kubenswrapper[4810]: I1201 16:15:49.603143 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b833c4d1-603f-476a-928a-2cb312cdebcb","Type":"ContainerStarted","Data":"418411aa6fa8dbd3cf409abf2c65e687e43f8275fd309ca9491be136ad9947a9"} Dec 01 16:15:49 crc kubenswrapper[4810]: I1201 16:15:49.603505 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b833c4d1-603f-476a-928a-2cb312cdebcb","Type":"ContainerStarted","Data":"20c09e08fc167aa2aa622534472c8866787947691d9ca5f7a384e4a16bfc2248"} Dec 01 16:15:49 crc kubenswrapper[4810]: I1201 16:15:49.640109 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.6400829679999998 podStartE2EDuration="2.640082968s" podCreationTimestamp="2025-12-01 16:15:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:15:49.630423135 +0000 UTC m=+6115.393932738" watchObservedRunningTime="2025-12-01 16:15:49.640082968 +0000 UTC m=+6115.403592571" Dec 01 16:15:56 crc kubenswrapper[4810]: I1201 16:15:56.926937 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-flbt8"] Dec 01 16:15:56 crc kubenswrapper[4810]: I1201 16:15:56.929528 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-flbt8" Dec 01 16:15:56 crc kubenswrapper[4810]: I1201 16:15:56.962849 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-flbt8"] Dec 01 16:15:57 crc kubenswrapper[4810]: I1201 16:15:57.014460 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6380c11-56f3-4737-893d-60c095fb0a4b-catalog-content\") pod \"certified-operators-flbt8\" (UID: \"c6380c11-56f3-4737-893d-60c095fb0a4b\") " pod="openshift-marketplace/certified-operators-flbt8" Dec 01 16:15:57 crc kubenswrapper[4810]: I1201 16:15:57.014587 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6380c11-56f3-4737-893d-60c095fb0a4b-utilities\") pod \"certified-operators-flbt8\" (UID: \"c6380c11-56f3-4737-893d-60c095fb0a4b\") " pod="openshift-marketplace/certified-operators-flbt8" Dec 01 16:15:57 crc kubenswrapper[4810]: I1201 16:15:57.014634 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw2kc\" (UniqueName: \"kubernetes.io/projected/c6380c11-56f3-4737-893d-60c095fb0a4b-kube-api-access-jw2kc\") pod \"certified-operators-flbt8\" (UID: \"c6380c11-56f3-4737-893d-60c095fb0a4b\") " pod="openshift-marketplace/certified-operators-flbt8" Dec 01 16:15:57 crc kubenswrapper[4810]: I1201 16:15:57.115803 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6380c11-56f3-4737-893d-60c095fb0a4b-utilities\") pod \"certified-operators-flbt8\" (UID: \"c6380c11-56f3-4737-893d-60c095fb0a4b\") " pod="openshift-marketplace/certified-operators-flbt8" Dec 01 16:15:57 crc kubenswrapper[4810]: I1201 16:15:57.115870 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw2kc\" (UniqueName: \"kubernetes.io/projected/c6380c11-56f3-4737-893d-60c095fb0a4b-kube-api-access-jw2kc\") pod \"certified-operators-flbt8\" (UID: \"c6380c11-56f3-4737-893d-60c095fb0a4b\") " pod="openshift-marketplace/certified-operators-flbt8" Dec 01 16:15:57 crc kubenswrapper[4810]: I1201 16:15:57.115930 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6380c11-56f3-4737-893d-60c095fb0a4b-catalog-content\") pod \"certified-operators-flbt8\" (UID: \"c6380c11-56f3-4737-893d-60c095fb0a4b\") " pod="openshift-marketplace/certified-operators-flbt8" Dec 01 16:15:57 crc kubenswrapper[4810]: I1201 16:15:57.116323 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6380c11-56f3-4737-893d-60c095fb0a4b-catalog-content\") pod \"certified-operators-flbt8\" (UID: \"c6380c11-56f3-4737-893d-60c095fb0a4b\") " pod="openshift-marketplace/certified-operators-flbt8" Dec 01 16:15:57 crc kubenswrapper[4810]: I1201 16:15:57.116324 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6380c11-56f3-4737-893d-60c095fb0a4b-utilities\") pod \"certified-operators-flbt8\" (UID: \"c6380c11-56f3-4737-893d-60c095fb0a4b\") " pod="openshift-marketplace/certified-operators-flbt8" Dec 01 16:15:57 crc kubenswrapper[4810]: I1201 16:15:57.136832 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw2kc\" (UniqueName: \"kubernetes.io/projected/c6380c11-56f3-4737-893d-60c095fb0a4b-kube-api-access-jw2kc\") pod \"certified-operators-flbt8\" (UID: \"c6380c11-56f3-4737-893d-60c095fb0a4b\") " pod="openshift-marketplace/certified-operators-flbt8" Dec 01 16:15:57 crc kubenswrapper[4810]: I1201 16:15:57.269059 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-flbt8" Dec 01 16:15:57 crc kubenswrapper[4810]: I1201 16:15:57.675196 4810 generic.go:334] "Generic (PLEG): container finished" podID="febfc322-9d2b-470b-bfdf-f7ee31996fc8" containerID="6a2eec563994f549af7272b80ded697cd4379e34fc30dd094901d08ad0f9408c" exitCode=137 Dec 01 16:15:57 crc kubenswrapper[4810]: I1201 16:15:57.675271 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"febfc322-9d2b-470b-bfdf-f7ee31996fc8","Type":"ContainerDied","Data":"6a2eec563994f549af7272b80ded697cd4379e34fc30dd094901d08ad0f9408c"} Dec 01 16:15:57 crc kubenswrapper[4810]: I1201 16:15:57.753261 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-flbt8"] Dec 01 16:15:57 crc kubenswrapper[4810]: W1201 16:15:57.757054 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6380c11_56f3_4737_893d_60c095fb0a4b.slice/crio-793040c2e72c79da17fd25af2874e361336393473330f2993caa4098747eb5e5 WatchSource:0}: Error finding container 793040c2e72c79da17fd25af2874e361336393473330f2993caa4098747eb5e5: Status 404 returned error can't find the container with id 793040c2e72c79da17fd25af2874e361336393473330f2993caa4098747eb5e5 Dec 01 16:15:57 crc kubenswrapper[4810]: I1201 16:15:57.982277 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 16:15:57 crc kubenswrapper[4810]: I1201 16:15:57.984389 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.616337 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.687783 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"febfc322-9d2b-470b-bfdf-f7ee31996fc8","Type":"ContainerDied","Data":"5157e0ad3bc0bb968412975155d21da950ea5f0a80421aea49fc9f7ffcdab2f2"} Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.687845 4810 scope.go:117] "RemoveContainer" containerID="6a2eec563994f549af7272b80ded697cd4379e34fc30dd094901d08ad0f9408c" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.687982 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.687985 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.690872 4810 generic.go:334] "Generic (PLEG): container finished" podID="c6380c11-56f3-4737-893d-60c095fb0a4b" containerID="819705bbb25234456ebfd926e1782ee866c42f44e3e3d39b440a0df7170e66d6" exitCode=0 Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.691820 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flbt8" event={"ID":"c6380c11-56f3-4737-893d-60c095fb0a4b","Type":"ContainerDied","Data":"819705bbb25234456ebfd926e1782ee866c42f44e3e3d39b440a0df7170e66d6"} Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.692005 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flbt8" event={"ID":"c6380c11-56f3-4737-893d-60c095fb0a4b","Type":"ContainerStarted","Data":"793040c2e72c79da17fd25af2874e361336393473330f2993caa4098747eb5e5"} Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.693350 4810 generic.go:334] "Generic (PLEG): container finished" podID="be84761b-974e-451a-be7d-c4ba9a8ea300" containerID="39455d721b30c8b092f65e569a04eb48ffb6a7963ed9eb3b789f8fb07ac618dc" exitCode=0 Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.693384 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be84761b-974e-451a-be7d-c4ba9a8ea300","Type":"ContainerDied","Data":"39455d721b30c8b092f65e569a04eb48ffb6a7963ed9eb3b789f8fb07ac618dc"} Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.693409 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be84761b-974e-451a-be7d-c4ba9a8ea300","Type":"ContainerDied","Data":"2ebe0d93273497c4b999dc795ec11d3dc1ee46ad426b4de3a1be4708010a1e1f"} Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.693460 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.717956 4810 scope.go:117] "RemoveContainer" containerID="39455d721b30c8b092f65e569a04eb48ffb6a7963ed9eb3b789f8fb07ac618dc" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.760455 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/febfc322-9d2b-470b-bfdf-f7ee31996fc8-combined-ca-bundle\") pod \"febfc322-9d2b-470b-bfdf-f7ee31996fc8\" (UID: \"febfc322-9d2b-470b-bfdf-f7ee31996fc8\") " Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.760603 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2qfj\" (UniqueName: \"kubernetes.io/projected/febfc322-9d2b-470b-bfdf-f7ee31996fc8-kube-api-access-s2qfj\") pod \"febfc322-9d2b-470b-bfdf-f7ee31996fc8\" (UID: \"febfc322-9d2b-470b-bfdf-f7ee31996fc8\") " Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.760748 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/febfc322-9d2b-470b-bfdf-f7ee31996fc8-config-data\") pod \"febfc322-9d2b-470b-bfdf-f7ee31996fc8\" (UID: \"febfc322-9d2b-470b-bfdf-f7ee31996fc8\") " Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.764875 4810 scope.go:117] "RemoveContainer" containerID="c422294469fa10663a6b0fc3c4839a5955f3b74bca3ceacf651466ce00c3af1c" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.765598 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/febfc322-9d2b-470b-bfdf-f7ee31996fc8-kube-api-access-s2qfj" (OuterVolumeSpecName: "kube-api-access-s2qfj") pod "febfc322-9d2b-470b-bfdf-f7ee31996fc8" (UID: "febfc322-9d2b-470b-bfdf-f7ee31996fc8"). InnerVolumeSpecName "kube-api-access-s2qfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.782234 4810 scope.go:117] "RemoveContainer" containerID="39455d721b30c8b092f65e569a04eb48ffb6a7963ed9eb3b789f8fb07ac618dc" Dec 01 16:15:58 crc kubenswrapper[4810]: E1201 16:15:58.782816 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39455d721b30c8b092f65e569a04eb48ffb6a7963ed9eb3b789f8fb07ac618dc\": container with ID starting with 39455d721b30c8b092f65e569a04eb48ffb6a7963ed9eb3b789f8fb07ac618dc not found: ID does not exist" containerID="39455d721b30c8b092f65e569a04eb48ffb6a7963ed9eb3b789f8fb07ac618dc" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.782869 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39455d721b30c8b092f65e569a04eb48ffb6a7963ed9eb3b789f8fb07ac618dc"} err="failed to get container status \"39455d721b30c8b092f65e569a04eb48ffb6a7963ed9eb3b789f8fb07ac618dc\": rpc error: code = NotFound desc = could not find container \"39455d721b30c8b092f65e569a04eb48ffb6a7963ed9eb3b789f8fb07ac618dc\": container with ID starting with 39455d721b30c8b092f65e569a04eb48ffb6a7963ed9eb3b789f8fb07ac618dc not found: ID does not exist" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.782895 4810 scope.go:117] "RemoveContainer" containerID="c422294469fa10663a6b0fc3c4839a5955f3b74bca3ceacf651466ce00c3af1c" Dec 01 16:15:58 crc kubenswrapper[4810]: E1201 16:15:58.783234 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c422294469fa10663a6b0fc3c4839a5955f3b74bca3ceacf651466ce00c3af1c\": container with ID starting with c422294469fa10663a6b0fc3c4839a5955f3b74bca3ceacf651466ce00c3af1c not found: ID does not exist" containerID="c422294469fa10663a6b0fc3c4839a5955f3b74bca3ceacf651466ce00c3af1c" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.783270 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c422294469fa10663a6b0fc3c4839a5955f3b74bca3ceacf651466ce00c3af1c"} err="failed to get container status \"c422294469fa10663a6b0fc3c4839a5955f3b74bca3ceacf651466ce00c3af1c\": rpc error: code = NotFound desc = could not find container \"c422294469fa10663a6b0fc3c4839a5955f3b74bca3ceacf651466ce00c3af1c\": container with ID starting with c422294469fa10663a6b0fc3c4839a5955f3b74bca3ceacf651466ce00c3af1c not found: ID does not exist" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.786287 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/febfc322-9d2b-470b-bfdf-f7ee31996fc8-config-data" (OuterVolumeSpecName: "config-data") pod "febfc322-9d2b-470b-bfdf-f7ee31996fc8" (UID: "febfc322-9d2b-470b-bfdf-f7ee31996fc8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.788104 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/febfc322-9d2b-470b-bfdf-f7ee31996fc8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "febfc322-9d2b-470b-bfdf-f7ee31996fc8" (UID: "febfc322-9d2b-470b-bfdf-f7ee31996fc8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.862190 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be84761b-974e-451a-be7d-c4ba9a8ea300-config-data\") pod \"be84761b-974e-451a-be7d-c4ba9a8ea300\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.862292 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be84761b-974e-451a-be7d-c4ba9a8ea300-combined-ca-bundle\") pod \"be84761b-974e-451a-be7d-c4ba9a8ea300\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.862326 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jx54s\" (UniqueName: \"kubernetes.io/projected/be84761b-974e-451a-be7d-c4ba9a8ea300-kube-api-access-jx54s\") pod \"be84761b-974e-451a-be7d-c4ba9a8ea300\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.862418 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/be84761b-974e-451a-be7d-c4ba9a8ea300-nova-metadata-tls-certs\") pod \"be84761b-974e-451a-be7d-c4ba9a8ea300\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.862556 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be84761b-974e-451a-be7d-c4ba9a8ea300-logs\") pod \"be84761b-974e-451a-be7d-c4ba9a8ea300\" (UID: \"be84761b-974e-451a-be7d-c4ba9a8ea300\") " Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.863170 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2qfj\" (UniqueName: \"kubernetes.io/projected/febfc322-9d2b-470b-bfdf-f7ee31996fc8-kube-api-access-s2qfj\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.863202 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/febfc322-9d2b-470b-bfdf-f7ee31996fc8-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.863217 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/febfc322-9d2b-470b-bfdf-f7ee31996fc8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.864917 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be84761b-974e-451a-be7d-c4ba9a8ea300-logs" (OuterVolumeSpecName: "logs") pod "be84761b-974e-451a-be7d-c4ba9a8ea300" (UID: "be84761b-974e-451a-be7d-c4ba9a8ea300"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.867215 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be84761b-974e-451a-be7d-c4ba9a8ea300-kube-api-access-jx54s" (OuterVolumeSpecName: "kube-api-access-jx54s") pod "be84761b-974e-451a-be7d-c4ba9a8ea300" (UID: "be84761b-974e-451a-be7d-c4ba9a8ea300"). InnerVolumeSpecName "kube-api-access-jx54s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.891414 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be84761b-974e-451a-be7d-c4ba9a8ea300-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be84761b-974e-451a-be7d-c4ba9a8ea300" (UID: "be84761b-974e-451a-be7d-c4ba9a8ea300"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.892310 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be84761b-974e-451a-be7d-c4ba9a8ea300-config-data" (OuterVolumeSpecName: "config-data") pod "be84761b-974e-451a-be7d-c4ba9a8ea300" (UID: "be84761b-974e-451a-be7d-c4ba9a8ea300"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.913814 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be84761b-974e-451a-be7d-c4ba9a8ea300-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "be84761b-974e-451a-be7d-c4ba9a8ea300" (UID: "be84761b-974e-451a-be7d-c4ba9a8ea300"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.967923 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be84761b-974e-451a-be7d-c4ba9a8ea300-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.967990 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be84761b-974e-451a-be7d-c4ba9a8ea300-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.968008 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jx54s\" (UniqueName: \"kubernetes.io/projected/be84761b-974e-451a-be7d-c4ba9a8ea300-kube-api-access-jx54s\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.968022 4810 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/be84761b-974e-451a-be7d-c4ba9a8ea300-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:58 crc kubenswrapper[4810]: I1201 16:15:58.968034 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be84761b-974e-451a-be7d-c4ba9a8ea300-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.035242 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.061546 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.068011 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b833c4d1-603f-476a-928a-2cb312cdebcb" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.89:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.068318 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b833c4d1-603f-476a-928a-2cb312cdebcb" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.89:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.075143 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 16:15:59 crc kubenswrapper[4810]: E1201 16:15:59.075756 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be84761b-974e-451a-be7d-c4ba9a8ea300" containerName="nova-metadata-metadata" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.075779 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="be84761b-974e-451a-be7d-c4ba9a8ea300" containerName="nova-metadata-metadata" Dec 01 16:15:59 crc kubenswrapper[4810]: E1201 16:15:59.075810 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="febfc322-9d2b-470b-bfdf-f7ee31996fc8" containerName="nova-scheduler-scheduler" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.075820 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="febfc322-9d2b-470b-bfdf-f7ee31996fc8" containerName="nova-scheduler-scheduler" Dec 01 16:15:59 crc kubenswrapper[4810]: E1201 16:15:59.075902 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be84761b-974e-451a-be7d-c4ba9a8ea300" containerName="nova-metadata-log" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.075914 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="be84761b-974e-451a-be7d-c4ba9a8ea300" containerName="nova-metadata-log" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.076157 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="febfc322-9d2b-470b-bfdf-f7ee31996fc8" containerName="nova-scheduler-scheduler" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.076183 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="be84761b-974e-451a-be7d-c4ba9a8ea300" containerName="nova-metadata-log" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.076200 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="be84761b-974e-451a-be7d-c4ba9a8ea300" containerName="nova-metadata-metadata" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.090247 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.093388 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.101036 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.117323 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.126309 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.134431 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.136627 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.140372 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.140425 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.143066 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.273055 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90143656-4ff7-4511-8824-88d158f05993-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"90143656-4ff7-4511-8824-88d158f05993\") " pod="openstack/nova-scheduler-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.273272 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cd84ceb-8494-4328-8c92-0e818910f06e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " pod="openstack/nova-metadata-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.273326 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cd84ceb-8494-4328-8c92-0e818910f06e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " pod="openstack/nova-metadata-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.273557 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cd84ceb-8494-4328-8c92-0e818910f06e-config-data\") pod \"nova-metadata-0\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " pod="openstack/nova-metadata-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.273629 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrf2n\" (UniqueName: \"kubernetes.io/projected/5cd84ceb-8494-4328-8c92-0e818910f06e-kube-api-access-qrf2n\") pod \"nova-metadata-0\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " pod="openstack/nova-metadata-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.273753 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cd84ceb-8494-4328-8c92-0e818910f06e-logs\") pod \"nova-metadata-0\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " pod="openstack/nova-metadata-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.273785 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90143656-4ff7-4511-8824-88d158f05993-config-data\") pod \"nova-scheduler-0\" (UID: \"90143656-4ff7-4511-8824-88d158f05993\") " pod="openstack/nova-scheduler-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.274002 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz6fj\" (UniqueName: \"kubernetes.io/projected/90143656-4ff7-4511-8824-88d158f05993-kube-api-access-nz6fj\") pod \"nova-scheduler-0\" (UID: \"90143656-4ff7-4511-8824-88d158f05993\") " pod="openstack/nova-scheduler-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.375603 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cd84ceb-8494-4328-8c92-0e818910f06e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " pod="openstack/nova-metadata-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.375972 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cd84ceb-8494-4328-8c92-0e818910f06e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " pod="openstack/nova-metadata-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.376560 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cd84ceb-8494-4328-8c92-0e818910f06e-config-data\") pod \"nova-metadata-0\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " pod="openstack/nova-metadata-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.376609 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrf2n\" (UniqueName: \"kubernetes.io/projected/5cd84ceb-8494-4328-8c92-0e818910f06e-kube-api-access-qrf2n\") pod \"nova-metadata-0\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " pod="openstack/nova-metadata-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.376675 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cd84ceb-8494-4328-8c92-0e818910f06e-logs\") pod \"nova-metadata-0\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " pod="openstack/nova-metadata-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.376703 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90143656-4ff7-4511-8824-88d158f05993-config-data\") pod \"nova-scheduler-0\" (UID: \"90143656-4ff7-4511-8824-88d158f05993\") " pod="openstack/nova-scheduler-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.376803 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz6fj\" (UniqueName: \"kubernetes.io/projected/90143656-4ff7-4511-8824-88d158f05993-kube-api-access-nz6fj\") pod \"nova-scheduler-0\" (UID: \"90143656-4ff7-4511-8824-88d158f05993\") " pod="openstack/nova-scheduler-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.376844 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90143656-4ff7-4511-8824-88d158f05993-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"90143656-4ff7-4511-8824-88d158f05993\") " pod="openstack/nova-scheduler-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.377541 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cd84ceb-8494-4328-8c92-0e818910f06e-logs\") pod \"nova-metadata-0\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " pod="openstack/nova-metadata-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.390078 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90143656-4ff7-4511-8824-88d158f05993-config-data\") pod \"nova-scheduler-0\" (UID: \"90143656-4ff7-4511-8824-88d158f05993\") " pod="openstack/nova-scheduler-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.391533 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cd84ceb-8494-4328-8c92-0e818910f06e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " pod="openstack/nova-metadata-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.393184 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cd84ceb-8494-4328-8c92-0e818910f06e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " pod="openstack/nova-metadata-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.393394 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cd84ceb-8494-4328-8c92-0e818910f06e-config-data\") pod \"nova-metadata-0\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " pod="openstack/nova-metadata-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.393838 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90143656-4ff7-4511-8824-88d158f05993-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"90143656-4ff7-4511-8824-88d158f05993\") " pod="openstack/nova-scheduler-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.394268 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrf2n\" (UniqueName: \"kubernetes.io/projected/5cd84ceb-8494-4328-8c92-0e818910f06e-kube-api-access-qrf2n\") pod \"nova-metadata-0\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " pod="openstack/nova-metadata-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.395186 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz6fj\" (UniqueName: \"kubernetes.io/projected/90143656-4ff7-4511-8824-88d158f05993-kube-api-access-nz6fj\") pod \"nova-scheduler-0\" (UID: \"90143656-4ff7-4511-8824-88d158f05993\") " pod="openstack/nova-scheduler-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.413696 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.546924 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 16:15:59 crc kubenswrapper[4810]: I1201 16:15:59.888952 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 16:15:59 crc kubenswrapper[4810]: W1201 16:15:59.904834 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90143656_4ff7_4511_8824_88d158f05993.slice/crio-d3add81d4db38a45e441e110f9e1a43a9361e4f71ff23d464d24dfe826c72bd0 WatchSource:0}: Error finding container d3add81d4db38a45e441e110f9e1a43a9361e4f71ff23d464d24dfe826c72bd0: Status 404 returned error can't find the container with id d3add81d4db38a45e441e110f9e1a43a9361e4f71ff23d464d24dfe826c72bd0 Dec 01 16:16:00 crc kubenswrapper[4810]: I1201 16:16:00.026380 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 16:16:00 crc kubenswrapper[4810]: W1201 16:16:00.033640 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cd84ceb_8494_4328_8c92_0e818910f06e.slice/crio-f72abb675295eaf87b10cde1dedea6029b101469ceaaf0e22524854850ad6ca2 WatchSource:0}: Error finding container f72abb675295eaf87b10cde1dedea6029b101469ceaaf0e22524854850ad6ca2: Status 404 returned error can't find the container with id f72abb675295eaf87b10cde1dedea6029b101469ceaaf0e22524854850ad6ca2 Dec 01 16:16:00 crc kubenswrapper[4810]: I1201 16:16:00.501834 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be84761b-974e-451a-be7d-c4ba9a8ea300" path="/var/lib/kubelet/pods/be84761b-974e-451a-be7d-c4ba9a8ea300/volumes" Dec 01 16:16:00 crc kubenswrapper[4810]: I1201 16:16:00.502975 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="febfc322-9d2b-470b-bfdf-f7ee31996fc8" path="/var/lib/kubelet/pods/febfc322-9d2b-470b-bfdf-f7ee31996fc8/volumes" Dec 01 16:16:00 crc kubenswrapper[4810]: I1201 16:16:00.720816 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5cd84ceb-8494-4328-8c92-0e818910f06e","Type":"ContainerStarted","Data":"23e5a07093be2eca9b799a768274021963eb0c7a728239a891950c901ac444d8"} Dec 01 16:16:00 crc kubenswrapper[4810]: I1201 16:16:00.721207 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5cd84ceb-8494-4328-8c92-0e818910f06e","Type":"ContainerStarted","Data":"3db908e2f008cd7cbe619bad66f45cf79b61c7c1c2e6812ee0b8632d24bbf459"} Dec 01 16:16:00 crc kubenswrapper[4810]: I1201 16:16:00.721224 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5cd84ceb-8494-4328-8c92-0e818910f06e","Type":"ContainerStarted","Data":"f72abb675295eaf87b10cde1dedea6029b101469ceaaf0e22524854850ad6ca2"} Dec 01 16:16:00 crc kubenswrapper[4810]: I1201 16:16:00.723138 4810 generic.go:334] "Generic (PLEG): container finished" podID="c6380c11-56f3-4737-893d-60c095fb0a4b" containerID="36313b447af7a81677f7974fc1c49aa3a4c9364ccfa74fcb45f009a91a1abb2b" exitCode=0 Dec 01 16:16:00 crc kubenswrapper[4810]: I1201 16:16:00.723269 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flbt8" event={"ID":"c6380c11-56f3-4737-893d-60c095fb0a4b","Type":"ContainerDied","Data":"36313b447af7a81677f7974fc1c49aa3a4c9364ccfa74fcb45f009a91a1abb2b"} Dec 01 16:16:00 crc kubenswrapper[4810]: I1201 16:16:00.725129 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"90143656-4ff7-4511-8824-88d158f05993","Type":"ContainerStarted","Data":"6ac9d4f430ea9f350983bbe17059146649a35ad2ed281b9a2c05e082a9ecfdd6"} Dec 01 16:16:00 crc kubenswrapper[4810]: I1201 16:16:00.725174 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"90143656-4ff7-4511-8824-88d158f05993","Type":"ContainerStarted","Data":"d3add81d4db38a45e441e110f9e1a43a9361e4f71ff23d464d24dfe826c72bd0"} Dec 01 16:16:00 crc kubenswrapper[4810]: I1201 16:16:00.763230 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.7632076859999999 podStartE2EDuration="1.763207686s" podCreationTimestamp="2025-12-01 16:15:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:16:00.751920619 +0000 UTC m=+6126.515430222" watchObservedRunningTime="2025-12-01 16:16:00.763207686 +0000 UTC m=+6126.526717289" Dec 01 16:16:00 crc kubenswrapper[4810]: I1201 16:16:00.790098 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.790073995 podStartE2EDuration="1.790073995s" podCreationTimestamp="2025-12-01 16:15:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:16:00.782428268 +0000 UTC m=+6126.545937911" watchObservedRunningTime="2025-12-01 16:16:00.790073995 +0000 UTC m=+6126.553583598" Dec 01 16:16:01 crc kubenswrapper[4810]: I1201 16:16:01.492002 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:16:01 crc kubenswrapper[4810]: E1201 16:16:01.492330 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:16:02 crc kubenswrapper[4810]: I1201 16:16:02.743615 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flbt8" event={"ID":"c6380c11-56f3-4737-893d-60c095fb0a4b","Type":"ContainerStarted","Data":"cf7b6f5dd11533a73af088f70d110217b3ccd1deea252713edff7a7b03a00567"} Dec 01 16:16:02 crc kubenswrapper[4810]: I1201 16:16:02.761296 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-flbt8" podStartSLOduration=3.35217726 podStartE2EDuration="6.761279727s" podCreationTimestamp="2025-12-01 16:15:56 +0000 UTC" firstStartedPulling="2025-12-01 16:15:58.695304218 +0000 UTC m=+6124.458813821" lastFinishedPulling="2025-12-01 16:16:02.104406685 +0000 UTC m=+6127.867916288" observedRunningTime="2025-12-01 16:16:02.758464001 +0000 UTC m=+6128.521973624" watchObservedRunningTime="2025-12-01 16:16:02.761279727 +0000 UTC m=+6128.524789330" Dec 01 16:16:04 crc kubenswrapper[4810]: I1201 16:16:04.414024 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 16:16:04 crc kubenswrapper[4810]: I1201 16:16:04.547073 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 16:16:04 crc kubenswrapper[4810]: I1201 16:16:04.547323 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 16:16:07 crc kubenswrapper[4810]: I1201 16:16:07.269858 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-flbt8" Dec 01 16:16:07 crc kubenswrapper[4810]: I1201 16:16:07.270192 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-flbt8" Dec 01 16:16:07 crc kubenswrapper[4810]: I1201 16:16:07.318139 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-flbt8" Dec 01 16:16:07 crc kubenswrapper[4810]: I1201 16:16:07.899431 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-flbt8" Dec 01 16:16:07 crc kubenswrapper[4810]: I1201 16:16:07.955988 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-flbt8"] Dec 01 16:16:07 crc kubenswrapper[4810]: I1201 16:16:07.985989 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 16:16:07 crc kubenswrapper[4810]: I1201 16:16:07.986943 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 16:16:07 crc kubenswrapper[4810]: I1201 16:16:07.990297 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 16:16:07 crc kubenswrapper[4810]: I1201 16:16:07.990450 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 16:16:08 crc kubenswrapper[4810]: I1201 16:16:08.802113 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 16:16:08 crc kubenswrapper[4810]: I1201 16:16:08.805167 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.005513 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-797f6f57f5-n86v7"] Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.007457 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.033783 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-ovsdbserver-nb\") pod \"dnsmasq-dns-797f6f57f5-n86v7\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.033897 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-797f6f57f5-n86v7"] Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.033911 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzf5s\" (UniqueName: \"kubernetes.io/projected/4c524b0f-d099-489b-9675-a44654fc1c53-kube-api-access-qzf5s\") pod \"dnsmasq-dns-797f6f57f5-n86v7\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.033953 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-dns-svc\") pod \"dnsmasq-dns-797f6f57f5-n86v7\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.033995 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-config\") pod \"dnsmasq-dns-797f6f57f5-n86v7\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.034059 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-ovsdbserver-sb\") pod \"dnsmasq-dns-797f6f57f5-n86v7\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.135125 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-ovsdbserver-nb\") pod \"dnsmasq-dns-797f6f57f5-n86v7\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.135234 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzf5s\" (UniqueName: \"kubernetes.io/projected/4c524b0f-d099-489b-9675-a44654fc1c53-kube-api-access-qzf5s\") pod \"dnsmasq-dns-797f6f57f5-n86v7\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.135272 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-dns-svc\") pod \"dnsmasq-dns-797f6f57f5-n86v7\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.135298 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-config\") pod \"dnsmasq-dns-797f6f57f5-n86v7\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.135327 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-ovsdbserver-sb\") pod \"dnsmasq-dns-797f6f57f5-n86v7\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.136738 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-ovsdbserver-nb\") pod \"dnsmasq-dns-797f6f57f5-n86v7\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.137016 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-dns-svc\") pod \"dnsmasq-dns-797f6f57f5-n86v7\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.137297 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-config\") pod \"dnsmasq-dns-797f6f57f5-n86v7\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.138367 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-ovsdbserver-sb\") pod \"dnsmasq-dns-797f6f57f5-n86v7\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.155219 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzf5s\" (UniqueName: \"kubernetes.io/projected/4c524b0f-d099-489b-9675-a44654fc1c53-kube-api-access-qzf5s\") pod \"dnsmasq-dns-797f6f57f5-n86v7\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.338963 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.414803 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.492786 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.547805 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.547852 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.811331 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-flbt8" podUID="c6380c11-56f3-4737-893d-60c095fb0a4b" containerName="registry-server" containerID="cri-o://cf7b6f5dd11533a73af088f70d110217b3ccd1deea252713edff7a7b03a00567" gracePeriod=2 Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.823593 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-797f6f57f5-n86v7"] Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.844083 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 16:16:09 crc kubenswrapper[4810]: I1201 16:16:09.970501 4810 scope.go:117] "RemoveContainer" containerID="5f1d8eaa3d64ad44d9bd25eb7824637f056506dd2b88b09e4774d4cf0605d56f" Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.018559 4810 scope.go:117] "RemoveContainer" containerID="bbda8149a7db5b2157b310da3506a0525a2e2aed87b3b5d64c5c8d7e2d3f9757" Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.314867 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-flbt8" Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.392745 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6380c11-56f3-4737-893d-60c095fb0a4b-catalog-content\") pod \"c6380c11-56f3-4737-893d-60c095fb0a4b\" (UID: \"c6380c11-56f3-4737-893d-60c095fb0a4b\") " Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.392997 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jw2kc\" (UniqueName: \"kubernetes.io/projected/c6380c11-56f3-4737-893d-60c095fb0a4b-kube-api-access-jw2kc\") pod \"c6380c11-56f3-4737-893d-60c095fb0a4b\" (UID: \"c6380c11-56f3-4737-893d-60c095fb0a4b\") " Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.393103 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6380c11-56f3-4737-893d-60c095fb0a4b-utilities\") pod \"c6380c11-56f3-4737-893d-60c095fb0a4b\" (UID: \"c6380c11-56f3-4737-893d-60c095fb0a4b\") " Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.394460 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6380c11-56f3-4737-893d-60c095fb0a4b-utilities" (OuterVolumeSpecName: "utilities") pod "c6380c11-56f3-4737-893d-60c095fb0a4b" (UID: "c6380c11-56f3-4737-893d-60c095fb0a4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.400682 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6380c11-56f3-4737-893d-60c095fb0a4b-kube-api-access-jw2kc" (OuterVolumeSpecName: "kube-api-access-jw2kc") pod "c6380c11-56f3-4737-893d-60c095fb0a4b" (UID: "c6380c11-56f3-4737-893d-60c095fb0a4b"). InnerVolumeSpecName "kube-api-access-jw2kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.464794 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6380c11-56f3-4737-893d-60c095fb0a4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c6380c11-56f3-4737-893d-60c095fb0a4b" (UID: "c6380c11-56f3-4737-893d-60c095fb0a4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.495409 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6380c11-56f3-4737-893d-60c095fb0a4b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.495442 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jw2kc\" (UniqueName: \"kubernetes.io/projected/c6380c11-56f3-4737-893d-60c095fb0a4b-kube-api-access-jw2kc\") on node \"crc\" DevicePath \"\"" Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.495455 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6380c11-56f3-4737-893d-60c095fb0a4b-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.564664 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5cd84ceb-8494-4328-8c92-0e818910f06e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.92:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.564700 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5cd84ceb-8494-4328-8c92-0e818910f06e" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.92:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.836958 4810 generic.go:334] "Generic (PLEG): container finished" podID="c6380c11-56f3-4737-893d-60c095fb0a4b" containerID="cf7b6f5dd11533a73af088f70d110217b3ccd1deea252713edff7a7b03a00567" exitCode=0 Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.837048 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flbt8" event={"ID":"c6380c11-56f3-4737-893d-60c095fb0a4b","Type":"ContainerDied","Data":"cf7b6f5dd11533a73af088f70d110217b3ccd1deea252713edff7a7b03a00567"} Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.837107 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-flbt8" event={"ID":"c6380c11-56f3-4737-893d-60c095fb0a4b","Type":"ContainerDied","Data":"793040c2e72c79da17fd25af2874e361336393473330f2993caa4098747eb5e5"} Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.837136 4810 scope.go:117] "RemoveContainer" containerID="cf7b6f5dd11533a73af088f70d110217b3ccd1deea252713edff7a7b03a00567" Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.837137 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-flbt8" Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.840713 4810 generic.go:334] "Generic (PLEG): container finished" podID="4c524b0f-d099-489b-9675-a44654fc1c53" containerID="a4d3a67c7d0b60488a4e6ca4913a6bb709ca4a854614390cf68e4af7b9d97eb2" exitCode=0 Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.840767 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" event={"ID":"4c524b0f-d099-489b-9675-a44654fc1c53","Type":"ContainerDied","Data":"a4d3a67c7d0b60488a4e6ca4913a6bb709ca4a854614390cf68e4af7b9d97eb2"} Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.841136 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" event={"ID":"4c524b0f-d099-489b-9675-a44654fc1c53","Type":"ContainerStarted","Data":"2fa2a4a45b28c455713878d35c67cedc9caafed8ac22cf6a3bf2e2e1c394b22f"} Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.873378 4810 scope.go:117] "RemoveContainer" containerID="36313b447af7a81677f7974fc1c49aa3a4c9364ccfa74fcb45f009a91a1abb2b" Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.889541 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-flbt8"] Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.912560 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-flbt8"] Dec 01 16:16:10 crc kubenswrapper[4810]: I1201 16:16:10.989685 4810 scope.go:117] "RemoveContainer" containerID="819705bbb25234456ebfd926e1782ee866c42f44e3e3d39b440a0df7170e66d6" Dec 01 16:16:11 crc kubenswrapper[4810]: I1201 16:16:11.022661 4810 scope.go:117] "RemoveContainer" containerID="cf7b6f5dd11533a73af088f70d110217b3ccd1deea252713edff7a7b03a00567" Dec 01 16:16:11 crc kubenswrapper[4810]: E1201 16:16:11.023891 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf7b6f5dd11533a73af088f70d110217b3ccd1deea252713edff7a7b03a00567\": container with ID starting with cf7b6f5dd11533a73af088f70d110217b3ccd1deea252713edff7a7b03a00567 not found: ID does not exist" containerID="cf7b6f5dd11533a73af088f70d110217b3ccd1deea252713edff7a7b03a00567" Dec 01 16:16:11 crc kubenswrapper[4810]: I1201 16:16:11.024369 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf7b6f5dd11533a73af088f70d110217b3ccd1deea252713edff7a7b03a00567"} err="failed to get container status \"cf7b6f5dd11533a73af088f70d110217b3ccd1deea252713edff7a7b03a00567\": rpc error: code = NotFound desc = could not find container \"cf7b6f5dd11533a73af088f70d110217b3ccd1deea252713edff7a7b03a00567\": container with ID starting with cf7b6f5dd11533a73af088f70d110217b3ccd1deea252713edff7a7b03a00567 not found: ID does not exist" Dec 01 16:16:11 crc kubenswrapper[4810]: I1201 16:16:11.024400 4810 scope.go:117] "RemoveContainer" containerID="36313b447af7a81677f7974fc1c49aa3a4c9364ccfa74fcb45f009a91a1abb2b" Dec 01 16:16:11 crc kubenswrapper[4810]: E1201 16:16:11.025138 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36313b447af7a81677f7974fc1c49aa3a4c9364ccfa74fcb45f009a91a1abb2b\": container with ID starting with 36313b447af7a81677f7974fc1c49aa3a4c9364ccfa74fcb45f009a91a1abb2b not found: ID does not exist" containerID="36313b447af7a81677f7974fc1c49aa3a4c9364ccfa74fcb45f009a91a1abb2b" Dec 01 16:16:11 crc kubenswrapper[4810]: I1201 16:16:11.025181 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36313b447af7a81677f7974fc1c49aa3a4c9364ccfa74fcb45f009a91a1abb2b"} err="failed to get container status \"36313b447af7a81677f7974fc1c49aa3a4c9364ccfa74fcb45f009a91a1abb2b\": rpc error: code = NotFound desc = could not find container \"36313b447af7a81677f7974fc1c49aa3a4c9364ccfa74fcb45f009a91a1abb2b\": container with ID starting with 36313b447af7a81677f7974fc1c49aa3a4c9364ccfa74fcb45f009a91a1abb2b not found: ID does not exist" Dec 01 16:16:11 crc kubenswrapper[4810]: I1201 16:16:11.025212 4810 scope.go:117] "RemoveContainer" containerID="819705bbb25234456ebfd926e1782ee866c42f44e3e3d39b440a0df7170e66d6" Dec 01 16:16:11 crc kubenswrapper[4810]: E1201 16:16:11.027045 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"819705bbb25234456ebfd926e1782ee866c42f44e3e3d39b440a0df7170e66d6\": container with ID starting with 819705bbb25234456ebfd926e1782ee866c42f44e3e3d39b440a0df7170e66d6 not found: ID does not exist" containerID="819705bbb25234456ebfd926e1782ee866c42f44e3e3d39b440a0df7170e66d6" Dec 01 16:16:11 crc kubenswrapper[4810]: I1201 16:16:11.027078 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"819705bbb25234456ebfd926e1782ee866c42f44e3e3d39b440a0df7170e66d6"} err="failed to get container status \"819705bbb25234456ebfd926e1782ee866c42f44e3e3d39b440a0df7170e66d6\": rpc error: code = NotFound desc = could not find container \"819705bbb25234456ebfd926e1782ee866c42f44e3e3d39b440a0df7170e66d6\": container with ID starting with 819705bbb25234456ebfd926e1782ee866c42f44e3e3d39b440a0df7170e66d6 not found: ID does not exist" Dec 01 16:16:11 crc kubenswrapper[4810]: I1201 16:16:11.853928 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" event={"ID":"4c524b0f-d099-489b-9675-a44654fc1c53","Type":"ContainerStarted","Data":"461e34c10cc81aba237ad62874026ed718e6581607abf907f1cab8e9ae099d1a"} Dec 01 16:16:11 crc kubenswrapper[4810]: I1201 16:16:11.854358 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:11 crc kubenswrapper[4810]: I1201 16:16:11.915914 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" podStartSLOduration=3.915892966 podStartE2EDuration="3.915892966s" podCreationTimestamp="2025-12-01 16:16:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:16:11.886975292 +0000 UTC m=+6137.650484915" watchObservedRunningTime="2025-12-01 16:16:11.915892966 +0000 UTC m=+6137.679402569" Dec 01 16:16:11 crc kubenswrapper[4810]: I1201 16:16:11.919742 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 16:16:11 crc kubenswrapper[4810]: I1201 16:16:11.919969 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b833c4d1-603f-476a-928a-2cb312cdebcb" containerName="nova-api-log" containerID="cri-o://20c09e08fc167aa2aa622534472c8866787947691d9ca5f7a384e4a16bfc2248" gracePeriod=30 Dec 01 16:16:11 crc kubenswrapper[4810]: I1201 16:16:11.920059 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b833c4d1-603f-476a-928a-2cb312cdebcb" containerName="nova-api-api" containerID="cri-o://418411aa6fa8dbd3cf409abf2c65e687e43f8275fd309ca9491be136ad9947a9" gracePeriod=30 Dec 01 16:16:12 crc kubenswrapper[4810]: I1201 16:16:12.500389 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6380c11-56f3-4737-893d-60c095fb0a4b" path="/var/lib/kubelet/pods/c6380c11-56f3-4737-893d-60c095fb0a4b/volumes" Dec 01 16:16:12 crc kubenswrapper[4810]: I1201 16:16:12.866179 4810 generic.go:334] "Generic (PLEG): container finished" podID="b833c4d1-603f-476a-928a-2cb312cdebcb" containerID="20c09e08fc167aa2aa622534472c8866787947691d9ca5f7a384e4a16bfc2248" exitCode=143 Dec 01 16:16:12 crc kubenswrapper[4810]: I1201 16:16:12.866244 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b833c4d1-603f-476a-928a-2cb312cdebcb","Type":"ContainerDied","Data":"20c09e08fc167aa2aa622534472c8866787947691d9ca5f7a384e4a16bfc2248"} Dec 01 16:16:13 crc kubenswrapper[4810]: I1201 16:16:13.491334 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:16:13 crc kubenswrapper[4810]: E1201 16:16:13.491997 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.676024 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.740643 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b833c4d1-603f-476a-928a-2cb312cdebcb-combined-ca-bundle\") pod \"b833c4d1-603f-476a-928a-2cb312cdebcb\" (UID: \"b833c4d1-603f-476a-928a-2cb312cdebcb\") " Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.740713 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zs5vb\" (UniqueName: \"kubernetes.io/projected/b833c4d1-603f-476a-928a-2cb312cdebcb-kube-api-access-zs5vb\") pod \"b833c4d1-603f-476a-928a-2cb312cdebcb\" (UID: \"b833c4d1-603f-476a-928a-2cb312cdebcb\") " Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.740891 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b833c4d1-603f-476a-928a-2cb312cdebcb-config-data\") pod \"b833c4d1-603f-476a-928a-2cb312cdebcb\" (UID: \"b833c4d1-603f-476a-928a-2cb312cdebcb\") " Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.740992 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b833c4d1-603f-476a-928a-2cb312cdebcb-logs\") pod \"b833c4d1-603f-476a-928a-2cb312cdebcb\" (UID: \"b833c4d1-603f-476a-928a-2cb312cdebcb\") " Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.741674 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b833c4d1-603f-476a-928a-2cb312cdebcb-logs" (OuterVolumeSpecName: "logs") pod "b833c4d1-603f-476a-928a-2cb312cdebcb" (UID: "b833c4d1-603f-476a-928a-2cb312cdebcb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.746896 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b833c4d1-603f-476a-928a-2cb312cdebcb-kube-api-access-zs5vb" (OuterVolumeSpecName: "kube-api-access-zs5vb") pod "b833c4d1-603f-476a-928a-2cb312cdebcb" (UID: "b833c4d1-603f-476a-928a-2cb312cdebcb"). InnerVolumeSpecName "kube-api-access-zs5vb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.774675 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b833c4d1-603f-476a-928a-2cb312cdebcb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b833c4d1-603f-476a-928a-2cb312cdebcb" (UID: "b833c4d1-603f-476a-928a-2cb312cdebcb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.796639 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b833c4d1-603f-476a-928a-2cb312cdebcb-config-data" (OuterVolumeSpecName: "config-data") pod "b833c4d1-603f-476a-928a-2cb312cdebcb" (UID: "b833c4d1-603f-476a-928a-2cb312cdebcb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.844443 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b833c4d1-603f-476a-928a-2cb312cdebcb-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.844500 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b833c4d1-603f-476a-928a-2cb312cdebcb-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.844512 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b833c4d1-603f-476a-928a-2cb312cdebcb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.844537 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zs5vb\" (UniqueName: \"kubernetes.io/projected/b833c4d1-603f-476a-928a-2cb312cdebcb-kube-api-access-zs5vb\") on node \"crc\" DevicePath \"\"" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.917517 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gd5zz"] Dec 01 16:16:16 crc kubenswrapper[4810]: E1201 16:16:16.917926 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6380c11-56f3-4737-893d-60c095fb0a4b" containerName="extract-content" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.917943 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6380c11-56f3-4737-893d-60c095fb0a4b" containerName="extract-content" Dec 01 16:16:16 crc kubenswrapper[4810]: E1201 16:16:16.917962 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6380c11-56f3-4737-893d-60c095fb0a4b" containerName="extract-utilities" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.917970 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6380c11-56f3-4737-893d-60c095fb0a4b" containerName="extract-utilities" Dec 01 16:16:16 crc kubenswrapper[4810]: E1201 16:16:16.917984 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b833c4d1-603f-476a-928a-2cb312cdebcb" containerName="nova-api-log" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.917992 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b833c4d1-603f-476a-928a-2cb312cdebcb" containerName="nova-api-log" Dec 01 16:16:16 crc kubenswrapper[4810]: E1201 16:16:16.918018 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b833c4d1-603f-476a-928a-2cb312cdebcb" containerName="nova-api-api" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.918024 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b833c4d1-603f-476a-928a-2cb312cdebcb" containerName="nova-api-api" Dec 01 16:16:16 crc kubenswrapper[4810]: E1201 16:16:16.918031 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6380c11-56f3-4737-893d-60c095fb0a4b" containerName="registry-server" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.918036 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6380c11-56f3-4737-893d-60c095fb0a4b" containerName="registry-server" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.918196 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b833c4d1-603f-476a-928a-2cb312cdebcb" containerName="nova-api-log" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.918217 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b833c4d1-603f-476a-928a-2cb312cdebcb" containerName="nova-api-api" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.918228 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6380c11-56f3-4737-893d-60c095fb0a4b" containerName="registry-server" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.919617 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gd5zz" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.923390 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gd5zz"] Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.945800 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e8daf3c-58ed-4feb-b574-8bda66dd9c79-utilities\") pod \"community-operators-gd5zz\" (UID: \"9e8daf3c-58ed-4feb-b574-8bda66dd9c79\") " pod="openshift-marketplace/community-operators-gd5zz" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.945888 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt72v\" (UniqueName: \"kubernetes.io/projected/9e8daf3c-58ed-4feb-b574-8bda66dd9c79-kube-api-access-jt72v\") pod \"community-operators-gd5zz\" (UID: \"9e8daf3c-58ed-4feb-b574-8bda66dd9c79\") " pod="openshift-marketplace/community-operators-gd5zz" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.945920 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e8daf3c-58ed-4feb-b574-8bda66dd9c79-catalog-content\") pod \"community-operators-gd5zz\" (UID: \"9e8daf3c-58ed-4feb-b574-8bda66dd9c79\") " pod="openshift-marketplace/community-operators-gd5zz" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.956429 4810 generic.go:334] "Generic (PLEG): container finished" podID="b833c4d1-603f-476a-928a-2cb312cdebcb" containerID="418411aa6fa8dbd3cf409abf2c65e687e43f8275fd309ca9491be136ad9947a9" exitCode=0 Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.956496 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b833c4d1-603f-476a-928a-2cb312cdebcb","Type":"ContainerDied","Data":"418411aa6fa8dbd3cf409abf2c65e687e43f8275fd309ca9491be136ad9947a9"} Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.956526 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b833c4d1-603f-476a-928a-2cb312cdebcb","Type":"ContainerDied","Data":"105f56e0ed466b2191f96211c4d59a734cf87103c7b9f9f7404c43e98cfb36a7"} Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.956544 4810 scope.go:117] "RemoveContainer" containerID="418411aa6fa8dbd3cf409abf2c65e687e43f8275fd309ca9491be136ad9947a9" Dec 01 16:16:16 crc kubenswrapper[4810]: I1201 16:16:16.956727 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.025104 4810 scope.go:117] "RemoveContainer" containerID="20c09e08fc167aa2aa622534472c8866787947691d9ca5f7a384e4a16bfc2248" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.039644 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.047898 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e8daf3c-58ed-4feb-b574-8bda66dd9c79-utilities\") pod \"community-operators-gd5zz\" (UID: \"9e8daf3c-58ed-4feb-b574-8bda66dd9c79\") " pod="openshift-marketplace/community-operators-gd5zz" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.047958 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt72v\" (UniqueName: \"kubernetes.io/projected/9e8daf3c-58ed-4feb-b574-8bda66dd9c79-kube-api-access-jt72v\") pod \"community-operators-gd5zz\" (UID: \"9e8daf3c-58ed-4feb-b574-8bda66dd9c79\") " pod="openshift-marketplace/community-operators-gd5zz" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.047982 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e8daf3c-58ed-4feb-b574-8bda66dd9c79-catalog-content\") pod \"community-operators-gd5zz\" (UID: \"9e8daf3c-58ed-4feb-b574-8bda66dd9c79\") " pod="openshift-marketplace/community-operators-gd5zz" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.048506 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e8daf3c-58ed-4feb-b574-8bda66dd9c79-catalog-content\") pod \"community-operators-gd5zz\" (UID: \"9e8daf3c-58ed-4feb-b574-8bda66dd9c79\") " pod="openshift-marketplace/community-operators-gd5zz" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.048735 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e8daf3c-58ed-4feb-b574-8bda66dd9c79-utilities\") pod \"community-operators-gd5zz\" (UID: \"9e8daf3c-58ed-4feb-b574-8bda66dd9c79\") " pod="openshift-marketplace/community-operators-gd5zz" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.057692 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.057896 4810 scope.go:117] "RemoveContainer" containerID="418411aa6fa8dbd3cf409abf2c65e687e43f8275fd309ca9491be136ad9947a9" Dec 01 16:16:17 crc kubenswrapper[4810]: E1201 16:16:17.060159 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"418411aa6fa8dbd3cf409abf2c65e687e43f8275fd309ca9491be136ad9947a9\": container with ID starting with 418411aa6fa8dbd3cf409abf2c65e687e43f8275fd309ca9491be136ad9947a9 not found: ID does not exist" containerID="418411aa6fa8dbd3cf409abf2c65e687e43f8275fd309ca9491be136ad9947a9" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.060196 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"418411aa6fa8dbd3cf409abf2c65e687e43f8275fd309ca9491be136ad9947a9"} err="failed to get container status \"418411aa6fa8dbd3cf409abf2c65e687e43f8275fd309ca9491be136ad9947a9\": rpc error: code = NotFound desc = could not find container \"418411aa6fa8dbd3cf409abf2c65e687e43f8275fd309ca9491be136ad9947a9\": container with ID starting with 418411aa6fa8dbd3cf409abf2c65e687e43f8275fd309ca9491be136ad9947a9 not found: ID does not exist" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.060222 4810 scope.go:117] "RemoveContainer" containerID="20c09e08fc167aa2aa622534472c8866787947691d9ca5f7a384e4a16bfc2248" Dec 01 16:16:17 crc kubenswrapper[4810]: E1201 16:16:17.062662 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20c09e08fc167aa2aa622534472c8866787947691d9ca5f7a384e4a16bfc2248\": container with ID starting with 20c09e08fc167aa2aa622534472c8866787947691d9ca5f7a384e4a16bfc2248 not found: ID does not exist" containerID="20c09e08fc167aa2aa622534472c8866787947691d9ca5f7a384e4a16bfc2248" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.062695 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20c09e08fc167aa2aa622534472c8866787947691d9ca5f7a384e4a16bfc2248"} err="failed to get container status \"20c09e08fc167aa2aa622534472c8866787947691d9ca5f7a384e4a16bfc2248\": rpc error: code = NotFound desc = could not find container \"20c09e08fc167aa2aa622534472c8866787947691d9ca5f7a384e4a16bfc2248\": container with ID starting with 20c09e08fc167aa2aa622534472c8866787947691d9ca5f7a384e4a16bfc2248 not found: ID does not exist" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.070060 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.071634 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.078633 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt72v\" (UniqueName: \"kubernetes.io/projected/9e8daf3c-58ed-4feb-b574-8bda66dd9c79-kube-api-access-jt72v\") pod \"community-operators-gd5zz\" (UID: \"9e8daf3c-58ed-4feb-b574-8bda66dd9c79\") " pod="openshift-marketplace/community-operators-gd5zz" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.081420 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.082420 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.084595 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.094044 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.149224 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj4hq\" (UniqueName: \"kubernetes.io/projected/30cf5aff-9c39-4531-8ee6-1284597c7d00-kube-api-access-hj4hq\") pod \"nova-api-0\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.149392 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-internal-tls-certs\") pod \"nova-api-0\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.149481 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-public-tls-certs\") pod \"nova-api-0\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.149504 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30cf5aff-9c39-4531-8ee6-1284597c7d00-logs\") pod \"nova-api-0\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.149654 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-config-data\") pod \"nova-api-0\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.149688 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.251160 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj4hq\" (UniqueName: \"kubernetes.io/projected/30cf5aff-9c39-4531-8ee6-1284597c7d00-kube-api-access-hj4hq\") pod \"nova-api-0\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.251261 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-internal-tls-certs\") pod \"nova-api-0\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.251292 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-public-tls-certs\") pod \"nova-api-0\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.251308 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30cf5aff-9c39-4531-8ee6-1284597c7d00-logs\") pod \"nova-api-0\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.251348 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-config-data\") pod \"nova-api-0\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.251362 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.252097 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30cf5aff-9c39-4531-8ee6-1284597c7d00-logs\") pod \"nova-api-0\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.254704 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-internal-tls-certs\") pod \"nova-api-0\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.254783 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-public-tls-certs\") pod \"nova-api-0\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.255525 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-config-data\") pod \"nova-api-0\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.255594 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.270954 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj4hq\" (UniqueName: \"kubernetes.io/projected/30cf5aff-9c39-4531-8ee6-1284597c7d00-kube-api-access-hj4hq\") pod \"nova-api-0\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.309298 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gd5zz" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.391393 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.812164 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gd5zz"] Dec 01 16:16:17 crc kubenswrapper[4810]: W1201 16:16:17.815790 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e8daf3c_58ed_4feb_b574_8bda66dd9c79.slice/crio-95b782f30c7ed0b4957bc75b4b410c18537f624234aa421300686434963c1ee6 WatchSource:0}: Error finding container 95b782f30c7ed0b4957bc75b4b410c18537f624234aa421300686434963c1ee6: Status 404 returned error can't find the container with id 95b782f30c7ed0b4957bc75b4b410c18537f624234aa421300686434963c1ee6 Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.966807 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gd5zz" event={"ID":"9e8daf3c-58ed-4feb-b574-8bda66dd9c79","Type":"ContainerStarted","Data":"a867b5b0fcfc7f3ba0b7f6938f28527326a38e9eee6fae8b4a7b30a8a0f62d67"} Dec 01 16:16:17 crc kubenswrapper[4810]: I1201 16:16:17.966859 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gd5zz" event={"ID":"9e8daf3c-58ed-4feb-b574-8bda66dd9c79","Type":"ContainerStarted","Data":"95b782f30c7ed0b4957bc75b4b410c18537f624234aa421300686434963c1ee6"} Dec 01 16:16:18 crc kubenswrapper[4810]: I1201 16:16:18.008519 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 16:16:18 crc kubenswrapper[4810]: W1201 16:16:18.038308 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30cf5aff_9c39_4531_8ee6_1284597c7d00.slice/crio-19f62766348c357265510cc3791251351783c7de7e96e035ed79b683c0340489 WatchSource:0}: Error finding container 19f62766348c357265510cc3791251351783c7de7e96e035ed79b683c0340489: Status 404 returned error can't find the container with id 19f62766348c357265510cc3791251351783c7de7e96e035ed79b683c0340489 Dec 01 16:16:18 crc kubenswrapper[4810]: I1201 16:16:18.500616 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b833c4d1-603f-476a-928a-2cb312cdebcb" path="/var/lib/kubelet/pods/b833c4d1-603f-476a-928a-2cb312cdebcb/volumes" Dec 01 16:16:18 crc kubenswrapper[4810]: I1201 16:16:18.982935 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30cf5aff-9c39-4531-8ee6-1284597c7d00","Type":"ContainerStarted","Data":"46a37a18b93213b4d6649471f8576ba09563847420101eb829523635836fd3b1"} Dec 01 16:16:18 crc kubenswrapper[4810]: I1201 16:16:18.982976 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30cf5aff-9c39-4531-8ee6-1284597c7d00","Type":"ContainerStarted","Data":"06fab8219fdf4668dcf89d9f07d815244dc4097c711665bc071ec2661de80ad7"} Dec 01 16:16:18 crc kubenswrapper[4810]: I1201 16:16:18.982985 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30cf5aff-9c39-4531-8ee6-1284597c7d00","Type":"ContainerStarted","Data":"19f62766348c357265510cc3791251351783c7de7e96e035ed79b683c0340489"} Dec 01 16:16:18 crc kubenswrapper[4810]: I1201 16:16:18.984283 4810 generic.go:334] "Generic (PLEG): container finished" podID="9e8daf3c-58ed-4feb-b574-8bda66dd9c79" containerID="a867b5b0fcfc7f3ba0b7f6938f28527326a38e9eee6fae8b4a7b30a8a0f62d67" exitCode=0 Dec 01 16:16:18 crc kubenswrapper[4810]: I1201 16:16:18.984335 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gd5zz" event={"ID":"9e8daf3c-58ed-4feb-b574-8bda66dd9c79","Type":"ContainerDied","Data":"a867b5b0fcfc7f3ba0b7f6938f28527326a38e9eee6fae8b4a7b30a8a0f62d67"} Dec 01 16:16:19 crc kubenswrapper[4810]: I1201 16:16:19.038641 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.038623947 podStartE2EDuration="2.038623947s" podCreationTimestamp="2025-12-01 16:16:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:16:19.013060703 +0000 UTC m=+6144.776570326" watchObservedRunningTime="2025-12-01 16:16:19.038623947 +0000 UTC m=+6144.802133550" Dec 01 16:16:19 crc kubenswrapper[4810]: I1201 16:16:19.341647 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:16:19 crc kubenswrapper[4810]: I1201 16:16:19.418717 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-867b4b5c7c-k9wk9"] Dec 01 16:16:19 crc kubenswrapper[4810]: I1201 16:16:19.418961 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" podUID="59e48d6b-851f-4977-bb5e-9720f697ef6c" containerName="dnsmasq-dns" containerID="cri-o://a1543b61db7c6141aeb0e45fa2812f91cf44f55889a9e387463962a25656553d" gracePeriod=10 Dec 01 16:16:19 crc kubenswrapper[4810]: I1201 16:16:19.555300 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 16:16:19 crc kubenswrapper[4810]: I1201 16:16:19.560221 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 16:16:19 crc kubenswrapper[4810]: I1201 16:16:19.560338 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 16:16:19 crc kubenswrapper[4810]: I1201 16:16:19.939157 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.011638 4810 generic.go:334] "Generic (PLEG): container finished" podID="59e48d6b-851f-4977-bb5e-9720f697ef6c" containerID="a1543b61db7c6141aeb0e45fa2812f91cf44f55889a9e387463962a25656553d" exitCode=0 Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.012672 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.013215 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" event={"ID":"59e48d6b-851f-4977-bb5e-9720f697ef6c","Type":"ContainerDied","Data":"a1543b61db7c6141aeb0e45fa2812f91cf44f55889a9e387463962a25656553d"} Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.013244 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867b4b5c7c-k9wk9" event={"ID":"59e48d6b-851f-4977-bb5e-9720f697ef6c","Type":"ContainerDied","Data":"89cbcd209b6be7d6eb189ddc52c94915d59aef477f10dd3ff4ba0e200e1a2486"} Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.013289 4810 scope.go:117] "RemoveContainer" containerID="a1543b61db7c6141aeb0e45fa2812f91cf44f55889a9e387463962a25656553d" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.028550 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.048403 4810 scope.go:117] "RemoveContainer" containerID="78572dc9cd3125f374819f01e0e59b8ed100f18300c7c8d7e9edb5257ad15a66" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.091832 4810 scope.go:117] "RemoveContainer" containerID="a1543b61db7c6141aeb0e45fa2812f91cf44f55889a9e387463962a25656553d" Dec 01 16:16:20 crc kubenswrapper[4810]: E1201 16:16:20.092275 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1543b61db7c6141aeb0e45fa2812f91cf44f55889a9e387463962a25656553d\": container with ID starting with a1543b61db7c6141aeb0e45fa2812f91cf44f55889a9e387463962a25656553d not found: ID does not exist" containerID="a1543b61db7c6141aeb0e45fa2812f91cf44f55889a9e387463962a25656553d" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.092317 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1543b61db7c6141aeb0e45fa2812f91cf44f55889a9e387463962a25656553d"} err="failed to get container status \"a1543b61db7c6141aeb0e45fa2812f91cf44f55889a9e387463962a25656553d\": rpc error: code = NotFound desc = could not find container \"a1543b61db7c6141aeb0e45fa2812f91cf44f55889a9e387463962a25656553d\": container with ID starting with a1543b61db7c6141aeb0e45fa2812f91cf44f55889a9e387463962a25656553d not found: ID does not exist" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.092345 4810 scope.go:117] "RemoveContainer" containerID="78572dc9cd3125f374819f01e0e59b8ed100f18300c7c8d7e9edb5257ad15a66" Dec 01 16:16:20 crc kubenswrapper[4810]: E1201 16:16:20.092761 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78572dc9cd3125f374819f01e0e59b8ed100f18300c7c8d7e9edb5257ad15a66\": container with ID starting with 78572dc9cd3125f374819f01e0e59b8ed100f18300c7c8d7e9edb5257ad15a66 not found: ID does not exist" containerID="78572dc9cd3125f374819f01e0e59b8ed100f18300c7c8d7e9edb5257ad15a66" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.092785 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78572dc9cd3125f374819f01e0e59b8ed100f18300c7c8d7e9edb5257ad15a66"} err="failed to get container status \"78572dc9cd3125f374819f01e0e59b8ed100f18300c7c8d7e9edb5257ad15a66\": rpc error: code = NotFound desc = could not find container \"78572dc9cd3125f374819f01e0e59b8ed100f18300c7c8d7e9edb5257ad15a66\": container with ID starting with 78572dc9cd3125f374819f01e0e59b8ed100f18300c7c8d7e9edb5257ad15a66 not found: ID does not exist" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.095731 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-config\") pod \"59e48d6b-851f-4977-bb5e-9720f697ef6c\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.095904 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-dns-svc\") pod \"59e48d6b-851f-4977-bb5e-9720f697ef6c\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.095957 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-ovsdbserver-nb\") pod \"59e48d6b-851f-4977-bb5e-9720f697ef6c\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.095977 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-ovsdbserver-sb\") pod \"59e48d6b-851f-4977-bb5e-9720f697ef6c\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.096039 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ps7h\" (UniqueName: \"kubernetes.io/projected/59e48d6b-851f-4977-bb5e-9720f697ef6c-kube-api-access-8ps7h\") pod \"59e48d6b-851f-4977-bb5e-9720f697ef6c\" (UID: \"59e48d6b-851f-4977-bb5e-9720f697ef6c\") " Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.101351 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59e48d6b-851f-4977-bb5e-9720f697ef6c-kube-api-access-8ps7h" (OuterVolumeSpecName: "kube-api-access-8ps7h") pod "59e48d6b-851f-4977-bb5e-9720f697ef6c" (UID: "59e48d6b-851f-4977-bb5e-9720f697ef6c"). InnerVolumeSpecName "kube-api-access-8ps7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.152023 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "59e48d6b-851f-4977-bb5e-9720f697ef6c" (UID: "59e48d6b-851f-4977-bb5e-9720f697ef6c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.157876 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "59e48d6b-851f-4977-bb5e-9720f697ef6c" (UID: "59e48d6b-851f-4977-bb5e-9720f697ef6c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.162660 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "59e48d6b-851f-4977-bb5e-9720f697ef6c" (UID: "59e48d6b-851f-4977-bb5e-9720f697ef6c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.180546 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-config" (OuterVolumeSpecName: "config") pod "59e48d6b-851f-4977-bb5e-9720f697ef6c" (UID: "59e48d6b-851f-4977-bb5e-9720f697ef6c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.198159 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ps7h\" (UniqueName: \"kubernetes.io/projected/59e48d6b-851f-4977-bb5e-9720f697ef6c-kube-api-access-8ps7h\") on node \"crc\" DevicePath \"\"" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.199674 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.199827 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.199854 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.199870 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59e48d6b-851f-4977-bb5e-9720f697ef6c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.357340 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-867b4b5c7c-k9wk9"] Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.366124 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-867b4b5c7c-k9wk9"] Dec 01 16:16:20 crc kubenswrapper[4810]: I1201 16:16:20.503419 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59e48d6b-851f-4977-bb5e-9720f697ef6c" path="/var/lib/kubelet/pods/59e48d6b-851f-4977-bb5e-9720f697ef6c/volumes" Dec 01 16:16:24 crc kubenswrapper[4810]: I1201 16:16:24.051600 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gd5zz" event={"ID":"9e8daf3c-58ed-4feb-b574-8bda66dd9c79","Type":"ContainerStarted","Data":"7f3937cd12a9163825ad2bf7b02f451ce92a5016d602be0051815d390550e6e2"} Dec 01 16:16:25 crc kubenswrapper[4810]: I1201 16:16:25.063556 4810 generic.go:334] "Generic (PLEG): container finished" podID="9e8daf3c-58ed-4feb-b574-8bda66dd9c79" containerID="7f3937cd12a9163825ad2bf7b02f451ce92a5016d602be0051815d390550e6e2" exitCode=0 Dec 01 16:16:25 crc kubenswrapper[4810]: I1201 16:16:25.063709 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gd5zz" event={"ID":"9e8daf3c-58ed-4feb-b574-8bda66dd9c79","Type":"ContainerDied","Data":"7f3937cd12a9163825ad2bf7b02f451ce92a5016d602be0051815d390550e6e2"} Dec 01 16:16:26 crc kubenswrapper[4810]: I1201 16:16:26.073832 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gd5zz" event={"ID":"9e8daf3c-58ed-4feb-b574-8bda66dd9c79","Type":"ContainerStarted","Data":"44632dd9584a65c8447e93ce8db198c67c23e6cb0e0e4361c92967632df77581"} Dec 01 16:16:26 crc kubenswrapper[4810]: I1201 16:16:26.491347 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:16:26 crc kubenswrapper[4810]: E1201 16:16:26.491940 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:16:27 crc kubenswrapper[4810]: I1201 16:16:27.309413 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gd5zz" Dec 01 16:16:27 crc kubenswrapper[4810]: I1201 16:16:27.309463 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gd5zz" Dec 01 16:16:27 crc kubenswrapper[4810]: I1201 16:16:27.370096 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gd5zz" Dec 01 16:16:27 crc kubenswrapper[4810]: I1201 16:16:27.390213 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gd5zz" podStartSLOduration=4.791848742 podStartE2EDuration="11.390195866s" podCreationTimestamp="2025-12-01 16:16:16 +0000 UTC" firstStartedPulling="2025-12-01 16:16:18.98567242 +0000 UTC m=+6144.749182023" lastFinishedPulling="2025-12-01 16:16:25.584019544 +0000 UTC m=+6151.347529147" observedRunningTime="2025-12-01 16:16:26.094363238 +0000 UTC m=+6151.857872851" watchObservedRunningTime="2025-12-01 16:16:27.390195866 +0000 UTC m=+6153.153705469" Dec 01 16:16:27 crc kubenswrapper[4810]: I1201 16:16:27.392652 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 16:16:27 crc kubenswrapper[4810]: I1201 16:16:27.392693 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 16:16:28 crc kubenswrapper[4810]: I1201 16:16:28.405634 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="30cf5aff-9c39-4531-8ee6-1284597c7d00" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.95:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 16:16:28 crc kubenswrapper[4810]: I1201 16:16:28.405754 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="30cf5aff-9c39-4531-8ee6-1284597c7d00" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.95:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 16:16:37 crc kubenswrapper[4810]: I1201 16:16:37.358690 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gd5zz" Dec 01 16:16:37 crc kubenswrapper[4810]: I1201 16:16:37.398737 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 16:16:37 crc kubenswrapper[4810]: I1201 16:16:37.399476 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 16:16:37 crc kubenswrapper[4810]: I1201 16:16:37.410300 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 16:16:37 crc kubenswrapper[4810]: I1201 16:16:37.411102 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 16:16:37 crc kubenswrapper[4810]: I1201 16:16:37.417142 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gd5zz"] Dec 01 16:16:37 crc kubenswrapper[4810]: I1201 16:16:37.477956 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6vngw"] Dec 01 16:16:37 crc kubenswrapper[4810]: I1201 16:16:37.478573 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6vngw" podUID="8ab7c438-ec65-4e1e-ad29-379ba4ddd02e" containerName="registry-server" containerID="cri-o://10e20b6875ec34e50fbf418ef27485b5a3d6468329c93d4f9c33d584f2187a91" gracePeriod=2 Dec 01 16:16:37 crc kubenswrapper[4810]: I1201 16:16:37.969634 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vngw" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.152087 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e-catalog-content\") pod \"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e\" (UID: \"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e\") " Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.152343 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e-utilities\") pod \"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e\" (UID: \"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e\") " Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.152518 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qx8mc\" (UniqueName: \"kubernetes.io/projected/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e-kube-api-access-qx8mc\") pod \"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e\" (UID: \"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e\") " Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.153951 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e-utilities" (OuterVolumeSpecName: "utilities") pod "8ab7c438-ec65-4e1e-ad29-379ba4ddd02e" (UID: "8ab7c438-ec65-4e1e-ad29-379ba4ddd02e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.154360 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.165344 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e-kube-api-access-qx8mc" (OuterVolumeSpecName: "kube-api-access-qx8mc") pod "8ab7c438-ec65-4e1e-ad29-379ba4ddd02e" (UID: "8ab7c438-ec65-4e1e-ad29-379ba4ddd02e"). InnerVolumeSpecName "kube-api-access-qx8mc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.213705 4810 generic.go:334] "Generic (PLEG): container finished" podID="8ab7c438-ec65-4e1e-ad29-379ba4ddd02e" containerID="10e20b6875ec34e50fbf418ef27485b5a3d6468329c93d4f9c33d584f2187a91" exitCode=0 Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.213906 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vngw" event={"ID":"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e","Type":"ContainerDied","Data":"10e20b6875ec34e50fbf418ef27485b5a3d6468329c93d4f9c33d584f2187a91"} Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.213957 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vngw" event={"ID":"8ab7c438-ec65-4e1e-ad29-379ba4ddd02e","Type":"ContainerDied","Data":"6040774dc49082eddf82fbddbf6a72f460f4aab17b015b549797719e97d7c61c"} Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.213978 4810 scope.go:117] "RemoveContainer" containerID="10e20b6875ec34e50fbf418ef27485b5a3d6468329c93d4f9c33d584f2187a91" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.214150 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vngw" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.214883 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.223588 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ab7c438-ec65-4e1e-ad29-379ba4ddd02e" (UID: "8ab7c438-ec65-4e1e-ad29-379ba4ddd02e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.229381 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.246902 4810 scope.go:117] "RemoveContainer" containerID="f5a19743a393a588fd41badaefd677c7892a774a00e76e4929d2cc1d7e21aee6" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.256819 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.256866 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qx8mc\" (UniqueName: \"kubernetes.io/projected/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e-kube-api-access-qx8mc\") on node \"crc\" DevicePath \"\"" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.284417 4810 scope.go:117] "RemoveContainer" containerID="e2af957a9c9d52dbfe9e1c5a479e146e1367c535beb6ea1e1e341afac591cf55" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.328849 4810 scope.go:117] "RemoveContainer" containerID="10e20b6875ec34e50fbf418ef27485b5a3d6468329c93d4f9c33d584f2187a91" Dec 01 16:16:38 crc kubenswrapper[4810]: E1201 16:16:38.329332 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10e20b6875ec34e50fbf418ef27485b5a3d6468329c93d4f9c33d584f2187a91\": container with ID starting with 10e20b6875ec34e50fbf418ef27485b5a3d6468329c93d4f9c33d584f2187a91 not found: ID does not exist" containerID="10e20b6875ec34e50fbf418ef27485b5a3d6468329c93d4f9c33d584f2187a91" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.329373 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10e20b6875ec34e50fbf418ef27485b5a3d6468329c93d4f9c33d584f2187a91"} err="failed to get container status \"10e20b6875ec34e50fbf418ef27485b5a3d6468329c93d4f9c33d584f2187a91\": rpc error: code = NotFound desc = could not find container \"10e20b6875ec34e50fbf418ef27485b5a3d6468329c93d4f9c33d584f2187a91\": container with ID starting with 10e20b6875ec34e50fbf418ef27485b5a3d6468329c93d4f9c33d584f2187a91 not found: ID does not exist" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.329419 4810 scope.go:117] "RemoveContainer" containerID="f5a19743a393a588fd41badaefd677c7892a774a00e76e4929d2cc1d7e21aee6" Dec 01 16:16:38 crc kubenswrapper[4810]: E1201 16:16:38.329810 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5a19743a393a588fd41badaefd677c7892a774a00e76e4929d2cc1d7e21aee6\": container with ID starting with f5a19743a393a588fd41badaefd677c7892a774a00e76e4929d2cc1d7e21aee6 not found: ID does not exist" containerID="f5a19743a393a588fd41badaefd677c7892a774a00e76e4929d2cc1d7e21aee6" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.329845 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5a19743a393a588fd41badaefd677c7892a774a00e76e4929d2cc1d7e21aee6"} err="failed to get container status \"f5a19743a393a588fd41badaefd677c7892a774a00e76e4929d2cc1d7e21aee6\": rpc error: code = NotFound desc = could not find container \"f5a19743a393a588fd41badaefd677c7892a774a00e76e4929d2cc1d7e21aee6\": container with ID starting with f5a19743a393a588fd41badaefd677c7892a774a00e76e4929d2cc1d7e21aee6 not found: ID does not exist" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.329867 4810 scope.go:117] "RemoveContainer" containerID="e2af957a9c9d52dbfe9e1c5a479e146e1367c535beb6ea1e1e341afac591cf55" Dec 01 16:16:38 crc kubenswrapper[4810]: E1201 16:16:38.330145 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2af957a9c9d52dbfe9e1c5a479e146e1367c535beb6ea1e1e341afac591cf55\": container with ID starting with e2af957a9c9d52dbfe9e1c5a479e146e1367c535beb6ea1e1e341afac591cf55 not found: ID does not exist" containerID="e2af957a9c9d52dbfe9e1c5a479e146e1367c535beb6ea1e1e341afac591cf55" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.330174 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2af957a9c9d52dbfe9e1c5a479e146e1367c535beb6ea1e1e341afac591cf55"} err="failed to get container status \"e2af957a9c9d52dbfe9e1c5a479e146e1367c535beb6ea1e1e341afac591cf55\": rpc error: code = NotFound desc = could not find container \"e2af957a9c9d52dbfe9e1c5a479e146e1367c535beb6ea1e1e341afac591cf55\": container with ID starting with e2af957a9c9d52dbfe9e1c5a479e146e1367c535beb6ea1e1e341afac591cf55 not found: ID does not exist" Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.538350 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6vngw"] Dec 01 16:16:38 crc kubenswrapper[4810]: I1201 16:16:38.547570 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6vngw"] Dec 01 16:16:40 crc kubenswrapper[4810]: I1201 16:16:40.492011 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:16:40 crc kubenswrapper[4810]: E1201 16:16:40.492808 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:16:40 crc kubenswrapper[4810]: I1201 16:16:40.503445 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ab7c438-ec65-4e1e-ad29-379ba4ddd02e" path="/var/lib/kubelet/pods/8ab7c438-ec65-4e1e-ad29-379ba4ddd02e/volumes" Dec 01 16:16:54 crc kubenswrapper[4810]: I1201 16:16:54.498065 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:16:54 crc kubenswrapper[4810]: E1201 16:16:54.498934 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:16:58 crc kubenswrapper[4810]: I1201 16:16:58.890981 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4kw8j"] Dec 01 16:16:58 crc kubenswrapper[4810]: E1201 16:16:58.891846 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59e48d6b-851f-4977-bb5e-9720f697ef6c" containerName="init" Dec 01 16:16:58 crc kubenswrapper[4810]: I1201 16:16:58.891858 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="59e48d6b-851f-4977-bb5e-9720f697ef6c" containerName="init" Dec 01 16:16:58 crc kubenswrapper[4810]: E1201 16:16:58.891888 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ab7c438-ec65-4e1e-ad29-379ba4ddd02e" containerName="extract-content" Dec 01 16:16:58 crc kubenswrapper[4810]: I1201 16:16:58.891894 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ab7c438-ec65-4e1e-ad29-379ba4ddd02e" containerName="extract-content" Dec 01 16:16:58 crc kubenswrapper[4810]: E1201 16:16:58.891912 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59e48d6b-851f-4977-bb5e-9720f697ef6c" containerName="dnsmasq-dns" Dec 01 16:16:58 crc kubenswrapper[4810]: I1201 16:16:58.891920 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="59e48d6b-851f-4977-bb5e-9720f697ef6c" containerName="dnsmasq-dns" Dec 01 16:16:58 crc kubenswrapper[4810]: E1201 16:16:58.891936 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ab7c438-ec65-4e1e-ad29-379ba4ddd02e" containerName="extract-utilities" Dec 01 16:16:58 crc kubenswrapper[4810]: I1201 16:16:58.891941 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ab7c438-ec65-4e1e-ad29-379ba4ddd02e" containerName="extract-utilities" Dec 01 16:16:58 crc kubenswrapper[4810]: E1201 16:16:58.891947 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ab7c438-ec65-4e1e-ad29-379ba4ddd02e" containerName="registry-server" Dec 01 16:16:58 crc kubenswrapper[4810]: I1201 16:16:58.891953 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ab7c438-ec65-4e1e-ad29-379ba4ddd02e" containerName="registry-server" Dec 01 16:16:58 crc kubenswrapper[4810]: I1201 16:16:58.892129 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ab7c438-ec65-4e1e-ad29-379ba4ddd02e" containerName="registry-server" Dec 01 16:16:58 crc kubenswrapper[4810]: I1201 16:16:58.892147 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="59e48d6b-851f-4977-bb5e-9720f697ef6c" containerName="dnsmasq-dns" Dec 01 16:16:58 crc kubenswrapper[4810]: I1201 16:16:58.892748 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:58 crc kubenswrapper[4810]: I1201 16:16:58.894729 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 01 16:16:58 crc kubenswrapper[4810]: I1201 16:16:58.897107 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-tcbfz" Dec 01 16:16:58 crc kubenswrapper[4810]: I1201 16:16:58.902321 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 01 16:16:58 crc kubenswrapper[4810]: I1201 16:16:58.907148 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4kw8j"] Dec 01 16:16:58 crc kubenswrapper[4810]: I1201 16:16:58.949905 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-rq44f"] Dec 01 16:16:58 crc kubenswrapper[4810]: I1201 16:16:58.952822 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:58 crc kubenswrapper[4810]: I1201 16:16:58.986778 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-rq44f"] Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.049664 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8a092636-4ef4-4e56-91ad-cf0518f2c88d-var-log-ovn\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.049864 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6340cc9b-5129-437b-90fe-8b031224943c-etc-ovs\") pod \"ovn-controller-ovs-rq44f\" (UID: \"6340cc9b-5129-437b-90fe-8b031224943c\") " pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.049978 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfzmj\" (UniqueName: \"kubernetes.io/projected/6340cc9b-5129-437b-90fe-8b031224943c-kube-api-access-xfzmj\") pod \"ovn-controller-ovs-rq44f\" (UID: \"6340cc9b-5129-437b-90fe-8b031224943c\") " pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.050031 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6340cc9b-5129-437b-90fe-8b031224943c-var-run\") pod \"ovn-controller-ovs-rq44f\" (UID: \"6340cc9b-5129-437b-90fe-8b031224943c\") " pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.050068 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8a092636-4ef4-4e56-91ad-cf0518f2c88d-var-run-ovn\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.050157 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6340cc9b-5129-437b-90fe-8b031224943c-var-lib\") pod \"ovn-controller-ovs-rq44f\" (UID: \"6340cc9b-5129-437b-90fe-8b031224943c\") " pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.050273 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5dbw\" (UniqueName: \"kubernetes.io/projected/8a092636-4ef4-4e56-91ad-cf0518f2c88d-kube-api-access-k5dbw\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.050369 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a092636-4ef4-4e56-91ad-cf0518f2c88d-scripts\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.050399 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8a092636-4ef4-4e56-91ad-cf0518f2c88d-var-run\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.050450 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6340cc9b-5129-437b-90fe-8b031224943c-var-log\") pod \"ovn-controller-ovs-rq44f\" (UID: \"6340cc9b-5129-437b-90fe-8b031224943c\") " pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.050566 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a092636-4ef4-4e56-91ad-cf0518f2c88d-combined-ca-bundle\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.050649 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a092636-4ef4-4e56-91ad-cf0518f2c88d-ovn-controller-tls-certs\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.050696 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6340cc9b-5129-437b-90fe-8b031224943c-scripts\") pod \"ovn-controller-ovs-rq44f\" (UID: \"6340cc9b-5129-437b-90fe-8b031224943c\") " pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.151788 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6340cc9b-5129-437b-90fe-8b031224943c-var-lib\") pod \"ovn-controller-ovs-rq44f\" (UID: \"6340cc9b-5129-437b-90fe-8b031224943c\") " pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.151855 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5dbw\" (UniqueName: \"kubernetes.io/projected/8a092636-4ef4-4e56-91ad-cf0518f2c88d-kube-api-access-k5dbw\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.151894 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a092636-4ef4-4e56-91ad-cf0518f2c88d-scripts\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.151919 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8a092636-4ef4-4e56-91ad-cf0518f2c88d-var-run\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.151934 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6340cc9b-5129-437b-90fe-8b031224943c-var-log\") pod \"ovn-controller-ovs-rq44f\" (UID: \"6340cc9b-5129-437b-90fe-8b031224943c\") " pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.151979 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a092636-4ef4-4e56-91ad-cf0518f2c88d-combined-ca-bundle\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.152023 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a092636-4ef4-4e56-91ad-cf0518f2c88d-ovn-controller-tls-certs\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.152044 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6340cc9b-5129-437b-90fe-8b031224943c-scripts\") pod \"ovn-controller-ovs-rq44f\" (UID: \"6340cc9b-5129-437b-90fe-8b031224943c\") " pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.152062 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8a092636-4ef4-4e56-91ad-cf0518f2c88d-var-log-ovn\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.152099 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6340cc9b-5129-437b-90fe-8b031224943c-etc-ovs\") pod \"ovn-controller-ovs-rq44f\" (UID: \"6340cc9b-5129-437b-90fe-8b031224943c\") " pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.152133 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfzmj\" (UniqueName: \"kubernetes.io/projected/6340cc9b-5129-437b-90fe-8b031224943c-kube-api-access-xfzmj\") pod \"ovn-controller-ovs-rq44f\" (UID: \"6340cc9b-5129-437b-90fe-8b031224943c\") " pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.152156 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6340cc9b-5129-437b-90fe-8b031224943c-var-run\") pod \"ovn-controller-ovs-rq44f\" (UID: \"6340cc9b-5129-437b-90fe-8b031224943c\") " pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.152179 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8a092636-4ef4-4e56-91ad-cf0518f2c88d-var-run-ovn\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.152314 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8a092636-4ef4-4e56-91ad-cf0518f2c88d-var-run-ovn\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.153038 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6340cc9b-5129-437b-90fe-8b031224943c-etc-ovs\") pod \"ovn-controller-ovs-rq44f\" (UID: \"6340cc9b-5129-437b-90fe-8b031224943c\") " pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.153215 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6340cc9b-5129-437b-90fe-8b031224943c-var-run\") pod \"ovn-controller-ovs-rq44f\" (UID: \"6340cc9b-5129-437b-90fe-8b031224943c\") " pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.153623 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8a092636-4ef4-4e56-91ad-cf0518f2c88d-var-log-ovn\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.153665 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8a092636-4ef4-4e56-91ad-cf0518f2c88d-var-run\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.153705 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6340cc9b-5129-437b-90fe-8b031224943c-var-log\") pod \"ovn-controller-ovs-rq44f\" (UID: \"6340cc9b-5129-437b-90fe-8b031224943c\") " pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.152144 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6340cc9b-5129-437b-90fe-8b031224943c-var-lib\") pod \"ovn-controller-ovs-rq44f\" (UID: \"6340cc9b-5129-437b-90fe-8b031224943c\") " pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.154745 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a092636-4ef4-4e56-91ad-cf0518f2c88d-scripts\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.162020 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a092636-4ef4-4e56-91ad-cf0518f2c88d-combined-ca-bundle\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.162127 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a092636-4ef4-4e56-91ad-cf0518f2c88d-ovn-controller-tls-certs\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.165219 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6340cc9b-5129-437b-90fe-8b031224943c-scripts\") pod \"ovn-controller-ovs-rq44f\" (UID: \"6340cc9b-5129-437b-90fe-8b031224943c\") " pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.170314 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfzmj\" (UniqueName: \"kubernetes.io/projected/6340cc9b-5129-437b-90fe-8b031224943c-kube-api-access-xfzmj\") pod \"ovn-controller-ovs-rq44f\" (UID: \"6340cc9b-5129-437b-90fe-8b031224943c\") " pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.183788 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5dbw\" (UniqueName: \"kubernetes.io/projected/8a092636-4ef4-4e56-91ad-cf0518f2c88d-kube-api-access-k5dbw\") pod \"ovn-controller-4kw8j\" (UID: \"8a092636-4ef4-4e56-91ad-cf0518f2c88d\") " pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.254757 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kw8j" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.277359 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:16:59 crc kubenswrapper[4810]: I1201 16:16:59.719635 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4kw8j"] Dec 01 16:17:00 crc kubenswrapper[4810]: I1201 16:17:00.172686 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-rq44f"] Dec 01 16:17:00 crc kubenswrapper[4810]: I1201 16:17:00.398277 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rq44f" event={"ID":"6340cc9b-5129-437b-90fe-8b031224943c","Type":"ContainerStarted","Data":"e5be85de6525bb7ab61287a94b5613db13821c7dcada8dc2096f31efa0036810"} Dec 01 16:17:00 crc kubenswrapper[4810]: I1201 16:17:00.400183 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kw8j" event={"ID":"8a092636-4ef4-4e56-91ad-cf0518f2c88d","Type":"ContainerStarted","Data":"617356f1885ffb4ddd282f0a2146342ef3ef6d765d7bced65aa1f03f26590a6f"} Dec 01 16:17:00 crc kubenswrapper[4810]: I1201 16:17:00.400209 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kw8j" event={"ID":"8a092636-4ef4-4e56-91ad-cf0518f2c88d","Type":"ContainerStarted","Data":"2f6589d52b3f36e685064813cf7e817829bca6f877d75b1f2ef5c4072525ac28"} Dec 01 16:17:00 crc kubenswrapper[4810]: I1201 16:17:00.401099 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-4kw8j" Dec 01 16:17:00 crc kubenswrapper[4810]: I1201 16:17:00.420977 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-4kw8j" podStartSLOduration=2.420960199 podStartE2EDuration="2.420960199s" podCreationTimestamp="2025-12-01 16:16:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:17:00.418250966 +0000 UTC m=+6186.181760579" watchObservedRunningTime="2025-12-01 16:17:00.420960199 +0000 UTC m=+6186.184469802" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.077422 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-mggf6"] Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.079214 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-mggf6" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.090324 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-mggf6"] Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.192613 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wttz\" (UniqueName: \"kubernetes.io/projected/7db4342d-31f0-48c3-8c8c-eccc2a1324ad-kube-api-access-4wttz\") pod \"octavia-db-create-mggf6\" (UID: \"7db4342d-31f0-48c3-8c8c-eccc2a1324ad\") " pod="openstack/octavia-db-create-mggf6" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.192673 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7db4342d-31f0-48c3-8c8c-eccc2a1324ad-operator-scripts\") pod \"octavia-db-create-mggf6\" (UID: \"7db4342d-31f0-48c3-8c8c-eccc2a1324ad\") " pod="openstack/octavia-db-create-mggf6" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.295062 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wttz\" (UniqueName: \"kubernetes.io/projected/7db4342d-31f0-48c3-8c8c-eccc2a1324ad-kube-api-access-4wttz\") pod \"octavia-db-create-mggf6\" (UID: \"7db4342d-31f0-48c3-8c8c-eccc2a1324ad\") " pod="openstack/octavia-db-create-mggf6" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.295139 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7db4342d-31f0-48c3-8c8c-eccc2a1324ad-operator-scripts\") pod \"octavia-db-create-mggf6\" (UID: \"7db4342d-31f0-48c3-8c8c-eccc2a1324ad\") " pod="openstack/octavia-db-create-mggf6" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.296101 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7db4342d-31f0-48c3-8c8c-eccc2a1324ad-operator-scripts\") pod \"octavia-db-create-mggf6\" (UID: \"7db4342d-31f0-48c3-8c8c-eccc2a1324ad\") " pod="openstack/octavia-db-create-mggf6" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.315709 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wttz\" (UniqueName: \"kubernetes.io/projected/7db4342d-31f0-48c3-8c8c-eccc2a1324ad-kube-api-access-4wttz\") pod \"octavia-db-create-mggf6\" (UID: \"7db4342d-31f0-48c3-8c8c-eccc2a1324ad\") " pod="openstack/octavia-db-create-mggf6" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.405008 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-mggf6" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.410515 4810 generic.go:334] "Generic (PLEG): container finished" podID="6340cc9b-5129-437b-90fe-8b031224943c" containerID="f4f360a7cd98b375ede49b1cf8c3abcc9b747596dbb814b682b277b06591ccb1" exitCode=0 Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.410729 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rq44f" event={"ID":"6340cc9b-5129-437b-90fe-8b031224943c","Type":"ContainerDied","Data":"f4f360a7cd98b375ede49b1cf8c3abcc9b747596dbb814b682b277b06591ccb1"} Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.487182 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-tmvcm"] Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.488740 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.492389 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.500523 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-tmvcm"] Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.606134 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/59c20ba6-6bec-443f-911d-02d9139ac260-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-tmvcm\" (UID: \"59c20ba6-6bec-443f-911d-02d9139ac260\") " pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.606181 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59c20ba6-6bec-443f-911d-02d9139ac260-combined-ca-bundle\") pod \"ovn-controller-metrics-tmvcm\" (UID: \"59c20ba6-6bec-443f-911d-02d9139ac260\") " pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.606206 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59c20ba6-6bec-443f-911d-02d9139ac260-config\") pod \"ovn-controller-metrics-tmvcm\" (UID: \"59c20ba6-6bec-443f-911d-02d9139ac260\") " pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.606233 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqprd\" (UniqueName: \"kubernetes.io/projected/59c20ba6-6bec-443f-911d-02d9139ac260-kube-api-access-cqprd\") pod \"ovn-controller-metrics-tmvcm\" (UID: \"59c20ba6-6bec-443f-911d-02d9139ac260\") " pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.606342 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/59c20ba6-6bec-443f-911d-02d9139ac260-ovs-rundir\") pod \"ovn-controller-metrics-tmvcm\" (UID: \"59c20ba6-6bec-443f-911d-02d9139ac260\") " pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.606816 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/59c20ba6-6bec-443f-911d-02d9139ac260-ovn-rundir\") pod \"ovn-controller-metrics-tmvcm\" (UID: \"59c20ba6-6bec-443f-911d-02d9139ac260\") " pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.709524 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/59c20ba6-6bec-443f-911d-02d9139ac260-ovn-rundir\") pod \"ovn-controller-metrics-tmvcm\" (UID: \"59c20ba6-6bec-443f-911d-02d9139ac260\") " pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.709881 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/59c20ba6-6bec-443f-911d-02d9139ac260-ovn-rundir\") pod \"ovn-controller-metrics-tmvcm\" (UID: \"59c20ba6-6bec-443f-911d-02d9139ac260\") " pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.709988 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/59c20ba6-6bec-443f-911d-02d9139ac260-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-tmvcm\" (UID: \"59c20ba6-6bec-443f-911d-02d9139ac260\") " pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.710072 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59c20ba6-6bec-443f-911d-02d9139ac260-combined-ca-bundle\") pod \"ovn-controller-metrics-tmvcm\" (UID: \"59c20ba6-6bec-443f-911d-02d9139ac260\") " pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.710104 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59c20ba6-6bec-443f-911d-02d9139ac260-config\") pod \"ovn-controller-metrics-tmvcm\" (UID: \"59c20ba6-6bec-443f-911d-02d9139ac260\") " pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.710145 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqprd\" (UniqueName: \"kubernetes.io/projected/59c20ba6-6bec-443f-911d-02d9139ac260-kube-api-access-cqprd\") pod \"ovn-controller-metrics-tmvcm\" (UID: \"59c20ba6-6bec-443f-911d-02d9139ac260\") " pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.710225 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/59c20ba6-6bec-443f-911d-02d9139ac260-ovs-rundir\") pod \"ovn-controller-metrics-tmvcm\" (UID: \"59c20ba6-6bec-443f-911d-02d9139ac260\") " pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.710321 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/59c20ba6-6bec-443f-911d-02d9139ac260-ovs-rundir\") pod \"ovn-controller-metrics-tmvcm\" (UID: \"59c20ba6-6bec-443f-911d-02d9139ac260\") " pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.711064 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59c20ba6-6bec-443f-911d-02d9139ac260-config\") pod \"ovn-controller-metrics-tmvcm\" (UID: \"59c20ba6-6bec-443f-911d-02d9139ac260\") " pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.716219 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/59c20ba6-6bec-443f-911d-02d9139ac260-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-tmvcm\" (UID: \"59c20ba6-6bec-443f-911d-02d9139ac260\") " pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.717103 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59c20ba6-6bec-443f-911d-02d9139ac260-combined-ca-bundle\") pod \"ovn-controller-metrics-tmvcm\" (UID: \"59c20ba6-6bec-443f-911d-02d9139ac260\") " pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.737290 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqprd\" (UniqueName: \"kubernetes.io/projected/59c20ba6-6bec-443f-911d-02d9139ac260-kube-api-access-cqprd\") pod \"ovn-controller-metrics-tmvcm\" (UID: \"59c20ba6-6bec-443f-911d-02d9139ac260\") " pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.901139 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-tmvcm" Dec 01 16:17:01 crc kubenswrapper[4810]: I1201 16:17:01.956187 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-mggf6"] Dec 01 16:17:01 crc kubenswrapper[4810]: W1201 16:17:01.961317 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7db4342d_31f0_48c3_8c8c_eccc2a1324ad.slice/crio-8faba87ecf3786f4708255a87d3f39a4880acc0aa4c2462155118274f8ca7aa3 WatchSource:0}: Error finding container 8faba87ecf3786f4708255a87d3f39a4880acc0aa4c2462155118274f8ca7aa3: Status 404 returned error can't find the container with id 8faba87ecf3786f4708255a87d3f39a4880acc0aa4c2462155118274f8ca7aa3 Dec 01 16:17:02 crc kubenswrapper[4810]: I1201 16:17:02.404608 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-tmvcm"] Dec 01 16:17:02 crc kubenswrapper[4810]: I1201 16:17:02.426296 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rq44f" event={"ID":"6340cc9b-5129-437b-90fe-8b031224943c","Type":"ContainerStarted","Data":"8ce853387d0333de87c0b61fc355e59b2eaae06e4aa9b38e9aac8de74eb2546a"} Dec 01 16:17:02 crc kubenswrapper[4810]: I1201 16:17:02.426344 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rq44f" event={"ID":"6340cc9b-5129-437b-90fe-8b031224943c","Type":"ContainerStarted","Data":"8bf853060efd8542f4b0b79151b79934356b8c550f47476d16c3e4f3d394ad14"} Dec 01 16:17:02 crc kubenswrapper[4810]: I1201 16:17:02.426721 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:17:02 crc kubenswrapper[4810]: I1201 16:17:02.426891 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:17:02 crc kubenswrapper[4810]: I1201 16:17:02.429937 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-tmvcm" event={"ID":"59c20ba6-6bec-443f-911d-02d9139ac260","Type":"ContainerStarted","Data":"a57622358d6ad60e9b0620bbe3462ac3650c0bcd7b489e3ab9b38ca62752dc82"} Dec 01 16:17:02 crc kubenswrapper[4810]: I1201 16:17:02.431431 4810 generic.go:334] "Generic (PLEG): container finished" podID="7db4342d-31f0-48c3-8c8c-eccc2a1324ad" containerID="f3878656db426ae1896e62188b3c9738ffa6c6a41583f5419dd40bb9018b9ad2" exitCode=0 Dec 01 16:17:02 crc kubenswrapper[4810]: I1201 16:17:02.432211 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-mggf6" event={"ID":"7db4342d-31f0-48c3-8c8c-eccc2a1324ad","Type":"ContainerDied","Data":"f3878656db426ae1896e62188b3c9738ffa6c6a41583f5419dd40bb9018b9ad2"} Dec 01 16:17:02 crc kubenswrapper[4810]: I1201 16:17:02.432236 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-mggf6" event={"ID":"7db4342d-31f0-48c3-8c8c-eccc2a1324ad","Type":"ContainerStarted","Data":"8faba87ecf3786f4708255a87d3f39a4880acc0aa4c2462155118274f8ca7aa3"} Dec 01 16:17:02 crc kubenswrapper[4810]: I1201 16:17:02.449540 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-rq44f" podStartSLOduration=4.449522669 podStartE2EDuration="4.449522669s" podCreationTimestamp="2025-12-01 16:16:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:17:02.446055265 +0000 UTC m=+6188.209564888" watchObservedRunningTime="2025-12-01 16:17:02.449522669 +0000 UTC m=+6188.213032272" Dec 01 16:17:03 crc kubenswrapper[4810]: I1201 16:17:03.110127 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-1e8e-account-create-update-9mvxm"] Dec 01 16:17:03 crc kubenswrapper[4810]: I1201 16:17:03.112279 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1e8e-account-create-update-9mvxm" Dec 01 16:17:03 crc kubenswrapper[4810]: I1201 16:17:03.114119 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Dec 01 16:17:03 crc kubenswrapper[4810]: I1201 16:17:03.123085 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-1e8e-account-create-update-9mvxm"] Dec 01 16:17:03 crc kubenswrapper[4810]: I1201 16:17:03.237761 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c02cfaaf-96b5-444f-a03f-e4421bf3b8a1-operator-scripts\") pod \"octavia-1e8e-account-create-update-9mvxm\" (UID: \"c02cfaaf-96b5-444f-a03f-e4421bf3b8a1\") " pod="openstack/octavia-1e8e-account-create-update-9mvxm" Dec 01 16:17:03 crc kubenswrapper[4810]: I1201 16:17:03.237827 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cksz8\" (UniqueName: \"kubernetes.io/projected/c02cfaaf-96b5-444f-a03f-e4421bf3b8a1-kube-api-access-cksz8\") pod \"octavia-1e8e-account-create-update-9mvxm\" (UID: \"c02cfaaf-96b5-444f-a03f-e4421bf3b8a1\") " pod="openstack/octavia-1e8e-account-create-update-9mvxm" Dec 01 16:17:03 crc kubenswrapper[4810]: I1201 16:17:03.340149 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cksz8\" (UniqueName: \"kubernetes.io/projected/c02cfaaf-96b5-444f-a03f-e4421bf3b8a1-kube-api-access-cksz8\") pod \"octavia-1e8e-account-create-update-9mvxm\" (UID: \"c02cfaaf-96b5-444f-a03f-e4421bf3b8a1\") " pod="openstack/octavia-1e8e-account-create-update-9mvxm" Dec 01 16:17:03 crc kubenswrapper[4810]: I1201 16:17:03.340645 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c02cfaaf-96b5-444f-a03f-e4421bf3b8a1-operator-scripts\") pod \"octavia-1e8e-account-create-update-9mvxm\" (UID: \"c02cfaaf-96b5-444f-a03f-e4421bf3b8a1\") " pod="openstack/octavia-1e8e-account-create-update-9mvxm" Dec 01 16:17:03 crc kubenswrapper[4810]: I1201 16:17:03.341554 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c02cfaaf-96b5-444f-a03f-e4421bf3b8a1-operator-scripts\") pod \"octavia-1e8e-account-create-update-9mvxm\" (UID: \"c02cfaaf-96b5-444f-a03f-e4421bf3b8a1\") " pod="openstack/octavia-1e8e-account-create-update-9mvxm" Dec 01 16:17:03 crc kubenswrapper[4810]: I1201 16:17:03.358250 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cksz8\" (UniqueName: \"kubernetes.io/projected/c02cfaaf-96b5-444f-a03f-e4421bf3b8a1-kube-api-access-cksz8\") pod \"octavia-1e8e-account-create-update-9mvxm\" (UID: \"c02cfaaf-96b5-444f-a03f-e4421bf3b8a1\") " pod="openstack/octavia-1e8e-account-create-update-9mvxm" Dec 01 16:17:03 crc kubenswrapper[4810]: I1201 16:17:03.436866 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1e8e-account-create-update-9mvxm" Dec 01 16:17:03 crc kubenswrapper[4810]: I1201 16:17:03.446593 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-tmvcm" event={"ID":"59c20ba6-6bec-443f-911d-02d9139ac260","Type":"ContainerStarted","Data":"e98777cf8f6d692881c524c21674dfffe0aa25c75ad225d46deb8fd7ed4e8459"} Dec 01 16:17:03 crc kubenswrapper[4810]: I1201 16:17:03.496686 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-tmvcm" podStartSLOduration=2.496664226 podStartE2EDuration="2.496664226s" podCreationTimestamp="2025-12-01 16:17:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:17:03.460847143 +0000 UTC m=+6189.224356746" watchObservedRunningTime="2025-12-01 16:17:03.496664226 +0000 UTC m=+6189.260173829" Dec 01 16:17:03 crc kubenswrapper[4810]: I1201 16:17:03.931381 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-mggf6" Dec 01 16:17:04 crc kubenswrapper[4810]: I1201 16:17:04.058290 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7db4342d-31f0-48c3-8c8c-eccc2a1324ad-operator-scripts\") pod \"7db4342d-31f0-48c3-8c8c-eccc2a1324ad\" (UID: \"7db4342d-31f0-48c3-8c8c-eccc2a1324ad\") " Dec 01 16:17:04 crc kubenswrapper[4810]: I1201 16:17:04.058774 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wttz\" (UniqueName: \"kubernetes.io/projected/7db4342d-31f0-48c3-8c8c-eccc2a1324ad-kube-api-access-4wttz\") pod \"7db4342d-31f0-48c3-8c8c-eccc2a1324ad\" (UID: \"7db4342d-31f0-48c3-8c8c-eccc2a1324ad\") " Dec 01 16:17:04 crc kubenswrapper[4810]: I1201 16:17:04.060194 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7db4342d-31f0-48c3-8c8c-eccc2a1324ad-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7db4342d-31f0-48c3-8c8c-eccc2a1324ad" (UID: "7db4342d-31f0-48c3-8c8c-eccc2a1324ad"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:17:04 crc kubenswrapper[4810]: I1201 16:17:04.065648 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7db4342d-31f0-48c3-8c8c-eccc2a1324ad-kube-api-access-4wttz" (OuterVolumeSpecName: "kube-api-access-4wttz") pod "7db4342d-31f0-48c3-8c8c-eccc2a1324ad" (UID: "7db4342d-31f0-48c3-8c8c-eccc2a1324ad"). InnerVolumeSpecName "kube-api-access-4wttz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:17:04 crc kubenswrapper[4810]: I1201 16:17:04.079854 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-1e8e-account-create-update-9mvxm"] Dec 01 16:17:04 crc kubenswrapper[4810]: W1201 16:17:04.080682 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc02cfaaf_96b5_444f_a03f_e4421bf3b8a1.slice/crio-0e7fa315cc256003e73a1bab9efc3cba30d88eefabea2f2531e909694639c0eb WatchSource:0}: Error finding container 0e7fa315cc256003e73a1bab9efc3cba30d88eefabea2f2531e909694639c0eb: Status 404 returned error can't find the container with id 0e7fa315cc256003e73a1bab9efc3cba30d88eefabea2f2531e909694639c0eb Dec 01 16:17:04 crc kubenswrapper[4810]: I1201 16:17:04.160940 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7db4342d-31f0-48c3-8c8c-eccc2a1324ad-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:04 crc kubenswrapper[4810]: I1201 16:17:04.160984 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wttz\" (UniqueName: \"kubernetes.io/projected/7db4342d-31f0-48c3-8c8c-eccc2a1324ad-kube-api-access-4wttz\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:04 crc kubenswrapper[4810]: I1201 16:17:04.457071 4810 generic.go:334] "Generic (PLEG): container finished" podID="c02cfaaf-96b5-444f-a03f-e4421bf3b8a1" containerID="cdb7bdf10e9e9ee7431f3d188d36e0aab4e07e9981b9931f20479e6933e688f4" exitCode=0 Dec 01 16:17:04 crc kubenswrapper[4810]: I1201 16:17:04.457119 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-1e8e-account-create-update-9mvxm" event={"ID":"c02cfaaf-96b5-444f-a03f-e4421bf3b8a1","Type":"ContainerDied","Data":"cdb7bdf10e9e9ee7431f3d188d36e0aab4e07e9981b9931f20479e6933e688f4"} Dec 01 16:17:04 crc kubenswrapper[4810]: I1201 16:17:04.457156 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-1e8e-account-create-update-9mvxm" event={"ID":"c02cfaaf-96b5-444f-a03f-e4421bf3b8a1","Type":"ContainerStarted","Data":"0e7fa315cc256003e73a1bab9efc3cba30d88eefabea2f2531e909694639c0eb"} Dec 01 16:17:04 crc kubenswrapper[4810]: I1201 16:17:04.459555 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-mggf6" event={"ID":"7db4342d-31f0-48c3-8c8c-eccc2a1324ad","Type":"ContainerDied","Data":"8faba87ecf3786f4708255a87d3f39a4880acc0aa4c2462155118274f8ca7aa3"} Dec 01 16:17:04 crc kubenswrapper[4810]: I1201 16:17:04.459585 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8faba87ecf3786f4708255a87d3f39a4880acc0aa4c2462155118274f8ca7aa3" Dec 01 16:17:04 crc kubenswrapper[4810]: I1201 16:17:04.459598 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-mggf6" Dec 01 16:17:05 crc kubenswrapper[4810]: I1201 16:17:05.491652 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:17:05 crc kubenswrapper[4810]: E1201 16:17:05.492407 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:17:05 crc kubenswrapper[4810]: I1201 16:17:05.842107 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1e8e-account-create-update-9mvxm" Dec 01 16:17:05 crc kubenswrapper[4810]: I1201 16:17:05.999027 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cksz8\" (UniqueName: \"kubernetes.io/projected/c02cfaaf-96b5-444f-a03f-e4421bf3b8a1-kube-api-access-cksz8\") pod \"c02cfaaf-96b5-444f-a03f-e4421bf3b8a1\" (UID: \"c02cfaaf-96b5-444f-a03f-e4421bf3b8a1\") " Dec 01 16:17:05 crc kubenswrapper[4810]: I1201 16:17:05.999078 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c02cfaaf-96b5-444f-a03f-e4421bf3b8a1-operator-scripts\") pod \"c02cfaaf-96b5-444f-a03f-e4421bf3b8a1\" (UID: \"c02cfaaf-96b5-444f-a03f-e4421bf3b8a1\") " Dec 01 16:17:06 crc kubenswrapper[4810]: I1201 16:17:05.999568 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c02cfaaf-96b5-444f-a03f-e4421bf3b8a1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c02cfaaf-96b5-444f-a03f-e4421bf3b8a1" (UID: "c02cfaaf-96b5-444f-a03f-e4421bf3b8a1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:17:06 crc kubenswrapper[4810]: I1201 16:17:05.999846 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c02cfaaf-96b5-444f-a03f-e4421bf3b8a1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:06 crc kubenswrapper[4810]: I1201 16:17:06.007218 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c02cfaaf-96b5-444f-a03f-e4421bf3b8a1-kube-api-access-cksz8" (OuterVolumeSpecName: "kube-api-access-cksz8") pod "c02cfaaf-96b5-444f-a03f-e4421bf3b8a1" (UID: "c02cfaaf-96b5-444f-a03f-e4421bf3b8a1"). InnerVolumeSpecName "kube-api-access-cksz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:17:06 crc kubenswrapper[4810]: I1201 16:17:06.101461 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cksz8\" (UniqueName: \"kubernetes.io/projected/c02cfaaf-96b5-444f-a03f-e4421bf3b8a1-kube-api-access-cksz8\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:06 crc kubenswrapper[4810]: I1201 16:17:06.483251 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-1e8e-account-create-update-9mvxm" event={"ID":"c02cfaaf-96b5-444f-a03f-e4421bf3b8a1","Type":"ContainerDied","Data":"0e7fa315cc256003e73a1bab9efc3cba30d88eefabea2f2531e909694639c0eb"} Dec 01 16:17:06 crc kubenswrapper[4810]: I1201 16:17:06.483289 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e7fa315cc256003e73a1bab9efc3cba30d88eefabea2f2531e909694639c0eb" Dec 01 16:17:06 crc kubenswrapper[4810]: I1201 16:17:06.483295 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1e8e-account-create-update-9mvxm" Dec 01 16:17:09 crc kubenswrapper[4810]: I1201 16:17:09.129018 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-pxplj"] Dec 01 16:17:09 crc kubenswrapper[4810]: E1201 16:17:09.129792 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7db4342d-31f0-48c3-8c8c-eccc2a1324ad" containerName="mariadb-database-create" Dec 01 16:17:09 crc kubenswrapper[4810]: I1201 16:17:09.129808 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="7db4342d-31f0-48c3-8c8c-eccc2a1324ad" containerName="mariadb-database-create" Dec 01 16:17:09 crc kubenswrapper[4810]: E1201 16:17:09.129826 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c02cfaaf-96b5-444f-a03f-e4421bf3b8a1" containerName="mariadb-account-create-update" Dec 01 16:17:09 crc kubenswrapper[4810]: I1201 16:17:09.129834 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c02cfaaf-96b5-444f-a03f-e4421bf3b8a1" containerName="mariadb-account-create-update" Dec 01 16:17:09 crc kubenswrapper[4810]: I1201 16:17:09.130076 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c02cfaaf-96b5-444f-a03f-e4421bf3b8a1" containerName="mariadb-account-create-update" Dec 01 16:17:09 crc kubenswrapper[4810]: I1201 16:17:09.130100 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="7db4342d-31f0-48c3-8c8c-eccc2a1324ad" containerName="mariadb-database-create" Dec 01 16:17:09 crc kubenswrapper[4810]: I1201 16:17:09.130939 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-pxplj" Dec 01 16:17:09 crc kubenswrapper[4810]: I1201 16:17:09.140945 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-pxplj"] Dec 01 16:17:09 crc kubenswrapper[4810]: I1201 16:17:09.261687 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g27w\" (UniqueName: \"kubernetes.io/projected/9edab071-f11b-4afe-a2a9-b63b96c4069c-kube-api-access-8g27w\") pod \"octavia-persistence-db-create-pxplj\" (UID: \"9edab071-f11b-4afe-a2a9-b63b96c4069c\") " pod="openstack/octavia-persistence-db-create-pxplj" Dec 01 16:17:09 crc kubenswrapper[4810]: I1201 16:17:09.261794 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9edab071-f11b-4afe-a2a9-b63b96c4069c-operator-scripts\") pod \"octavia-persistence-db-create-pxplj\" (UID: \"9edab071-f11b-4afe-a2a9-b63b96c4069c\") " pod="openstack/octavia-persistence-db-create-pxplj" Dec 01 16:17:09 crc kubenswrapper[4810]: I1201 16:17:09.363762 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9edab071-f11b-4afe-a2a9-b63b96c4069c-operator-scripts\") pod \"octavia-persistence-db-create-pxplj\" (UID: \"9edab071-f11b-4afe-a2a9-b63b96c4069c\") " pod="openstack/octavia-persistence-db-create-pxplj" Dec 01 16:17:09 crc kubenswrapper[4810]: I1201 16:17:09.364087 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g27w\" (UniqueName: \"kubernetes.io/projected/9edab071-f11b-4afe-a2a9-b63b96c4069c-kube-api-access-8g27w\") pod \"octavia-persistence-db-create-pxplj\" (UID: \"9edab071-f11b-4afe-a2a9-b63b96c4069c\") " pod="openstack/octavia-persistence-db-create-pxplj" Dec 01 16:17:09 crc kubenswrapper[4810]: I1201 16:17:09.364609 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9edab071-f11b-4afe-a2a9-b63b96c4069c-operator-scripts\") pod \"octavia-persistence-db-create-pxplj\" (UID: \"9edab071-f11b-4afe-a2a9-b63b96c4069c\") " pod="openstack/octavia-persistence-db-create-pxplj" Dec 01 16:17:09 crc kubenswrapper[4810]: I1201 16:17:09.395152 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g27w\" (UniqueName: \"kubernetes.io/projected/9edab071-f11b-4afe-a2a9-b63b96c4069c-kube-api-access-8g27w\") pod \"octavia-persistence-db-create-pxplj\" (UID: \"9edab071-f11b-4afe-a2a9-b63b96c4069c\") " pod="openstack/octavia-persistence-db-create-pxplj" Dec 01 16:17:09 crc kubenswrapper[4810]: I1201 16:17:09.497295 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-pxplj" Dec 01 16:17:09 crc kubenswrapper[4810]: I1201 16:17:09.949325 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-pxplj"] Dec 01 16:17:10 crc kubenswrapper[4810]: I1201 16:17:10.152599 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-11c5-account-create-update-b6n7h"] Dec 01 16:17:10 crc kubenswrapper[4810]: I1201 16:17:10.154284 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-11c5-account-create-update-b6n7h" Dec 01 16:17:10 crc kubenswrapper[4810]: I1201 16:17:10.156511 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Dec 01 16:17:10 crc kubenswrapper[4810]: I1201 16:17:10.164043 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-11c5-account-create-update-b6n7h"] Dec 01 16:17:10 crc kubenswrapper[4810]: I1201 16:17:10.281393 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/188b6e42-b278-4777-9be3-d86ffe3c16a5-operator-scripts\") pod \"octavia-11c5-account-create-update-b6n7h\" (UID: \"188b6e42-b278-4777-9be3-d86ffe3c16a5\") " pod="openstack/octavia-11c5-account-create-update-b6n7h" Dec 01 16:17:10 crc kubenswrapper[4810]: I1201 16:17:10.281450 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndh2c\" (UniqueName: \"kubernetes.io/projected/188b6e42-b278-4777-9be3-d86ffe3c16a5-kube-api-access-ndh2c\") pod \"octavia-11c5-account-create-update-b6n7h\" (UID: \"188b6e42-b278-4777-9be3-d86ffe3c16a5\") " pod="openstack/octavia-11c5-account-create-update-b6n7h" Dec 01 16:17:10 crc kubenswrapper[4810]: I1201 16:17:10.345555 4810 scope.go:117] "RemoveContainer" containerID="e30ac6813835f238af5ecd4c2ae2a64754ebabc796d14bb5c6d179b184bdfc67" Dec 01 16:17:10 crc kubenswrapper[4810]: I1201 16:17:10.365370 4810 scope.go:117] "RemoveContainer" containerID="12b03cbcefc24ecd02a6041471e8b3cf5452f647fd4d5d5bf8253ec98dc1401d" Dec 01 16:17:10 crc kubenswrapper[4810]: I1201 16:17:10.383712 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/188b6e42-b278-4777-9be3-d86ffe3c16a5-operator-scripts\") pod \"octavia-11c5-account-create-update-b6n7h\" (UID: \"188b6e42-b278-4777-9be3-d86ffe3c16a5\") " pod="openstack/octavia-11c5-account-create-update-b6n7h" Dec 01 16:17:10 crc kubenswrapper[4810]: I1201 16:17:10.383771 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndh2c\" (UniqueName: \"kubernetes.io/projected/188b6e42-b278-4777-9be3-d86ffe3c16a5-kube-api-access-ndh2c\") pod \"octavia-11c5-account-create-update-b6n7h\" (UID: \"188b6e42-b278-4777-9be3-d86ffe3c16a5\") " pod="openstack/octavia-11c5-account-create-update-b6n7h" Dec 01 16:17:10 crc kubenswrapper[4810]: I1201 16:17:10.384760 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/188b6e42-b278-4777-9be3-d86ffe3c16a5-operator-scripts\") pod \"octavia-11c5-account-create-update-b6n7h\" (UID: \"188b6e42-b278-4777-9be3-d86ffe3c16a5\") " pod="openstack/octavia-11c5-account-create-update-b6n7h" Dec 01 16:17:10 crc kubenswrapper[4810]: I1201 16:17:10.402350 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndh2c\" (UniqueName: \"kubernetes.io/projected/188b6e42-b278-4777-9be3-d86ffe3c16a5-kube-api-access-ndh2c\") pod \"octavia-11c5-account-create-update-b6n7h\" (UID: \"188b6e42-b278-4777-9be3-d86ffe3c16a5\") " pod="openstack/octavia-11c5-account-create-update-b6n7h" Dec 01 16:17:10 crc kubenswrapper[4810]: I1201 16:17:10.479971 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-11c5-account-create-update-b6n7h" Dec 01 16:17:10 crc kubenswrapper[4810]: I1201 16:17:10.526242 4810 generic.go:334] "Generic (PLEG): container finished" podID="9edab071-f11b-4afe-a2a9-b63b96c4069c" containerID="9b1bdb512c45ec6b07574c4ce36454471f78dd7ab8ff4a6e6fbf03b8f26a6008" exitCode=0 Dec 01 16:17:10 crc kubenswrapper[4810]: I1201 16:17:10.526306 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-pxplj" event={"ID":"9edab071-f11b-4afe-a2a9-b63b96c4069c","Type":"ContainerDied","Data":"9b1bdb512c45ec6b07574c4ce36454471f78dd7ab8ff4a6e6fbf03b8f26a6008"} Dec 01 16:17:10 crc kubenswrapper[4810]: I1201 16:17:10.526339 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-pxplj" event={"ID":"9edab071-f11b-4afe-a2a9-b63b96c4069c","Type":"ContainerStarted","Data":"346cbfbc11c91c86c515ff4f793cd9fe5483af21f7b77e32beac88fb509b44e4"} Dec 01 16:17:10 crc kubenswrapper[4810]: W1201 16:17:10.918313 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod188b6e42_b278_4777_9be3_d86ffe3c16a5.slice/crio-c06bdf8356fbf72d6a64b649ddde4949ddda9f9e0f08633fb83503fa1818b3ba WatchSource:0}: Error finding container c06bdf8356fbf72d6a64b649ddde4949ddda9f9e0f08633fb83503fa1818b3ba: Status 404 returned error can't find the container with id c06bdf8356fbf72d6a64b649ddde4949ddda9f9e0f08633fb83503fa1818b3ba Dec 01 16:17:10 crc kubenswrapper[4810]: I1201 16:17:10.926192 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-11c5-account-create-update-b6n7h"] Dec 01 16:17:11 crc kubenswrapper[4810]: I1201 16:17:11.538791 4810 generic.go:334] "Generic (PLEG): container finished" podID="188b6e42-b278-4777-9be3-d86ffe3c16a5" containerID="eb880792a3c2fbc18060174543da6970f79bc73c96c926f83aa10ec402bbbd6f" exitCode=0 Dec 01 16:17:11 crc kubenswrapper[4810]: I1201 16:17:11.538875 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-11c5-account-create-update-b6n7h" event={"ID":"188b6e42-b278-4777-9be3-d86ffe3c16a5","Type":"ContainerDied","Data":"eb880792a3c2fbc18060174543da6970f79bc73c96c926f83aa10ec402bbbd6f"} Dec 01 16:17:11 crc kubenswrapper[4810]: I1201 16:17:11.539573 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-11c5-account-create-update-b6n7h" event={"ID":"188b6e42-b278-4777-9be3-d86ffe3c16a5","Type":"ContainerStarted","Data":"c06bdf8356fbf72d6a64b649ddde4949ddda9f9e0f08633fb83503fa1818b3ba"} Dec 01 16:17:11 crc kubenswrapper[4810]: I1201 16:17:11.886972 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-pxplj" Dec 01 16:17:12 crc kubenswrapper[4810]: I1201 16:17:12.016042 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g27w\" (UniqueName: \"kubernetes.io/projected/9edab071-f11b-4afe-a2a9-b63b96c4069c-kube-api-access-8g27w\") pod \"9edab071-f11b-4afe-a2a9-b63b96c4069c\" (UID: \"9edab071-f11b-4afe-a2a9-b63b96c4069c\") " Dec 01 16:17:12 crc kubenswrapper[4810]: I1201 16:17:12.016153 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9edab071-f11b-4afe-a2a9-b63b96c4069c-operator-scripts\") pod \"9edab071-f11b-4afe-a2a9-b63b96c4069c\" (UID: \"9edab071-f11b-4afe-a2a9-b63b96c4069c\") " Dec 01 16:17:12 crc kubenswrapper[4810]: I1201 16:17:12.016638 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9edab071-f11b-4afe-a2a9-b63b96c4069c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9edab071-f11b-4afe-a2a9-b63b96c4069c" (UID: "9edab071-f11b-4afe-a2a9-b63b96c4069c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:17:12 crc kubenswrapper[4810]: I1201 16:17:12.022801 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9edab071-f11b-4afe-a2a9-b63b96c4069c-kube-api-access-8g27w" (OuterVolumeSpecName: "kube-api-access-8g27w") pod "9edab071-f11b-4afe-a2a9-b63b96c4069c" (UID: "9edab071-f11b-4afe-a2a9-b63b96c4069c"). InnerVolumeSpecName "kube-api-access-8g27w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:17:12 crc kubenswrapper[4810]: I1201 16:17:12.118128 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g27w\" (UniqueName: \"kubernetes.io/projected/9edab071-f11b-4afe-a2a9-b63b96c4069c-kube-api-access-8g27w\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:12 crc kubenswrapper[4810]: I1201 16:17:12.118159 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9edab071-f11b-4afe-a2a9-b63b96c4069c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:12 crc kubenswrapper[4810]: I1201 16:17:12.549564 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-pxplj" Dec 01 16:17:12 crc kubenswrapper[4810]: I1201 16:17:12.550245 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-pxplj" event={"ID":"9edab071-f11b-4afe-a2a9-b63b96c4069c","Type":"ContainerDied","Data":"346cbfbc11c91c86c515ff4f793cd9fe5483af21f7b77e32beac88fb509b44e4"} Dec 01 16:17:12 crc kubenswrapper[4810]: I1201 16:17:12.550266 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="346cbfbc11c91c86c515ff4f793cd9fe5483af21f7b77e32beac88fb509b44e4" Dec 01 16:17:12 crc kubenswrapper[4810]: I1201 16:17:12.867764 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-11c5-account-create-update-b6n7h" Dec 01 16:17:13 crc kubenswrapper[4810]: I1201 16:17:13.035756 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/188b6e42-b278-4777-9be3-d86ffe3c16a5-operator-scripts\") pod \"188b6e42-b278-4777-9be3-d86ffe3c16a5\" (UID: \"188b6e42-b278-4777-9be3-d86ffe3c16a5\") " Dec 01 16:17:13 crc kubenswrapper[4810]: I1201 16:17:13.035950 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndh2c\" (UniqueName: \"kubernetes.io/projected/188b6e42-b278-4777-9be3-d86ffe3c16a5-kube-api-access-ndh2c\") pod \"188b6e42-b278-4777-9be3-d86ffe3c16a5\" (UID: \"188b6e42-b278-4777-9be3-d86ffe3c16a5\") " Dec 01 16:17:13 crc kubenswrapper[4810]: I1201 16:17:13.036309 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/188b6e42-b278-4777-9be3-d86ffe3c16a5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "188b6e42-b278-4777-9be3-d86ffe3c16a5" (UID: "188b6e42-b278-4777-9be3-d86ffe3c16a5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:17:13 crc kubenswrapper[4810]: I1201 16:17:13.036587 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/188b6e42-b278-4777-9be3-d86ffe3c16a5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:13 crc kubenswrapper[4810]: I1201 16:17:13.046737 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/188b6e42-b278-4777-9be3-d86ffe3c16a5-kube-api-access-ndh2c" (OuterVolumeSpecName: "kube-api-access-ndh2c") pod "188b6e42-b278-4777-9be3-d86ffe3c16a5" (UID: "188b6e42-b278-4777-9be3-d86ffe3c16a5"). InnerVolumeSpecName "kube-api-access-ndh2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:17:13 crc kubenswrapper[4810]: I1201 16:17:13.062883 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-nhj96"] Dec 01 16:17:13 crc kubenswrapper[4810]: I1201 16:17:13.071338 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5de1-account-create-update-79k2v"] Dec 01 16:17:13 crc kubenswrapper[4810]: I1201 16:17:13.080643 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-5de1-account-create-update-79k2v"] Dec 01 16:17:13 crc kubenswrapper[4810]: I1201 16:17:13.088541 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-nhj96"] Dec 01 16:17:13 crc kubenswrapper[4810]: I1201 16:17:13.138370 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndh2c\" (UniqueName: \"kubernetes.io/projected/188b6e42-b278-4777-9be3-d86ffe3c16a5-kube-api-access-ndh2c\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:13 crc kubenswrapper[4810]: I1201 16:17:13.559534 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-11c5-account-create-update-b6n7h" event={"ID":"188b6e42-b278-4777-9be3-d86ffe3c16a5","Type":"ContainerDied","Data":"c06bdf8356fbf72d6a64b649ddde4949ddda9f9e0f08633fb83503fa1818b3ba"} Dec 01 16:17:13 crc kubenswrapper[4810]: I1201 16:17:13.559570 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c06bdf8356fbf72d6a64b649ddde4949ddda9f9e0f08633fb83503fa1818b3ba" Dec 01 16:17:13 crc kubenswrapper[4810]: I1201 16:17:13.559655 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-11c5-account-create-update-b6n7h" Dec 01 16:17:14 crc kubenswrapper[4810]: I1201 16:17:14.504675 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88b6fee6-788f-48c3-9d06-c5eaec035c49" path="/var/lib/kubelet/pods/88b6fee6-788f-48c3-9d06-c5eaec035c49/volumes" Dec 01 16:17:14 crc kubenswrapper[4810]: I1201 16:17:14.505589 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0bb8241-0cd8-487e-86d0-7c89995f2632" path="/var/lib/kubelet/pods/b0bb8241-0cd8-487e-86d0-7c89995f2632/volumes" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.186210 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-75d7f9cfbb-zdwbq"] Dec 01 16:17:16 crc kubenswrapper[4810]: E1201 16:17:16.186925 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9edab071-f11b-4afe-a2a9-b63b96c4069c" containerName="mariadb-database-create" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.186939 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9edab071-f11b-4afe-a2a9-b63b96c4069c" containerName="mariadb-database-create" Dec 01 16:17:16 crc kubenswrapper[4810]: E1201 16:17:16.186957 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188b6e42-b278-4777-9be3-d86ffe3c16a5" containerName="mariadb-account-create-update" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.186964 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="188b6e42-b278-4777-9be3-d86ffe3c16a5" containerName="mariadb-account-create-update" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.187149 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="188b6e42-b278-4777-9be3-d86ffe3c16a5" containerName="mariadb-account-create-update" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.187162 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9edab071-f11b-4afe-a2a9-b63b96c4069c" containerName="mariadb-database-create" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.188506 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.193072 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.193324 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-octavia-ovndbs" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.193594 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-znhmd" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.197026 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.203139 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-75d7f9cfbb-zdwbq"] Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.318352 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/fb2847c5-3667-4b51-a1b7-784fcea0a897-config-data-merged\") pod \"octavia-api-75d7f9cfbb-zdwbq\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.318455 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-config-data\") pod \"octavia-api-75d7f9cfbb-zdwbq\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.318524 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-scripts\") pod \"octavia-api-75d7f9cfbb-zdwbq\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.318561 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-combined-ca-bundle\") pod \"octavia-api-75d7f9cfbb-zdwbq\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.318594 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-ovndb-tls-certs\") pod \"octavia-api-75d7f9cfbb-zdwbq\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.318631 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/fb2847c5-3667-4b51-a1b7-784fcea0a897-octavia-run\") pod \"octavia-api-75d7f9cfbb-zdwbq\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.420220 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/fb2847c5-3667-4b51-a1b7-784fcea0a897-config-data-merged\") pod \"octavia-api-75d7f9cfbb-zdwbq\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.420324 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-config-data\") pod \"octavia-api-75d7f9cfbb-zdwbq\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.420383 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-scripts\") pod \"octavia-api-75d7f9cfbb-zdwbq\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.420422 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-combined-ca-bundle\") pod \"octavia-api-75d7f9cfbb-zdwbq\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.420525 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-ovndb-tls-certs\") pod \"octavia-api-75d7f9cfbb-zdwbq\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.420570 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/fb2847c5-3667-4b51-a1b7-784fcea0a897-octavia-run\") pod \"octavia-api-75d7f9cfbb-zdwbq\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.421223 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/fb2847c5-3667-4b51-a1b7-784fcea0a897-config-data-merged\") pod \"octavia-api-75d7f9cfbb-zdwbq\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.421228 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/fb2847c5-3667-4b51-a1b7-784fcea0a897-octavia-run\") pod \"octavia-api-75d7f9cfbb-zdwbq\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.426871 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-config-data\") pod \"octavia-api-75d7f9cfbb-zdwbq\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.426898 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-ovndb-tls-certs\") pod \"octavia-api-75d7f9cfbb-zdwbq\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.427631 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-scripts\") pod \"octavia-api-75d7f9cfbb-zdwbq\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.428238 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-combined-ca-bundle\") pod \"octavia-api-75d7f9cfbb-zdwbq\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.492450 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:17:16 crc kubenswrapper[4810]: E1201 16:17:16.492769 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.513141 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.990866 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-75d7f9cfbb-zdwbq"] Dec 01 16:17:16 crc kubenswrapper[4810]: I1201 16:17:16.995154 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:17:17 crc kubenswrapper[4810]: I1201 16:17:17.608583 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-75d7f9cfbb-zdwbq" event={"ID":"fb2847c5-3667-4b51-a1b7-784fcea0a897","Type":"ContainerStarted","Data":"ee0bd041bee15e99c0cb7f5ab43459b72f7394fd5867643f2d24a7304bbb952f"} Dec 01 16:17:20 crc kubenswrapper[4810]: I1201 16:17:20.035739 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-vt57p"] Dec 01 16:17:20 crc kubenswrapper[4810]: I1201 16:17:20.045309 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-vt57p"] Dec 01 16:17:20 crc kubenswrapper[4810]: I1201 16:17:20.505752 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cef9541-a570-4638-9360-8ecc123c441a" path="/var/lib/kubelet/pods/4cef9541-a570-4638-9360-8ecc123c441a/volumes" Dec 01 16:17:27 crc kubenswrapper[4810]: I1201 16:17:27.704464 4810 generic.go:334] "Generic (PLEG): container finished" podID="fb2847c5-3667-4b51-a1b7-784fcea0a897" containerID="73bc563fa06154ba3d1541eeb709c72ca50bc76ad1a19b5de38f5681f40486b1" exitCode=0 Dec 01 16:17:27 crc kubenswrapper[4810]: I1201 16:17:27.704545 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-75d7f9cfbb-zdwbq" event={"ID":"fb2847c5-3667-4b51-a1b7-784fcea0a897","Type":"ContainerDied","Data":"73bc563fa06154ba3d1541eeb709c72ca50bc76ad1a19b5de38f5681f40486b1"} Dec 01 16:17:28 crc kubenswrapper[4810]: I1201 16:17:28.714891 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-75d7f9cfbb-zdwbq" event={"ID":"fb2847c5-3667-4b51-a1b7-784fcea0a897","Type":"ContainerStarted","Data":"c001c32d77c6672cc553f637fdcbe0ee6887a96c667aaad887bd49951ee324fb"} Dec 01 16:17:28 crc kubenswrapper[4810]: I1201 16:17:28.715271 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-75d7f9cfbb-zdwbq" event={"ID":"fb2847c5-3667-4b51-a1b7-784fcea0a897","Type":"ContainerStarted","Data":"93cfcb2c6a52d38abd85a2c72025ec5a016e01f37d2531382f001961eb41ffeb"} Dec 01 16:17:28 crc kubenswrapper[4810]: I1201 16:17:28.715295 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:28 crc kubenswrapper[4810]: I1201 16:17:28.746358 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-75d7f9cfbb-zdwbq" podStartSLOduration=3.16749564 podStartE2EDuration="12.746341466s" podCreationTimestamp="2025-12-01 16:17:16 +0000 UTC" firstStartedPulling="2025-12-01 16:17:16.994961202 +0000 UTC m=+6202.758470805" lastFinishedPulling="2025-12-01 16:17:26.573807028 +0000 UTC m=+6212.337316631" observedRunningTime="2025-12-01 16:17:28.741140654 +0000 UTC m=+6214.504650267" watchObservedRunningTime="2025-12-01 16:17:28.746341466 +0000 UTC m=+6214.509851069" Dec 01 16:17:29 crc kubenswrapper[4810]: I1201 16:17:29.722563 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:30 crc kubenswrapper[4810]: I1201 16:17:30.491634 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:17:30 crc kubenswrapper[4810]: E1201 16:17:30.492254 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:17:33 crc kubenswrapper[4810]: I1201 16:17:33.055620 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-nbhcs"] Dec 01 16:17:33 crc kubenswrapper[4810]: I1201 16:17:33.070446 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-nbhcs"] Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.343618 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.345956 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-4kw8j" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.354498 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-rq44f" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.477843 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4kw8j-config-gghfv"] Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.479090 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.491893 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.504372 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="230f81a0-af9d-4d25-8319-4c9990813ffa" path="/var/lib/kubelet/pods/230f81a0-af9d-4d25-8319-4c9990813ffa/volumes" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.506205 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4kw8j-config-gghfv"] Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.600608 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a92504d-6e40-4994-938e-127a252b5491-var-run\") pod \"ovn-controller-4kw8j-config-gghfv\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.600662 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a92504d-6e40-4994-938e-127a252b5491-scripts\") pod \"ovn-controller-4kw8j-config-gghfv\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.600696 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a92504d-6e40-4994-938e-127a252b5491-var-log-ovn\") pod \"ovn-controller-4kw8j-config-gghfv\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.600820 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a92504d-6e40-4994-938e-127a252b5491-var-run-ovn\") pod \"ovn-controller-4kw8j-config-gghfv\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.600929 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs2cl\" (UniqueName: \"kubernetes.io/projected/6a92504d-6e40-4994-938e-127a252b5491-kube-api-access-vs2cl\") pod \"ovn-controller-4kw8j-config-gghfv\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.601046 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a92504d-6e40-4994-938e-127a252b5491-additional-scripts\") pod \"ovn-controller-4kw8j-config-gghfv\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.704199 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a92504d-6e40-4994-938e-127a252b5491-var-log-ovn\") pod \"ovn-controller-4kw8j-config-gghfv\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.704264 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a92504d-6e40-4994-938e-127a252b5491-var-run-ovn\") pod \"ovn-controller-4kw8j-config-gghfv\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.704288 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs2cl\" (UniqueName: \"kubernetes.io/projected/6a92504d-6e40-4994-938e-127a252b5491-kube-api-access-vs2cl\") pod \"ovn-controller-4kw8j-config-gghfv\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.704310 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a92504d-6e40-4994-938e-127a252b5491-additional-scripts\") pod \"ovn-controller-4kw8j-config-gghfv\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.704435 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a92504d-6e40-4994-938e-127a252b5491-var-run\") pod \"ovn-controller-4kw8j-config-gghfv\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.704457 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a92504d-6e40-4994-938e-127a252b5491-scripts\") pod \"ovn-controller-4kw8j-config-gghfv\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.704576 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a92504d-6e40-4994-938e-127a252b5491-var-log-ovn\") pod \"ovn-controller-4kw8j-config-gghfv\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.704608 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a92504d-6e40-4994-938e-127a252b5491-var-run-ovn\") pod \"ovn-controller-4kw8j-config-gghfv\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.704662 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a92504d-6e40-4994-938e-127a252b5491-var-run\") pod \"ovn-controller-4kw8j-config-gghfv\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.705366 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a92504d-6e40-4994-938e-127a252b5491-additional-scripts\") pod \"ovn-controller-4kw8j-config-gghfv\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.706448 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a92504d-6e40-4994-938e-127a252b5491-scripts\") pod \"ovn-controller-4kw8j-config-gghfv\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.728029 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs2cl\" (UniqueName: \"kubernetes.io/projected/6a92504d-6e40-4994-938e-127a252b5491-kube-api-access-vs2cl\") pod \"ovn-controller-4kw8j-config-gghfv\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:34 crc kubenswrapper[4810]: I1201 16:17:34.809895 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:35 crc kubenswrapper[4810]: I1201 16:17:35.449738 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4kw8j-config-gghfv"] Dec 01 16:17:35 crc kubenswrapper[4810]: I1201 16:17:35.787595 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kw8j-config-gghfv" event={"ID":"6a92504d-6e40-4994-938e-127a252b5491","Type":"ContainerStarted","Data":"9dbd1ef9b8801f43748894b2aae2ccf27db171d23df4908b86cc738a1d42b8db"} Dec 01 16:17:36 crc kubenswrapper[4810]: I1201 16:17:36.797622 4810 generic.go:334] "Generic (PLEG): container finished" podID="6a92504d-6e40-4994-938e-127a252b5491" containerID="d53fe94224d2bd2ea49669d0f69129401ffd965650a924d2f4da3e096289347d" exitCode=0 Dec 01 16:17:36 crc kubenswrapper[4810]: I1201 16:17:36.797676 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kw8j-config-gghfv" event={"ID":"6a92504d-6e40-4994-938e-127a252b5491","Type":"ContainerDied","Data":"d53fe94224d2bd2ea49669d0f69129401ffd965650a924d2f4da3e096289347d"} Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.162688 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.271789 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a92504d-6e40-4994-938e-127a252b5491-additional-scripts\") pod \"6a92504d-6e40-4994-938e-127a252b5491\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.271910 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a92504d-6e40-4994-938e-127a252b5491-var-log-ovn\") pod \"6a92504d-6e40-4994-938e-127a252b5491\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.271935 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a92504d-6e40-4994-938e-127a252b5491-var-run\") pod \"6a92504d-6e40-4994-938e-127a252b5491\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.272036 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a92504d-6e40-4994-938e-127a252b5491-scripts\") pod \"6a92504d-6e40-4994-938e-127a252b5491\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.272074 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vs2cl\" (UniqueName: \"kubernetes.io/projected/6a92504d-6e40-4994-938e-127a252b5491-kube-api-access-vs2cl\") pod \"6a92504d-6e40-4994-938e-127a252b5491\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.272060 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a92504d-6e40-4994-938e-127a252b5491-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6a92504d-6e40-4994-938e-127a252b5491" (UID: "6a92504d-6e40-4994-938e-127a252b5491"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.272173 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a92504d-6e40-4994-938e-127a252b5491-var-run-ovn\") pod \"6a92504d-6e40-4994-938e-127a252b5491\" (UID: \"6a92504d-6e40-4994-938e-127a252b5491\") " Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.272172 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a92504d-6e40-4994-938e-127a252b5491-var-run" (OuterVolumeSpecName: "var-run") pod "6a92504d-6e40-4994-938e-127a252b5491" (UID: "6a92504d-6e40-4994-938e-127a252b5491"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.272344 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a92504d-6e40-4994-938e-127a252b5491-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6a92504d-6e40-4994-938e-127a252b5491" (UID: "6a92504d-6e40-4994-938e-127a252b5491"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.272591 4810 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a92504d-6e40-4994-938e-127a252b5491-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.272616 4810 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a92504d-6e40-4994-938e-127a252b5491-var-run\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.272626 4810 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a92504d-6e40-4994-938e-127a252b5491-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.273120 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a92504d-6e40-4994-938e-127a252b5491-scripts" (OuterVolumeSpecName: "scripts") pod "6a92504d-6e40-4994-938e-127a252b5491" (UID: "6a92504d-6e40-4994-938e-127a252b5491"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.273425 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a92504d-6e40-4994-938e-127a252b5491-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "6a92504d-6e40-4994-938e-127a252b5491" (UID: "6a92504d-6e40-4994-938e-127a252b5491"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.276903 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a92504d-6e40-4994-938e-127a252b5491-kube-api-access-vs2cl" (OuterVolumeSpecName: "kube-api-access-vs2cl") pod "6a92504d-6e40-4994-938e-127a252b5491" (UID: "6a92504d-6e40-4994-938e-127a252b5491"). InnerVolumeSpecName "kube-api-access-vs2cl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.373956 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a92504d-6e40-4994-938e-127a252b5491-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.373999 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vs2cl\" (UniqueName: \"kubernetes.io/projected/6a92504d-6e40-4994-938e-127a252b5491-kube-api-access-vs2cl\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.374014 4810 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a92504d-6e40-4994-938e-127a252b5491-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.817253 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kw8j-config-gghfv" event={"ID":"6a92504d-6e40-4994-938e-127a252b5491","Type":"ContainerDied","Data":"9dbd1ef9b8801f43748894b2aae2ccf27db171d23df4908b86cc738a1d42b8db"} Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.817291 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9dbd1ef9b8801f43748894b2aae2ccf27db171d23df4908b86cc738a1d42b8db" Dec 01 16:17:38 crc kubenswrapper[4810]: I1201 16:17:38.817314 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kw8j-config-gghfv" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.257027 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4kw8j-config-gghfv"] Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.268534 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-4kw8j-config-gghfv"] Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.375360 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4kw8j-config-rrwb8"] Dec 01 16:17:39 crc kubenswrapper[4810]: E1201 16:17:39.375792 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a92504d-6e40-4994-938e-127a252b5491" containerName="ovn-config" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.375810 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a92504d-6e40-4994-938e-127a252b5491" containerName="ovn-config" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.376040 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a92504d-6e40-4994-938e-127a252b5491" containerName="ovn-config" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.376713 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.389836 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.390884 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4kw8j-config-rrwb8"] Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.496320 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-var-run\") pod \"ovn-controller-4kw8j-config-rrwb8\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.496371 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-additional-scripts\") pod \"ovn-controller-4kw8j-config-rrwb8\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.496388 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6l48\" (UniqueName: \"kubernetes.io/projected/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-kube-api-access-g6l48\") pod \"ovn-controller-4kw8j-config-rrwb8\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.496405 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-scripts\") pod \"ovn-controller-4kw8j-config-rrwb8\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.496439 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-var-run-ovn\") pod \"ovn-controller-4kw8j-config-rrwb8\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.496747 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-var-log-ovn\") pod \"ovn-controller-4kw8j-config-rrwb8\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.598744 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-var-run\") pod \"ovn-controller-4kw8j-config-rrwb8\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.598806 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6l48\" (UniqueName: \"kubernetes.io/projected/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-kube-api-access-g6l48\") pod \"ovn-controller-4kw8j-config-rrwb8\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.598835 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-additional-scripts\") pod \"ovn-controller-4kw8j-config-rrwb8\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.598870 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-scripts\") pod \"ovn-controller-4kw8j-config-rrwb8\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.598924 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-var-run-ovn\") pod \"ovn-controller-4kw8j-config-rrwb8\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.599036 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-var-log-ovn\") pod \"ovn-controller-4kw8j-config-rrwb8\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.599179 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-var-run\") pod \"ovn-controller-4kw8j-config-rrwb8\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.599193 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-var-log-ovn\") pod \"ovn-controller-4kw8j-config-rrwb8\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.599610 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-var-run-ovn\") pod \"ovn-controller-4kw8j-config-rrwb8\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.600499 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-additional-scripts\") pod \"ovn-controller-4kw8j-config-rrwb8\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.601411 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-scripts\") pod \"ovn-controller-4kw8j-config-rrwb8\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.620222 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6l48\" (UniqueName: \"kubernetes.io/projected/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-kube-api-access-g6l48\") pod \"ovn-controller-4kw8j-config-rrwb8\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:39 crc kubenswrapper[4810]: I1201 16:17:39.698451 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:40 crc kubenswrapper[4810]: I1201 16:17:40.199771 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4kw8j-config-rrwb8"] Dec 01 16:17:40 crc kubenswrapper[4810]: W1201 16:17:40.214824 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a9bfcf8_4b3e_4f76_a1ef_e6f5bf1bf6d5.slice/crio-1e6b229ed2771c1a9babd71b7a84de79622721472b641e8f31e1d7e1a67faf44 WatchSource:0}: Error finding container 1e6b229ed2771c1a9babd71b7a84de79622721472b641e8f31e1d7e1a67faf44: Status 404 returned error can't find the container with id 1e6b229ed2771c1a9babd71b7a84de79622721472b641e8f31e1d7e1a67faf44 Dec 01 16:17:40 crc kubenswrapper[4810]: I1201 16:17:40.503360 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a92504d-6e40-4994-938e-127a252b5491" path="/var/lib/kubelet/pods/6a92504d-6e40-4994-938e-127a252b5491/volumes" Dec 01 16:17:40 crc kubenswrapper[4810]: I1201 16:17:40.845141 4810 generic.go:334] "Generic (PLEG): container finished" podID="0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5" containerID="c4abe3f0f775cd77dd7ee7c754d96148cc205726254d214cbbb8fbbccc4f53be" exitCode=0 Dec 01 16:17:40 crc kubenswrapper[4810]: I1201 16:17:40.845187 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kw8j-config-rrwb8" event={"ID":"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5","Type":"ContainerDied","Data":"c4abe3f0f775cd77dd7ee7c754d96148cc205726254d214cbbb8fbbccc4f53be"} Dec 01 16:17:40 crc kubenswrapper[4810]: I1201 16:17:40.845214 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kw8j-config-rrwb8" event={"ID":"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5","Type":"ContainerStarted","Data":"1e6b229ed2771c1a9babd71b7a84de79622721472b641e8f31e1d7e1a67faf44"} Dec 01 16:17:41 crc kubenswrapper[4810]: I1201 16:17:41.609827 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d8glv"] Dec 01 16:17:41 crc kubenswrapper[4810]: I1201 16:17:41.613800 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d8glv" Dec 01 16:17:41 crc kubenswrapper[4810]: I1201 16:17:41.631695 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d8glv"] Dec 01 16:17:41 crc kubenswrapper[4810]: I1201 16:17:41.643311 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b286d0d-df8f-4501-9fa9-4436cb768aa5-catalog-content\") pod \"redhat-operators-d8glv\" (UID: \"1b286d0d-df8f-4501-9fa9-4436cb768aa5\") " pod="openshift-marketplace/redhat-operators-d8glv" Dec 01 16:17:41 crc kubenswrapper[4810]: I1201 16:17:41.643362 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fhls\" (UniqueName: \"kubernetes.io/projected/1b286d0d-df8f-4501-9fa9-4436cb768aa5-kube-api-access-5fhls\") pod \"redhat-operators-d8glv\" (UID: \"1b286d0d-df8f-4501-9fa9-4436cb768aa5\") " pod="openshift-marketplace/redhat-operators-d8glv" Dec 01 16:17:41 crc kubenswrapper[4810]: I1201 16:17:41.643393 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b286d0d-df8f-4501-9fa9-4436cb768aa5-utilities\") pod \"redhat-operators-d8glv\" (UID: \"1b286d0d-df8f-4501-9fa9-4436cb768aa5\") " pod="openshift-marketplace/redhat-operators-d8glv" Dec 01 16:17:41 crc kubenswrapper[4810]: I1201 16:17:41.745642 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b286d0d-df8f-4501-9fa9-4436cb768aa5-catalog-content\") pod \"redhat-operators-d8glv\" (UID: \"1b286d0d-df8f-4501-9fa9-4436cb768aa5\") " pod="openshift-marketplace/redhat-operators-d8glv" Dec 01 16:17:41 crc kubenswrapper[4810]: I1201 16:17:41.745719 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fhls\" (UniqueName: \"kubernetes.io/projected/1b286d0d-df8f-4501-9fa9-4436cb768aa5-kube-api-access-5fhls\") pod \"redhat-operators-d8glv\" (UID: \"1b286d0d-df8f-4501-9fa9-4436cb768aa5\") " pod="openshift-marketplace/redhat-operators-d8glv" Dec 01 16:17:41 crc kubenswrapper[4810]: I1201 16:17:41.745756 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b286d0d-df8f-4501-9fa9-4436cb768aa5-utilities\") pod \"redhat-operators-d8glv\" (UID: \"1b286d0d-df8f-4501-9fa9-4436cb768aa5\") " pod="openshift-marketplace/redhat-operators-d8glv" Dec 01 16:17:41 crc kubenswrapper[4810]: I1201 16:17:41.746495 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b286d0d-df8f-4501-9fa9-4436cb768aa5-utilities\") pod \"redhat-operators-d8glv\" (UID: \"1b286d0d-df8f-4501-9fa9-4436cb768aa5\") " pod="openshift-marketplace/redhat-operators-d8glv" Dec 01 16:17:41 crc kubenswrapper[4810]: I1201 16:17:41.746771 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b286d0d-df8f-4501-9fa9-4436cb768aa5-catalog-content\") pod \"redhat-operators-d8glv\" (UID: \"1b286d0d-df8f-4501-9fa9-4436cb768aa5\") " pod="openshift-marketplace/redhat-operators-d8glv" Dec 01 16:17:41 crc kubenswrapper[4810]: I1201 16:17:41.770077 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fhls\" (UniqueName: \"kubernetes.io/projected/1b286d0d-df8f-4501-9fa9-4436cb768aa5-kube-api-access-5fhls\") pod \"redhat-operators-d8glv\" (UID: \"1b286d0d-df8f-4501-9fa9-4436cb768aa5\") " pod="openshift-marketplace/redhat-operators-d8glv" Dec 01 16:17:41 crc kubenswrapper[4810]: I1201 16:17:41.945484 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d8glv" Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.250824 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.255167 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-scripts\") pod \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.255223 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6l48\" (UniqueName: \"kubernetes.io/projected/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-kube-api-access-g6l48\") pod \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.255245 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-var-run-ovn\") pod \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.255310 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-var-log-ovn\") pod \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.255328 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-additional-scripts\") pod \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.255351 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-var-run\") pod \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\" (UID: \"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5\") " Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.255531 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5" (UID: "0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.255574 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5" (UID: "0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.255581 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-var-run" (OuterVolumeSpecName: "var-run") pod "0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5" (UID: "0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.255966 4810 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.255981 4810 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.256353 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5" (UID: "0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.256595 4810 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-var-run\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.256716 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-scripts" (OuterVolumeSpecName: "scripts") pod "0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5" (UID: "0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.262001 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-kube-api-access-g6l48" (OuterVolumeSpecName: "kube-api-access-g6l48") pod "0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5" (UID: "0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5"). InnerVolumeSpecName "kube-api-access-g6l48". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.358843 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.358895 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6l48\" (UniqueName: \"kubernetes.io/projected/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-kube-api-access-g6l48\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.358912 4810 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.446739 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d8glv"] Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.863403 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kw8j-config-rrwb8" Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.863408 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kw8j-config-rrwb8" event={"ID":"0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5","Type":"ContainerDied","Data":"1e6b229ed2771c1a9babd71b7a84de79622721472b641e8f31e1d7e1a67faf44"} Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.864877 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e6b229ed2771c1a9babd71b7a84de79622721472b641e8f31e1d7e1a67faf44" Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.865527 4810 generic.go:334] "Generic (PLEG): container finished" podID="1b286d0d-df8f-4501-9fa9-4436cb768aa5" containerID="368b8552c8dd9c98a6d00f67ecbdaecc9b0625b2d9134c5352da6aa772d78284" exitCode=0 Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.865558 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8glv" event={"ID":"1b286d0d-df8f-4501-9fa9-4436cb768aa5","Type":"ContainerDied","Data":"368b8552c8dd9c98a6d00f67ecbdaecc9b0625b2d9134c5352da6aa772d78284"} Dec 01 16:17:42 crc kubenswrapper[4810]: I1201 16:17:42.865593 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8glv" event={"ID":"1b286d0d-df8f-4501-9fa9-4436cb768aa5","Type":"ContainerStarted","Data":"a535eac415e3d613280793bc083f37cf6cc9bfde376421883d5df3ec156675f2"} Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.333259 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4kw8j-config-rrwb8"] Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.341710 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-4kw8j-config-rrwb8"] Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.491750 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:17:43 crc kubenswrapper[4810]: E1201 16:17:43.492016 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.612316 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-qs2z8"] Dec 01 16:17:43 crc kubenswrapper[4810]: E1201 16:17:43.612988 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5" containerName="ovn-config" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.613020 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5" containerName="ovn-config" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.613303 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5" containerName="ovn-config" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.614826 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-qs2z8" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.618063 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.618334 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.618530 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.624803 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-qs2z8"] Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.790941 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30053c9f-67a1-4ec1-b2bf-c61ff171a0d9-scripts\") pod \"octavia-rsyslog-qs2z8\" (UID: \"30053c9f-67a1-4ec1-b2bf-c61ff171a0d9\") " pod="openstack/octavia-rsyslog-qs2z8" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.791061 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/30053c9f-67a1-4ec1-b2bf-c61ff171a0d9-config-data-merged\") pod \"octavia-rsyslog-qs2z8\" (UID: \"30053c9f-67a1-4ec1-b2bf-c61ff171a0d9\") " pod="openstack/octavia-rsyslog-qs2z8" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.791096 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30053c9f-67a1-4ec1-b2bf-c61ff171a0d9-config-data\") pod \"octavia-rsyslog-qs2z8\" (UID: \"30053c9f-67a1-4ec1-b2bf-c61ff171a0d9\") " pod="openstack/octavia-rsyslog-qs2z8" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.791199 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/30053c9f-67a1-4ec1-b2bf-c61ff171a0d9-hm-ports\") pod \"octavia-rsyslog-qs2z8\" (UID: \"30053c9f-67a1-4ec1-b2bf-c61ff171a0d9\") " pod="openstack/octavia-rsyslog-qs2z8" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.893227 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30053c9f-67a1-4ec1-b2bf-c61ff171a0d9-scripts\") pod \"octavia-rsyslog-qs2z8\" (UID: \"30053c9f-67a1-4ec1-b2bf-c61ff171a0d9\") " pod="openstack/octavia-rsyslog-qs2z8" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.893311 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/30053c9f-67a1-4ec1-b2bf-c61ff171a0d9-config-data-merged\") pod \"octavia-rsyslog-qs2z8\" (UID: \"30053c9f-67a1-4ec1-b2bf-c61ff171a0d9\") " pod="openstack/octavia-rsyslog-qs2z8" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.893355 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30053c9f-67a1-4ec1-b2bf-c61ff171a0d9-config-data\") pod \"octavia-rsyslog-qs2z8\" (UID: \"30053c9f-67a1-4ec1-b2bf-c61ff171a0d9\") " pod="openstack/octavia-rsyslog-qs2z8" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.893499 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/30053c9f-67a1-4ec1-b2bf-c61ff171a0d9-hm-ports\") pod \"octavia-rsyslog-qs2z8\" (UID: \"30053c9f-67a1-4ec1-b2bf-c61ff171a0d9\") " pod="openstack/octavia-rsyslog-qs2z8" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.893967 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/30053c9f-67a1-4ec1-b2bf-c61ff171a0d9-config-data-merged\") pod \"octavia-rsyslog-qs2z8\" (UID: \"30053c9f-67a1-4ec1-b2bf-c61ff171a0d9\") " pod="openstack/octavia-rsyslog-qs2z8" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.894302 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/30053c9f-67a1-4ec1-b2bf-c61ff171a0d9-hm-ports\") pod \"octavia-rsyslog-qs2z8\" (UID: \"30053c9f-67a1-4ec1-b2bf-c61ff171a0d9\") " pod="openstack/octavia-rsyslog-qs2z8" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.901292 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30053c9f-67a1-4ec1-b2bf-c61ff171a0d9-scripts\") pod \"octavia-rsyslog-qs2z8\" (UID: \"30053c9f-67a1-4ec1-b2bf-c61ff171a0d9\") " pod="openstack/octavia-rsyslog-qs2z8" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.902863 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30053c9f-67a1-4ec1-b2bf-c61ff171a0d9-config-data\") pod \"octavia-rsyslog-qs2z8\" (UID: \"30053c9f-67a1-4ec1-b2bf-c61ff171a0d9\") " pod="openstack/octavia-rsyslog-qs2z8" Dec 01 16:17:43 crc kubenswrapper[4810]: I1201 16:17:43.933874 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-qs2z8" Dec 01 16:17:44 crc kubenswrapper[4810]: I1201 16:17:44.509762 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5" path="/var/lib/kubelet/pods/0a9bfcf8-4b3e-4f76-a1ef-e6f5bf1bf6d5/volumes" Dec 01 16:17:44 crc kubenswrapper[4810]: I1201 16:17:44.546566 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-699f5f4fd7-mk5nx"] Dec 01 16:17:44 crc kubenswrapper[4810]: I1201 16:17:44.548494 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-699f5f4fd7-mk5nx" Dec 01 16:17:44 crc kubenswrapper[4810]: I1201 16:17:44.553164 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Dec 01 16:17:44 crc kubenswrapper[4810]: I1201 16:17:44.562731 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-qs2z8"] Dec 01 16:17:44 crc kubenswrapper[4810]: I1201 16:17:44.575259 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-699f5f4fd7-mk5nx"] Dec 01 16:17:44 crc kubenswrapper[4810]: I1201 16:17:44.610215 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0-httpd-config\") pod \"octavia-image-upload-699f5f4fd7-mk5nx\" (UID: \"b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0\") " pod="openstack/octavia-image-upload-699f5f4fd7-mk5nx" Dec 01 16:17:44 crc kubenswrapper[4810]: I1201 16:17:44.610551 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0-amphora-image\") pod \"octavia-image-upload-699f5f4fd7-mk5nx\" (UID: \"b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0\") " pod="openstack/octavia-image-upload-699f5f4fd7-mk5nx" Dec 01 16:17:44 crc kubenswrapper[4810]: I1201 16:17:44.711984 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0-amphora-image\") pod \"octavia-image-upload-699f5f4fd7-mk5nx\" (UID: \"b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0\") " pod="openstack/octavia-image-upload-699f5f4fd7-mk5nx" Dec 01 16:17:44 crc kubenswrapper[4810]: I1201 16:17:44.712366 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0-httpd-config\") pod \"octavia-image-upload-699f5f4fd7-mk5nx\" (UID: \"b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0\") " pod="openstack/octavia-image-upload-699f5f4fd7-mk5nx" Dec 01 16:17:44 crc kubenswrapper[4810]: I1201 16:17:44.712778 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0-amphora-image\") pod \"octavia-image-upload-699f5f4fd7-mk5nx\" (UID: \"b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0\") " pod="openstack/octavia-image-upload-699f5f4fd7-mk5nx" Dec 01 16:17:44 crc kubenswrapper[4810]: I1201 16:17:44.718456 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0-httpd-config\") pod \"octavia-image-upload-699f5f4fd7-mk5nx\" (UID: \"b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0\") " pod="openstack/octavia-image-upload-699f5f4fd7-mk5nx" Dec 01 16:17:44 crc kubenswrapper[4810]: I1201 16:17:44.886103 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-qs2z8" event={"ID":"30053c9f-67a1-4ec1-b2bf-c61ff171a0d9","Type":"ContainerStarted","Data":"1e4d5ac94dc2e6f9de3cc5573aa00cd50604aa398f8532a29095ee8caa57ce2d"} Dec 01 16:17:44 crc kubenswrapper[4810]: I1201 16:17:44.889056 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8glv" event={"ID":"1b286d0d-df8f-4501-9fa9-4436cb768aa5","Type":"ContainerStarted","Data":"c79aef5b334e87f6b1ae6a9e0c86d772e32a2b7fa206b69e355b46696494eab5"} Dec 01 16:17:44 crc kubenswrapper[4810]: I1201 16:17:44.944123 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-699f5f4fd7-mk5nx" Dec 01 16:17:45 crc kubenswrapper[4810]: I1201 16:17:45.391548 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-699f5f4fd7-mk5nx"] Dec 01 16:17:45 crc kubenswrapper[4810]: W1201 16:17:45.403604 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9b6f063_3e0a_4e8f_bcfe_31034b0f4ac0.slice/crio-0c468156da91da6cd949e5c719da0a910ff7af42b43435c84739569cce1adfb6 WatchSource:0}: Error finding container 0c468156da91da6cd949e5c719da0a910ff7af42b43435c84739569cce1adfb6: Status 404 returned error can't find the container with id 0c468156da91da6cd949e5c719da0a910ff7af42b43435c84739569cce1adfb6 Dec 01 16:17:45 crc kubenswrapper[4810]: I1201 16:17:45.907816 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-699f5f4fd7-mk5nx" event={"ID":"b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0","Type":"ContainerStarted","Data":"0c468156da91da6cd949e5c719da0a910ff7af42b43435c84739569cce1adfb6"} Dec 01 16:17:46 crc kubenswrapper[4810]: I1201 16:17:46.067397 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-bcw9j"] Dec 01 16:17:46 crc kubenswrapper[4810]: I1201 16:17:46.069219 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-bcw9j" Dec 01 16:17:46 crc kubenswrapper[4810]: I1201 16:17:46.071416 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Dec 01 16:17:46 crc kubenswrapper[4810]: I1201 16:17:46.108417 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-bcw9j"] Dec 01 16:17:46 crc kubenswrapper[4810]: I1201 16:17:46.142246 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89486609-b664-4b57-8776-c9b421186fab-scripts\") pod \"octavia-db-sync-bcw9j\" (UID: \"89486609-b664-4b57-8776-c9b421186fab\") " pod="openstack/octavia-db-sync-bcw9j" Dec 01 16:17:46 crc kubenswrapper[4810]: I1201 16:17:46.142354 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/89486609-b664-4b57-8776-c9b421186fab-config-data-merged\") pod \"octavia-db-sync-bcw9j\" (UID: \"89486609-b664-4b57-8776-c9b421186fab\") " pod="openstack/octavia-db-sync-bcw9j" Dec 01 16:17:46 crc kubenswrapper[4810]: I1201 16:17:46.142396 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89486609-b664-4b57-8776-c9b421186fab-config-data\") pod \"octavia-db-sync-bcw9j\" (UID: \"89486609-b664-4b57-8776-c9b421186fab\") " pod="openstack/octavia-db-sync-bcw9j" Dec 01 16:17:46 crc kubenswrapper[4810]: I1201 16:17:46.142782 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89486609-b664-4b57-8776-c9b421186fab-combined-ca-bundle\") pod \"octavia-db-sync-bcw9j\" (UID: \"89486609-b664-4b57-8776-c9b421186fab\") " pod="openstack/octavia-db-sync-bcw9j" Dec 01 16:17:46 crc kubenswrapper[4810]: I1201 16:17:46.266551 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89486609-b664-4b57-8776-c9b421186fab-scripts\") pod \"octavia-db-sync-bcw9j\" (UID: \"89486609-b664-4b57-8776-c9b421186fab\") " pod="openstack/octavia-db-sync-bcw9j" Dec 01 16:17:46 crc kubenswrapper[4810]: I1201 16:17:46.266643 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/89486609-b664-4b57-8776-c9b421186fab-config-data-merged\") pod \"octavia-db-sync-bcw9j\" (UID: \"89486609-b664-4b57-8776-c9b421186fab\") " pod="openstack/octavia-db-sync-bcw9j" Dec 01 16:17:46 crc kubenswrapper[4810]: I1201 16:17:46.266672 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89486609-b664-4b57-8776-c9b421186fab-config-data\") pod \"octavia-db-sync-bcw9j\" (UID: \"89486609-b664-4b57-8776-c9b421186fab\") " pod="openstack/octavia-db-sync-bcw9j" Dec 01 16:17:46 crc kubenswrapper[4810]: I1201 16:17:46.266727 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89486609-b664-4b57-8776-c9b421186fab-combined-ca-bundle\") pod \"octavia-db-sync-bcw9j\" (UID: \"89486609-b664-4b57-8776-c9b421186fab\") " pod="openstack/octavia-db-sync-bcw9j" Dec 01 16:17:46 crc kubenswrapper[4810]: I1201 16:17:46.267274 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/89486609-b664-4b57-8776-c9b421186fab-config-data-merged\") pod \"octavia-db-sync-bcw9j\" (UID: \"89486609-b664-4b57-8776-c9b421186fab\") " pod="openstack/octavia-db-sync-bcw9j" Dec 01 16:17:46 crc kubenswrapper[4810]: I1201 16:17:46.273505 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89486609-b664-4b57-8776-c9b421186fab-combined-ca-bundle\") pod \"octavia-db-sync-bcw9j\" (UID: \"89486609-b664-4b57-8776-c9b421186fab\") " pod="openstack/octavia-db-sync-bcw9j" Dec 01 16:17:46 crc kubenswrapper[4810]: I1201 16:17:46.274984 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89486609-b664-4b57-8776-c9b421186fab-config-data\") pod \"octavia-db-sync-bcw9j\" (UID: \"89486609-b664-4b57-8776-c9b421186fab\") " pod="openstack/octavia-db-sync-bcw9j" Dec 01 16:17:46 crc kubenswrapper[4810]: I1201 16:17:46.287634 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89486609-b664-4b57-8776-c9b421186fab-scripts\") pod \"octavia-db-sync-bcw9j\" (UID: \"89486609-b664-4b57-8776-c9b421186fab\") " pod="openstack/octavia-db-sync-bcw9j" Dec 01 16:17:46 crc kubenswrapper[4810]: I1201 16:17:46.427839 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-bcw9j" Dec 01 16:17:47 crc kubenswrapper[4810]: I1201 16:17:47.954859 4810 generic.go:334] "Generic (PLEG): container finished" podID="1b286d0d-df8f-4501-9fa9-4436cb768aa5" containerID="c79aef5b334e87f6b1ae6a9e0c86d772e32a2b7fa206b69e355b46696494eab5" exitCode=0 Dec 01 16:17:47 crc kubenswrapper[4810]: I1201 16:17:47.954926 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8glv" event={"ID":"1b286d0d-df8f-4501-9fa9-4436cb768aa5","Type":"ContainerDied","Data":"c79aef5b334e87f6b1ae6a9e0c86d772e32a2b7fa206b69e355b46696494eab5"} Dec 01 16:17:47 crc kubenswrapper[4810]: I1201 16:17:47.957640 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-bcw9j"] Dec 01 16:17:48 crc kubenswrapper[4810]: W1201 16:17:48.134951 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89486609_b664_4b57_8776_c9b421186fab.slice/crio-6e143ea5614a904cd1022d45628a21adf45aca15ae43e7a647144281b0d2c526 WatchSource:0}: Error finding container 6e143ea5614a904cd1022d45628a21adf45aca15ae43e7a647144281b0d2c526: Status 404 returned error can't find the container with id 6e143ea5614a904cd1022d45628a21adf45aca15ae43e7a647144281b0d2c526 Dec 01 16:17:48 crc kubenswrapper[4810]: I1201 16:17:48.967382 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8glv" event={"ID":"1b286d0d-df8f-4501-9fa9-4436cb768aa5","Type":"ContainerStarted","Data":"50b44a5721e643dbbfb789bbde99ee25a28ad6e968015d9cc8c2b4abf3ad78c3"} Dec 01 16:17:48 crc kubenswrapper[4810]: I1201 16:17:48.971635 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-bcw9j" event={"ID":"89486609-b664-4b57-8776-c9b421186fab","Type":"ContainerStarted","Data":"5780a34684dc0843aa275ef4948d4f619741a71249fff8e5310309fa0b40399f"} Dec 01 16:17:48 crc kubenswrapper[4810]: I1201 16:17:48.971678 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-bcw9j" event={"ID":"89486609-b664-4b57-8776-c9b421186fab","Type":"ContainerStarted","Data":"6e143ea5614a904cd1022d45628a21adf45aca15ae43e7a647144281b0d2c526"} Dec 01 16:17:48 crc kubenswrapper[4810]: I1201 16:17:48.986561 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d8glv" podStartSLOduration=2.189281917 podStartE2EDuration="7.986532844s" podCreationTimestamp="2025-12-01 16:17:41 +0000 UTC" firstStartedPulling="2025-12-01 16:17:42.87920169 +0000 UTC m=+6228.642711293" lastFinishedPulling="2025-12-01 16:17:48.676452617 +0000 UTC m=+6234.439962220" observedRunningTime="2025-12-01 16:17:48.982568698 +0000 UTC m=+6234.746078301" watchObservedRunningTime="2025-12-01 16:17:48.986532844 +0000 UTC m=+6234.750042447" Dec 01 16:17:49 crc kubenswrapper[4810]: I1201 16:17:49.986794 4810 generic.go:334] "Generic (PLEG): container finished" podID="89486609-b664-4b57-8776-c9b421186fab" containerID="5780a34684dc0843aa275ef4948d4f619741a71249fff8e5310309fa0b40399f" exitCode=0 Dec 01 16:17:49 crc kubenswrapper[4810]: I1201 16:17:49.986990 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-bcw9j" event={"ID":"89486609-b664-4b57-8776-c9b421186fab","Type":"ContainerDied","Data":"5780a34684dc0843aa275ef4948d4f619741a71249fff8e5310309fa0b40399f"} Dec 01 16:17:49 crc kubenswrapper[4810]: I1201 16:17:49.995501 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-qs2z8" event={"ID":"30053c9f-67a1-4ec1-b2bf-c61ff171a0d9","Type":"ContainerStarted","Data":"c1348d87640f37f54b0e730540bd49f6913e012e8b887b778bc4a07cd4050f49"} Dec 01 16:17:51 crc kubenswrapper[4810]: I1201 16:17:51.457025 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:51 crc kubenswrapper[4810]: I1201 16:17:51.741139 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:17:51 crc kubenswrapper[4810]: I1201 16:17:51.945601 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d8glv" Dec 01 16:17:51 crc kubenswrapper[4810]: I1201 16:17:51.945677 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d8glv" Dec 01 16:17:52 crc kubenswrapper[4810]: I1201 16:17:52.016160 4810 generic.go:334] "Generic (PLEG): container finished" podID="30053c9f-67a1-4ec1-b2bf-c61ff171a0d9" containerID="c1348d87640f37f54b0e730540bd49f6913e012e8b887b778bc4a07cd4050f49" exitCode=0 Dec 01 16:17:52 crc kubenswrapper[4810]: I1201 16:17:52.016249 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-qs2z8" event={"ID":"30053c9f-67a1-4ec1-b2bf-c61ff171a0d9","Type":"ContainerDied","Data":"c1348d87640f37f54b0e730540bd49f6913e012e8b887b778bc4a07cd4050f49"} Dec 01 16:17:52 crc kubenswrapper[4810]: I1201 16:17:52.020656 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-bcw9j" event={"ID":"89486609-b664-4b57-8776-c9b421186fab","Type":"ContainerStarted","Data":"e6ceddeec41cb81433b0b0fd627186faef0eba71d3692c4eaad6035e83b4bba2"} Dec 01 16:17:53 crc kubenswrapper[4810]: I1201 16:17:53.002947 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d8glv" podUID="1b286d0d-df8f-4501-9fa9-4436cb768aa5" containerName="registry-server" probeResult="failure" output=< Dec 01 16:17:53 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Dec 01 16:17:53 crc kubenswrapper[4810]: > Dec 01 16:17:53 crc kubenswrapper[4810]: I1201 16:17:53.062208 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-bcw9j" podStartSLOduration=7.062183786 podStartE2EDuration="7.062183786s" podCreationTimestamp="2025-12-01 16:17:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:17:53.055398892 +0000 UTC m=+6238.818908495" watchObservedRunningTime="2025-12-01 16:17:53.062183786 +0000 UTC m=+6238.825693389" Dec 01 16:17:56 crc kubenswrapper[4810]: I1201 16:17:56.075539 4810 generic.go:334] "Generic (PLEG): container finished" podID="89486609-b664-4b57-8776-c9b421186fab" containerID="e6ceddeec41cb81433b0b0fd627186faef0eba71d3692c4eaad6035e83b4bba2" exitCode=0 Dec 01 16:17:56 crc kubenswrapper[4810]: I1201 16:17:56.075737 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-bcw9j" event={"ID":"89486609-b664-4b57-8776-c9b421186fab","Type":"ContainerDied","Data":"e6ceddeec41cb81433b0b0fd627186faef0eba71d3692c4eaad6035e83b4bba2"} Dec 01 16:17:57 crc kubenswrapper[4810]: I1201 16:17:57.086821 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-699f5f4fd7-mk5nx" event={"ID":"b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0","Type":"ContainerStarted","Data":"f5bcf57a2563848964a597814fbf92a867df54ca7700bd64c902f38c6fd40f42"} Dec 01 16:17:57 crc kubenswrapper[4810]: I1201 16:17:57.090494 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-qs2z8" event={"ID":"30053c9f-67a1-4ec1-b2bf-c61ff171a0d9","Type":"ContainerStarted","Data":"0a2527a0baa77ac84810e4a2b998896a27016989cf64308da8f8fde38fb374be"} Dec 01 16:17:57 crc kubenswrapper[4810]: I1201 16:17:57.090917 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-qs2z8" Dec 01 16:17:57 crc kubenswrapper[4810]: I1201 16:17:57.130390 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-qs2z8" podStartSLOduration=2.328338506 podStartE2EDuration="14.130368235s" podCreationTimestamp="2025-12-01 16:17:43 +0000 UTC" firstStartedPulling="2025-12-01 16:17:44.583372453 +0000 UTC m=+6230.346882056" lastFinishedPulling="2025-12-01 16:17:56.385402172 +0000 UTC m=+6242.148911785" observedRunningTime="2025-12-01 16:17:57.122918133 +0000 UTC m=+6242.886427736" watchObservedRunningTime="2025-12-01 16:17:57.130368235 +0000 UTC m=+6242.893877828" Dec 01 16:17:58 crc kubenswrapper[4810]: I1201 16:17:58.101423 4810 generic.go:334] "Generic (PLEG): container finished" podID="b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0" containerID="f5bcf57a2563848964a597814fbf92a867df54ca7700bd64c902f38c6fd40f42" exitCode=0 Dec 01 16:17:58 crc kubenswrapper[4810]: I1201 16:17:58.101574 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-699f5f4fd7-mk5nx" event={"ID":"b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0","Type":"ContainerDied","Data":"f5bcf57a2563848964a597814fbf92a867df54ca7700bd64c902f38c6fd40f42"} Dec 01 16:17:58 crc kubenswrapper[4810]: I1201 16:17:58.104846 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-bcw9j" event={"ID":"89486609-b664-4b57-8776-c9b421186fab","Type":"ContainerDied","Data":"6e143ea5614a904cd1022d45628a21adf45aca15ae43e7a647144281b0d2c526"} Dec 01 16:17:58 crc kubenswrapper[4810]: I1201 16:17:58.104892 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e143ea5614a904cd1022d45628a21adf45aca15ae43e7a647144281b0d2c526" Dec 01 16:17:58 crc kubenswrapper[4810]: I1201 16:17:58.296207 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-bcw9j" Dec 01 16:17:58 crc kubenswrapper[4810]: I1201 16:17:58.422662 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89486609-b664-4b57-8776-c9b421186fab-combined-ca-bundle\") pod \"89486609-b664-4b57-8776-c9b421186fab\" (UID: \"89486609-b664-4b57-8776-c9b421186fab\") " Dec 01 16:17:58 crc kubenswrapper[4810]: I1201 16:17:58.422728 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89486609-b664-4b57-8776-c9b421186fab-config-data\") pod \"89486609-b664-4b57-8776-c9b421186fab\" (UID: \"89486609-b664-4b57-8776-c9b421186fab\") " Dec 01 16:17:58 crc kubenswrapper[4810]: I1201 16:17:58.422765 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89486609-b664-4b57-8776-c9b421186fab-scripts\") pod \"89486609-b664-4b57-8776-c9b421186fab\" (UID: \"89486609-b664-4b57-8776-c9b421186fab\") " Dec 01 16:17:58 crc kubenswrapper[4810]: I1201 16:17:58.422801 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/89486609-b664-4b57-8776-c9b421186fab-config-data-merged\") pod \"89486609-b664-4b57-8776-c9b421186fab\" (UID: \"89486609-b664-4b57-8776-c9b421186fab\") " Dec 01 16:17:58 crc kubenswrapper[4810]: I1201 16:17:58.448692 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89486609-b664-4b57-8776-c9b421186fab-config-data" (OuterVolumeSpecName: "config-data") pod "89486609-b664-4b57-8776-c9b421186fab" (UID: "89486609-b664-4b57-8776-c9b421186fab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:17:58 crc kubenswrapper[4810]: I1201 16:17:58.449263 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89486609-b664-4b57-8776-c9b421186fab-scripts" (OuterVolumeSpecName: "scripts") pod "89486609-b664-4b57-8776-c9b421186fab" (UID: "89486609-b664-4b57-8776-c9b421186fab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:17:58 crc kubenswrapper[4810]: I1201 16:17:58.451482 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89486609-b664-4b57-8776-c9b421186fab-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "89486609-b664-4b57-8776-c9b421186fab" (UID: "89486609-b664-4b57-8776-c9b421186fab"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:17:58 crc kubenswrapper[4810]: I1201 16:17:58.460186 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89486609-b664-4b57-8776-c9b421186fab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89486609-b664-4b57-8776-c9b421186fab" (UID: "89486609-b664-4b57-8776-c9b421186fab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:17:58 crc kubenswrapper[4810]: I1201 16:17:58.491611 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:17:58 crc kubenswrapper[4810]: E1201 16:17:58.491874 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:17:58 crc kubenswrapper[4810]: I1201 16:17:58.526068 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89486609-b664-4b57-8776-c9b421186fab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:58 crc kubenswrapper[4810]: I1201 16:17:58.526112 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89486609-b664-4b57-8776-c9b421186fab-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:58 crc kubenswrapper[4810]: I1201 16:17:58.526167 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89486609-b664-4b57-8776-c9b421186fab-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:58 crc kubenswrapper[4810]: I1201 16:17:58.526178 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/89486609-b664-4b57-8776-c9b421186fab-config-data-merged\") on node \"crc\" DevicePath \"\"" Dec 01 16:17:59 crc kubenswrapper[4810]: I1201 16:17:59.116010 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-699f5f4fd7-mk5nx" event={"ID":"b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0","Type":"ContainerStarted","Data":"28485e6f1b8597f7cbd985e18a93ee763e01698c5b57b4da347ed2e6374f4091"} Dec 01 16:17:59 crc kubenswrapper[4810]: I1201 16:17:59.116021 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-bcw9j" Dec 01 16:17:59 crc kubenswrapper[4810]: I1201 16:17:59.139604 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-699f5f4fd7-mk5nx" podStartSLOduration=4.083690096 podStartE2EDuration="15.139575309s" podCreationTimestamp="2025-12-01 16:17:44 +0000 UTC" firstStartedPulling="2025-12-01 16:17:45.410148337 +0000 UTC m=+6231.173657940" lastFinishedPulling="2025-12-01 16:17:56.46603355 +0000 UTC m=+6242.229543153" observedRunningTime="2025-12-01 16:17:59.127935503 +0000 UTC m=+6244.891445106" watchObservedRunningTime="2025-12-01 16:17:59.139575309 +0000 UTC m=+6244.903084912" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.236594 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-86b8958b56-x87pp"] Dec 01 16:18:00 crc kubenswrapper[4810]: E1201 16:18:00.237273 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89486609-b664-4b57-8776-c9b421186fab" containerName="init" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.237287 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="89486609-b664-4b57-8776-c9b421186fab" containerName="init" Dec 01 16:18:00 crc kubenswrapper[4810]: E1201 16:18:00.237319 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89486609-b664-4b57-8776-c9b421186fab" containerName="octavia-db-sync" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.237325 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="89486609-b664-4b57-8776-c9b421186fab" containerName="octavia-db-sync" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.237560 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="89486609-b664-4b57-8776-c9b421186fab" containerName="octavia-db-sync" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.238930 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.240793 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-octavia-public-svc" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.241397 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-octavia-internal-svc" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.249952 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-86b8958b56-x87pp"] Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.364434 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6821df4-0318-4d9c-8d47-116e602817fb-combined-ca-bundle\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.364515 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/a6821df4-0318-4d9c-8d47-116e602817fb-octavia-run\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.364636 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6821df4-0318-4d9c-8d47-116e602817fb-ovndb-tls-certs\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.364673 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/a6821df4-0318-4d9c-8d47-116e602817fb-config-data-merged\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.364740 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6821df4-0318-4d9c-8d47-116e602817fb-public-tls-certs\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.364842 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6821df4-0318-4d9c-8d47-116e602817fb-config-data\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.364929 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6821df4-0318-4d9c-8d47-116e602817fb-scripts\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.364962 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6821df4-0318-4d9c-8d47-116e602817fb-internal-tls-certs\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.466553 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/a6821df4-0318-4d9c-8d47-116e602817fb-octavia-run\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.466609 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6821df4-0318-4d9c-8d47-116e602817fb-combined-ca-bundle\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.466683 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6821df4-0318-4d9c-8d47-116e602817fb-ovndb-tls-certs\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.466702 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/a6821df4-0318-4d9c-8d47-116e602817fb-config-data-merged\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.466727 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6821df4-0318-4d9c-8d47-116e602817fb-public-tls-certs\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.466789 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6821df4-0318-4d9c-8d47-116e602817fb-config-data\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.466847 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6821df4-0318-4d9c-8d47-116e602817fb-scripts\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.466869 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6821df4-0318-4d9c-8d47-116e602817fb-internal-tls-certs\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.467153 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/a6821df4-0318-4d9c-8d47-116e602817fb-config-data-merged\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.467737 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/a6821df4-0318-4d9c-8d47-116e602817fb-octavia-run\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.471915 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6821df4-0318-4d9c-8d47-116e602817fb-config-data\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.471954 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6821df4-0318-4d9c-8d47-116e602817fb-combined-ca-bundle\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.471998 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6821df4-0318-4d9c-8d47-116e602817fb-scripts\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.480079 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6821df4-0318-4d9c-8d47-116e602817fb-internal-tls-certs\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.481624 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6821df4-0318-4d9c-8d47-116e602817fb-public-tls-certs\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.487553 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6821df4-0318-4d9c-8d47-116e602817fb-ovndb-tls-certs\") pod \"octavia-api-86b8958b56-x87pp\" (UID: \"a6821df4-0318-4d9c-8d47-116e602817fb\") " pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:00 crc kubenswrapper[4810]: I1201 16:18:00.564851 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:01 crc kubenswrapper[4810]: I1201 16:18:01.656180 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-86b8958b56-x87pp"] Dec 01 16:18:01 crc kubenswrapper[4810]: W1201 16:18:01.661287 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6821df4_0318_4d9c_8d47_116e602817fb.slice/crio-8a6540cdb858fc65da385b3972646aa0dc7a14309f932a845cbdffc0d8860e75 WatchSource:0}: Error finding container 8a6540cdb858fc65da385b3972646aa0dc7a14309f932a845cbdffc0d8860e75: Status 404 returned error can't find the container with id 8a6540cdb858fc65da385b3972646aa0dc7a14309f932a845cbdffc0d8860e75 Dec 01 16:18:02 crc kubenswrapper[4810]: I1201 16:18:02.143690 4810 generic.go:334] "Generic (PLEG): container finished" podID="a6821df4-0318-4d9c-8d47-116e602817fb" containerID="8549232b3390567217e592c471ec376896bec0e41e4d5711b24e6e6037d727e9" exitCode=0 Dec 01 16:18:02 crc kubenswrapper[4810]: I1201 16:18:02.143875 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-86b8958b56-x87pp" event={"ID":"a6821df4-0318-4d9c-8d47-116e602817fb","Type":"ContainerDied","Data":"8549232b3390567217e592c471ec376896bec0e41e4d5711b24e6e6037d727e9"} Dec 01 16:18:02 crc kubenswrapper[4810]: I1201 16:18:02.143974 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-86b8958b56-x87pp" event={"ID":"a6821df4-0318-4d9c-8d47-116e602817fb","Type":"ContainerStarted","Data":"8a6540cdb858fc65da385b3972646aa0dc7a14309f932a845cbdffc0d8860e75"} Dec 01 16:18:02 crc kubenswrapper[4810]: I1201 16:18:02.996025 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-d8glv" podUID="1b286d0d-df8f-4501-9fa9-4436cb768aa5" containerName="registry-server" probeResult="failure" output=< Dec 01 16:18:02 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Dec 01 16:18:02 crc kubenswrapper[4810]: > Dec 01 16:18:03 crc kubenswrapper[4810]: I1201 16:18:03.156148 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-86b8958b56-x87pp" event={"ID":"a6821df4-0318-4d9c-8d47-116e602817fb","Type":"ContainerStarted","Data":"5dc0a4d1e742acc79b42fbdec3a71ecd5252bd467ec48c637c5c7c740216e963"} Dec 01 16:18:03 crc kubenswrapper[4810]: I1201 16:18:03.156201 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-86b8958b56-x87pp" event={"ID":"a6821df4-0318-4d9c-8d47-116e602817fb","Type":"ContainerStarted","Data":"5c012df0c1ba3a82a6d7e1f36fecbb150a8c2852bac5546a8ebca9b556adcd14"} Dec 01 16:18:03 crc kubenswrapper[4810]: I1201 16:18:03.156286 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:03 crc kubenswrapper[4810]: I1201 16:18:03.186055 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-86b8958b56-x87pp" podStartSLOduration=3.186038718 podStartE2EDuration="3.186038718s" podCreationTimestamp="2025-12-01 16:18:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:18:03.185550026 +0000 UTC m=+6248.949059649" watchObservedRunningTime="2025-12-01 16:18:03.186038718 +0000 UTC m=+6248.949548331" Dec 01 16:18:04 crc kubenswrapper[4810]: I1201 16:18:04.167164 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:10 crc kubenswrapper[4810]: I1201 16:18:10.471958 4810 scope.go:117] "RemoveContainer" containerID="d31f441586726257a95607a056fe938965f790e148a7288e8d22984c6eaac50e" Dec 01 16:18:10 crc kubenswrapper[4810]: I1201 16:18:10.525756 4810 scope.go:117] "RemoveContainer" containerID="c2bb3483fe18662ea5bae4c150bf2664a3f69c896f0b0654b9a4e39c381eb3c3" Dec 01 16:18:10 crc kubenswrapper[4810]: I1201 16:18:10.560622 4810 scope.go:117] "RemoveContainer" containerID="2bd399de42e3ec006036ea6b138ef04cd65b13f1f8ef227bdf1dc3b4046c24bc" Dec 01 16:18:10 crc kubenswrapper[4810]: I1201 16:18:10.599065 4810 scope.go:117] "RemoveContainer" containerID="990e5b1039da24fd77b75aa90bdccc2baa567f6f1e4927c0333286693b7e9a03" Dec 01 16:18:10 crc kubenswrapper[4810]: I1201 16:18:10.628770 4810 scope.go:117] "RemoveContainer" containerID="a9ccc903529390bc2ba5cdda6d6e4b0fb29e6f47291bb54baf00ec6c2c910696" Dec 01 16:18:10 crc kubenswrapper[4810]: I1201 16:18:10.672748 4810 scope.go:117] "RemoveContainer" containerID="c6be535f8192f6d7b973867cf0fd2e46fb98d3297a6cf8ec785e44aae4cba1e6" Dec 01 16:18:10 crc kubenswrapper[4810]: I1201 16:18:10.694352 4810 scope.go:117] "RemoveContainer" containerID="c1119418971e85144659b81c9e22e1e466e311c7122e4a09f4cbbaf559b73f61" Dec 01 16:18:10 crc kubenswrapper[4810]: I1201 16:18:10.744663 4810 scope.go:117] "RemoveContainer" containerID="e12e791eb5079652c6fb5db332814d8bffc049a0a255bc677e536bfc4641e702" Dec 01 16:18:11 crc kubenswrapper[4810]: I1201 16:18:11.490900 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:18:11 crc kubenswrapper[4810]: E1201 16:18:11.492185 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:18:11 crc kubenswrapper[4810]: I1201 16:18:11.990663 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d8glv" Dec 01 16:18:12 crc kubenswrapper[4810]: I1201 16:18:12.056692 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d8glv" Dec 01 16:18:12 crc kubenswrapper[4810]: I1201 16:18:12.810964 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d8glv"] Dec 01 16:18:13 crc kubenswrapper[4810]: I1201 16:18:13.242243 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d8glv" podUID="1b286d0d-df8f-4501-9fa9-4436cb768aa5" containerName="registry-server" containerID="cri-o://50b44a5721e643dbbfb789bbde99ee25a28ad6e968015d9cc8c2b4abf3ad78c3" gracePeriod=2 Dec 01 16:18:13 crc kubenswrapper[4810]: I1201 16:18:13.754003 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d8glv" Dec 01 16:18:13 crc kubenswrapper[4810]: I1201 16:18:13.852424 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b286d0d-df8f-4501-9fa9-4436cb768aa5-utilities\") pod \"1b286d0d-df8f-4501-9fa9-4436cb768aa5\" (UID: \"1b286d0d-df8f-4501-9fa9-4436cb768aa5\") " Dec 01 16:18:13 crc kubenswrapper[4810]: I1201 16:18:13.852955 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b286d0d-df8f-4501-9fa9-4436cb768aa5-catalog-content\") pod \"1b286d0d-df8f-4501-9fa9-4436cb768aa5\" (UID: \"1b286d0d-df8f-4501-9fa9-4436cb768aa5\") " Dec 01 16:18:13 crc kubenswrapper[4810]: I1201 16:18:13.853101 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fhls\" (UniqueName: \"kubernetes.io/projected/1b286d0d-df8f-4501-9fa9-4436cb768aa5-kube-api-access-5fhls\") pod \"1b286d0d-df8f-4501-9fa9-4436cb768aa5\" (UID: \"1b286d0d-df8f-4501-9fa9-4436cb768aa5\") " Dec 01 16:18:13 crc kubenswrapper[4810]: I1201 16:18:13.853258 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b286d0d-df8f-4501-9fa9-4436cb768aa5-utilities" (OuterVolumeSpecName: "utilities") pod "1b286d0d-df8f-4501-9fa9-4436cb768aa5" (UID: "1b286d0d-df8f-4501-9fa9-4436cb768aa5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:18:13 crc kubenswrapper[4810]: I1201 16:18:13.853687 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b286d0d-df8f-4501-9fa9-4436cb768aa5-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:18:13 crc kubenswrapper[4810]: I1201 16:18:13.862868 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b286d0d-df8f-4501-9fa9-4436cb768aa5-kube-api-access-5fhls" (OuterVolumeSpecName: "kube-api-access-5fhls") pod "1b286d0d-df8f-4501-9fa9-4436cb768aa5" (UID: "1b286d0d-df8f-4501-9fa9-4436cb768aa5"). InnerVolumeSpecName "kube-api-access-5fhls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:18:13 crc kubenswrapper[4810]: I1201 16:18:13.956857 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fhls\" (UniqueName: \"kubernetes.io/projected/1b286d0d-df8f-4501-9fa9-4436cb768aa5-kube-api-access-5fhls\") on node \"crc\" DevicePath \"\"" Dec 01 16:18:13 crc kubenswrapper[4810]: I1201 16:18:13.960852 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b286d0d-df8f-4501-9fa9-4436cb768aa5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b286d0d-df8f-4501-9fa9-4436cb768aa5" (UID: "1b286d0d-df8f-4501-9fa9-4436cb768aa5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:18:13 crc kubenswrapper[4810]: I1201 16:18:13.976937 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-qs2z8" Dec 01 16:18:14 crc kubenswrapper[4810]: I1201 16:18:14.059256 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b286d0d-df8f-4501-9fa9-4436cb768aa5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:18:14 crc kubenswrapper[4810]: I1201 16:18:14.253343 4810 generic.go:334] "Generic (PLEG): container finished" podID="1b286d0d-df8f-4501-9fa9-4436cb768aa5" containerID="50b44a5721e643dbbfb789bbde99ee25a28ad6e968015d9cc8c2b4abf3ad78c3" exitCode=0 Dec 01 16:18:14 crc kubenswrapper[4810]: I1201 16:18:14.253414 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d8glv" Dec 01 16:18:14 crc kubenswrapper[4810]: I1201 16:18:14.253431 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8glv" event={"ID":"1b286d0d-df8f-4501-9fa9-4436cb768aa5","Type":"ContainerDied","Data":"50b44a5721e643dbbfb789bbde99ee25a28ad6e968015d9cc8c2b4abf3ad78c3"} Dec 01 16:18:14 crc kubenswrapper[4810]: I1201 16:18:14.253864 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8glv" event={"ID":"1b286d0d-df8f-4501-9fa9-4436cb768aa5","Type":"ContainerDied","Data":"a535eac415e3d613280793bc083f37cf6cc9bfde376421883d5df3ec156675f2"} Dec 01 16:18:14 crc kubenswrapper[4810]: I1201 16:18:14.253950 4810 scope.go:117] "RemoveContainer" containerID="50b44a5721e643dbbfb789bbde99ee25a28ad6e968015d9cc8c2b4abf3ad78c3" Dec 01 16:18:14 crc kubenswrapper[4810]: I1201 16:18:14.277077 4810 scope.go:117] "RemoveContainer" containerID="c79aef5b334e87f6b1ae6a9e0c86d772e32a2b7fa206b69e355b46696494eab5" Dec 01 16:18:14 crc kubenswrapper[4810]: I1201 16:18:14.305543 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d8glv"] Dec 01 16:18:14 crc kubenswrapper[4810]: I1201 16:18:14.314812 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d8glv"] Dec 01 16:18:14 crc kubenswrapper[4810]: I1201 16:18:14.349458 4810 scope.go:117] "RemoveContainer" containerID="368b8552c8dd9c98a6d00f67ecbdaecc9b0625b2d9134c5352da6aa772d78284" Dec 01 16:18:14 crc kubenswrapper[4810]: I1201 16:18:14.447567 4810 scope.go:117] "RemoveContainer" containerID="50b44a5721e643dbbfb789bbde99ee25a28ad6e968015d9cc8c2b4abf3ad78c3" Dec 01 16:18:14 crc kubenswrapper[4810]: E1201 16:18:14.448207 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50b44a5721e643dbbfb789bbde99ee25a28ad6e968015d9cc8c2b4abf3ad78c3\": container with ID starting with 50b44a5721e643dbbfb789bbde99ee25a28ad6e968015d9cc8c2b4abf3ad78c3 not found: ID does not exist" containerID="50b44a5721e643dbbfb789bbde99ee25a28ad6e968015d9cc8c2b4abf3ad78c3" Dec 01 16:18:14 crc kubenswrapper[4810]: I1201 16:18:14.448286 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50b44a5721e643dbbfb789bbde99ee25a28ad6e968015d9cc8c2b4abf3ad78c3"} err="failed to get container status \"50b44a5721e643dbbfb789bbde99ee25a28ad6e968015d9cc8c2b4abf3ad78c3\": rpc error: code = NotFound desc = could not find container \"50b44a5721e643dbbfb789bbde99ee25a28ad6e968015d9cc8c2b4abf3ad78c3\": container with ID starting with 50b44a5721e643dbbfb789bbde99ee25a28ad6e968015d9cc8c2b4abf3ad78c3 not found: ID does not exist" Dec 01 16:18:14 crc kubenswrapper[4810]: I1201 16:18:14.448330 4810 scope.go:117] "RemoveContainer" containerID="c79aef5b334e87f6b1ae6a9e0c86d772e32a2b7fa206b69e355b46696494eab5" Dec 01 16:18:14 crc kubenswrapper[4810]: E1201 16:18:14.448711 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c79aef5b334e87f6b1ae6a9e0c86d772e32a2b7fa206b69e355b46696494eab5\": container with ID starting with c79aef5b334e87f6b1ae6a9e0c86d772e32a2b7fa206b69e355b46696494eab5 not found: ID does not exist" containerID="c79aef5b334e87f6b1ae6a9e0c86d772e32a2b7fa206b69e355b46696494eab5" Dec 01 16:18:14 crc kubenswrapper[4810]: I1201 16:18:14.448740 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c79aef5b334e87f6b1ae6a9e0c86d772e32a2b7fa206b69e355b46696494eab5"} err="failed to get container status \"c79aef5b334e87f6b1ae6a9e0c86d772e32a2b7fa206b69e355b46696494eab5\": rpc error: code = NotFound desc = could not find container \"c79aef5b334e87f6b1ae6a9e0c86d772e32a2b7fa206b69e355b46696494eab5\": container with ID starting with c79aef5b334e87f6b1ae6a9e0c86d772e32a2b7fa206b69e355b46696494eab5 not found: ID does not exist" Dec 01 16:18:14 crc kubenswrapper[4810]: I1201 16:18:14.448760 4810 scope.go:117] "RemoveContainer" containerID="368b8552c8dd9c98a6d00f67ecbdaecc9b0625b2d9134c5352da6aa772d78284" Dec 01 16:18:14 crc kubenswrapper[4810]: E1201 16:18:14.449019 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"368b8552c8dd9c98a6d00f67ecbdaecc9b0625b2d9134c5352da6aa772d78284\": container with ID starting with 368b8552c8dd9c98a6d00f67ecbdaecc9b0625b2d9134c5352da6aa772d78284 not found: ID does not exist" containerID="368b8552c8dd9c98a6d00f67ecbdaecc9b0625b2d9134c5352da6aa772d78284" Dec 01 16:18:14 crc kubenswrapper[4810]: I1201 16:18:14.449040 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"368b8552c8dd9c98a6d00f67ecbdaecc9b0625b2d9134c5352da6aa772d78284"} err="failed to get container status \"368b8552c8dd9c98a6d00f67ecbdaecc9b0625b2d9134c5352da6aa772d78284\": rpc error: code = NotFound desc = could not find container \"368b8552c8dd9c98a6d00f67ecbdaecc9b0625b2d9134c5352da6aa772d78284\": container with ID starting with 368b8552c8dd9c98a6d00f67ecbdaecc9b0625b2d9134c5352da6aa772d78284 not found: ID does not exist" Dec 01 16:18:14 crc kubenswrapper[4810]: I1201 16:18:14.506550 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b286d0d-df8f-4501-9fa9-4436cb768aa5" path="/var/lib/kubelet/pods/1b286d0d-df8f-4501-9fa9-4436cb768aa5/volumes" Dec 01 16:18:19 crc kubenswrapper[4810]: I1201 16:18:19.989992 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:20 crc kubenswrapper[4810]: I1201 16:18:20.299709 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-86b8958b56-x87pp" Dec 01 16:18:20 crc kubenswrapper[4810]: I1201 16:18:20.389149 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-api-75d7f9cfbb-zdwbq"] Dec 01 16:18:20 crc kubenswrapper[4810]: I1201 16:18:20.390298 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-api-75d7f9cfbb-zdwbq" podUID="fb2847c5-3667-4b51-a1b7-784fcea0a897" containerName="octavia-api" containerID="cri-o://93cfcb2c6a52d38abd85a2c72025ec5a016e01f37d2531382f001961eb41ffeb" gracePeriod=30 Dec 01 16:18:20 crc kubenswrapper[4810]: I1201 16:18:20.390350 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-api-75d7f9cfbb-zdwbq" podUID="fb2847c5-3667-4b51-a1b7-784fcea0a897" containerName="octavia-api-provider-agent" containerID="cri-o://c001c32d77c6672cc553f637fdcbe0ee6887a96c667aaad887bd49951ee324fb" gracePeriod=30 Dec 01 16:18:21 crc kubenswrapper[4810]: I1201 16:18:21.338323 4810 generic.go:334] "Generic (PLEG): container finished" podID="fb2847c5-3667-4b51-a1b7-784fcea0a897" containerID="c001c32d77c6672cc553f637fdcbe0ee6887a96c667aaad887bd49951ee324fb" exitCode=0 Dec 01 16:18:21 crc kubenswrapper[4810]: I1201 16:18:21.338362 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-75d7f9cfbb-zdwbq" event={"ID":"fb2847c5-3667-4b51-a1b7-784fcea0a897","Type":"ContainerDied","Data":"c001c32d77c6672cc553f637fdcbe0ee6887a96c667aaad887bd49951ee324fb"} Dec 01 16:18:23 crc kubenswrapper[4810]: I1201 16:18:23.492050 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:18:23 crc kubenswrapper[4810]: E1201 16:18:23.492663 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.215363 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.289108 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/fb2847c5-3667-4b51-a1b7-784fcea0a897-config-data-merged\") pod \"fb2847c5-3667-4b51-a1b7-784fcea0a897\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.289188 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-scripts\") pod \"fb2847c5-3667-4b51-a1b7-784fcea0a897\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.289244 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-config-data\") pod \"fb2847c5-3667-4b51-a1b7-784fcea0a897\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.289271 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/fb2847c5-3667-4b51-a1b7-784fcea0a897-octavia-run\") pod \"fb2847c5-3667-4b51-a1b7-784fcea0a897\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.289507 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-combined-ca-bundle\") pod \"fb2847c5-3667-4b51-a1b7-784fcea0a897\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.289609 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-ovndb-tls-certs\") pod \"fb2847c5-3667-4b51-a1b7-784fcea0a897\" (UID: \"fb2847c5-3667-4b51-a1b7-784fcea0a897\") " Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.292290 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb2847c5-3667-4b51-a1b7-784fcea0a897-octavia-run" (OuterVolumeSpecName: "octavia-run") pod "fb2847c5-3667-4b51-a1b7-784fcea0a897" (UID: "fb2847c5-3667-4b51-a1b7-784fcea0a897"). InnerVolumeSpecName "octavia-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.300624 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-scripts" (OuterVolumeSpecName: "scripts") pod "fb2847c5-3667-4b51-a1b7-784fcea0a897" (UID: "fb2847c5-3667-4b51-a1b7-784fcea0a897"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.304178 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-config-data" (OuterVolumeSpecName: "config-data") pod "fb2847c5-3667-4b51-a1b7-784fcea0a897" (UID: "fb2847c5-3667-4b51-a1b7-784fcea0a897"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.368063 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb2847c5-3667-4b51-a1b7-784fcea0a897" (UID: "fb2847c5-3667-4b51-a1b7-784fcea0a897"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.370389 4810 generic.go:334] "Generic (PLEG): container finished" podID="fb2847c5-3667-4b51-a1b7-784fcea0a897" containerID="93cfcb2c6a52d38abd85a2c72025ec5a016e01f37d2531382f001961eb41ffeb" exitCode=0 Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.370443 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-75d7f9cfbb-zdwbq" event={"ID":"fb2847c5-3667-4b51-a1b7-784fcea0a897","Type":"ContainerDied","Data":"93cfcb2c6a52d38abd85a2c72025ec5a016e01f37d2531382f001961eb41ffeb"} Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.370454 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-75d7f9cfbb-zdwbq" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.370514 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-75d7f9cfbb-zdwbq" event={"ID":"fb2847c5-3667-4b51-a1b7-784fcea0a897","Type":"ContainerDied","Data":"ee0bd041bee15e99c0cb7f5ab43459b72f7394fd5867643f2d24a7304bbb952f"} Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.370528 4810 scope.go:117] "RemoveContainer" containerID="c001c32d77c6672cc553f637fdcbe0ee6887a96c667aaad887bd49951ee324fb" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.376242 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb2847c5-3667-4b51-a1b7-784fcea0a897-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "fb2847c5-3667-4b51-a1b7-784fcea0a897" (UID: "fb2847c5-3667-4b51-a1b7-784fcea0a897"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.392108 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.392140 4810 reconciler_common.go:293] "Volume detached for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/fb2847c5-3667-4b51-a1b7-784fcea0a897-octavia-run\") on node \"crc\" DevicePath \"\"" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.392150 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.392163 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/fb2847c5-3667-4b51-a1b7-784fcea0a897-config-data-merged\") on node \"crc\" DevicePath \"\"" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.392173 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.411439 4810 scope.go:117] "RemoveContainer" containerID="93cfcb2c6a52d38abd85a2c72025ec5a016e01f37d2531382f001961eb41ffeb" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.431818 4810 scope.go:117] "RemoveContainer" containerID="73bc563fa06154ba3d1541eeb709c72ca50bc76ad1a19b5de38f5681f40486b1" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.453618 4810 scope.go:117] "RemoveContainer" containerID="c001c32d77c6672cc553f637fdcbe0ee6887a96c667aaad887bd49951ee324fb" Dec 01 16:18:24 crc kubenswrapper[4810]: E1201 16:18:24.454102 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c001c32d77c6672cc553f637fdcbe0ee6887a96c667aaad887bd49951ee324fb\": container with ID starting with c001c32d77c6672cc553f637fdcbe0ee6887a96c667aaad887bd49951ee324fb not found: ID does not exist" containerID="c001c32d77c6672cc553f637fdcbe0ee6887a96c667aaad887bd49951ee324fb" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.454161 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c001c32d77c6672cc553f637fdcbe0ee6887a96c667aaad887bd49951ee324fb"} err="failed to get container status \"c001c32d77c6672cc553f637fdcbe0ee6887a96c667aaad887bd49951ee324fb\": rpc error: code = NotFound desc = could not find container \"c001c32d77c6672cc553f637fdcbe0ee6887a96c667aaad887bd49951ee324fb\": container with ID starting with c001c32d77c6672cc553f637fdcbe0ee6887a96c667aaad887bd49951ee324fb not found: ID does not exist" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.454195 4810 scope.go:117] "RemoveContainer" containerID="93cfcb2c6a52d38abd85a2c72025ec5a016e01f37d2531382f001961eb41ffeb" Dec 01 16:18:24 crc kubenswrapper[4810]: E1201 16:18:24.454442 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93cfcb2c6a52d38abd85a2c72025ec5a016e01f37d2531382f001961eb41ffeb\": container with ID starting with 93cfcb2c6a52d38abd85a2c72025ec5a016e01f37d2531382f001961eb41ffeb not found: ID does not exist" containerID="93cfcb2c6a52d38abd85a2c72025ec5a016e01f37d2531382f001961eb41ffeb" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.454496 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93cfcb2c6a52d38abd85a2c72025ec5a016e01f37d2531382f001961eb41ffeb"} err="failed to get container status \"93cfcb2c6a52d38abd85a2c72025ec5a016e01f37d2531382f001961eb41ffeb\": rpc error: code = NotFound desc = could not find container \"93cfcb2c6a52d38abd85a2c72025ec5a016e01f37d2531382f001961eb41ffeb\": container with ID starting with 93cfcb2c6a52d38abd85a2c72025ec5a016e01f37d2531382f001961eb41ffeb not found: ID does not exist" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.454518 4810 scope.go:117] "RemoveContainer" containerID="73bc563fa06154ba3d1541eeb709c72ca50bc76ad1a19b5de38f5681f40486b1" Dec 01 16:18:24 crc kubenswrapper[4810]: E1201 16:18:24.454785 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73bc563fa06154ba3d1541eeb709c72ca50bc76ad1a19b5de38f5681f40486b1\": container with ID starting with 73bc563fa06154ba3d1541eeb709c72ca50bc76ad1a19b5de38f5681f40486b1 not found: ID does not exist" containerID="73bc563fa06154ba3d1541eeb709c72ca50bc76ad1a19b5de38f5681f40486b1" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.454817 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73bc563fa06154ba3d1541eeb709c72ca50bc76ad1a19b5de38f5681f40486b1"} err="failed to get container status \"73bc563fa06154ba3d1541eeb709c72ca50bc76ad1a19b5de38f5681f40486b1\": rpc error: code = NotFound desc = could not find container \"73bc563fa06154ba3d1541eeb709c72ca50bc76ad1a19b5de38f5681f40486b1\": container with ID starting with 73bc563fa06154ba3d1541eeb709c72ca50bc76ad1a19b5de38f5681f40486b1 not found: ID does not exist" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.473105 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "fb2847c5-3667-4b51-a1b7-784fcea0a897" (UID: "fb2847c5-3667-4b51-a1b7-784fcea0a897"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.494876 4810 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb2847c5-3667-4b51-a1b7-784fcea0a897-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.695864 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-api-75d7f9cfbb-zdwbq"] Dec 01 16:18:24 crc kubenswrapper[4810]: I1201 16:18:24.705425 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-api-75d7f9cfbb-zdwbq"] Dec 01 16:18:25 crc kubenswrapper[4810]: I1201 16:18:25.425569 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-699f5f4fd7-mk5nx"] Dec 01 16:18:25 crc kubenswrapper[4810]: I1201 16:18:25.425838 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-699f5f4fd7-mk5nx" podUID="b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0" containerName="octavia-amphora-httpd" containerID="cri-o://28485e6f1b8597f7cbd985e18a93ee763e01698c5b57b4da347ed2e6374f4091" gracePeriod=30 Dec 01 16:18:25 crc kubenswrapper[4810]: I1201 16:18:25.952792 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-699f5f4fd7-mk5nx" Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.023986 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0-httpd-config\") pod \"b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0\" (UID: \"b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0\") " Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.024215 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0-amphora-image\") pod \"b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0\" (UID: \"b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0\") " Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.071829 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0" (UID: "b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.129717 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.138810 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0" (UID: "b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.230979 4810 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0-amphora-image\") on node \"crc\" DevicePath \"\"" Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.388137 4810 generic.go:334] "Generic (PLEG): container finished" podID="b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0" containerID="28485e6f1b8597f7cbd985e18a93ee763e01698c5b57b4da347ed2e6374f4091" exitCode=0 Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.388677 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-699f5f4fd7-mk5nx" event={"ID":"b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0","Type":"ContainerDied","Data":"28485e6f1b8597f7cbd985e18a93ee763e01698c5b57b4da347ed2e6374f4091"} Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.388710 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-699f5f4fd7-mk5nx" Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.388884 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-699f5f4fd7-mk5nx" event={"ID":"b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0","Type":"ContainerDied","Data":"0c468156da91da6cd949e5c719da0a910ff7af42b43435c84739569cce1adfb6"} Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.388965 4810 scope.go:117] "RemoveContainer" containerID="28485e6f1b8597f7cbd985e18a93ee763e01698c5b57b4da347ed2e6374f4091" Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.410443 4810 scope.go:117] "RemoveContainer" containerID="f5bcf57a2563848964a597814fbf92a867df54ca7700bd64c902f38c6fd40f42" Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.449123 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-699f5f4fd7-mk5nx"] Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.451245 4810 scope.go:117] "RemoveContainer" containerID="28485e6f1b8597f7cbd985e18a93ee763e01698c5b57b4da347ed2e6374f4091" Dec 01 16:18:26 crc kubenswrapper[4810]: E1201 16:18:26.451902 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28485e6f1b8597f7cbd985e18a93ee763e01698c5b57b4da347ed2e6374f4091\": container with ID starting with 28485e6f1b8597f7cbd985e18a93ee763e01698c5b57b4da347ed2e6374f4091 not found: ID does not exist" containerID="28485e6f1b8597f7cbd985e18a93ee763e01698c5b57b4da347ed2e6374f4091" Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.451958 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28485e6f1b8597f7cbd985e18a93ee763e01698c5b57b4da347ed2e6374f4091"} err="failed to get container status \"28485e6f1b8597f7cbd985e18a93ee763e01698c5b57b4da347ed2e6374f4091\": rpc error: code = NotFound desc = could not find container \"28485e6f1b8597f7cbd985e18a93ee763e01698c5b57b4da347ed2e6374f4091\": container with ID starting with 28485e6f1b8597f7cbd985e18a93ee763e01698c5b57b4da347ed2e6374f4091 not found: ID does not exist" Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.452010 4810 scope.go:117] "RemoveContainer" containerID="f5bcf57a2563848964a597814fbf92a867df54ca7700bd64c902f38c6fd40f42" Dec 01 16:18:26 crc kubenswrapper[4810]: E1201 16:18:26.452501 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5bcf57a2563848964a597814fbf92a867df54ca7700bd64c902f38c6fd40f42\": container with ID starting with f5bcf57a2563848964a597814fbf92a867df54ca7700bd64c902f38c6fd40f42 not found: ID does not exist" containerID="f5bcf57a2563848964a597814fbf92a867df54ca7700bd64c902f38c6fd40f42" Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.452532 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5bcf57a2563848964a597814fbf92a867df54ca7700bd64c902f38c6fd40f42"} err="failed to get container status \"f5bcf57a2563848964a597814fbf92a867df54ca7700bd64c902f38c6fd40f42\": rpc error: code = NotFound desc = could not find container \"f5bcf57a2563848964a597814fbf92a867df54ca7700bd64c902f38c6fd40f42\": container with ID starting with f5bcf57a2563848964a597814fbf92a867df54ca7700bd64c902f38c6fd40f42 not found: ID does not exist" Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.461088 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-699f5f4fd7-mk5nx"] Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.505088 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0" path="/var/lib/kubelet/pods/b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0/volumes" Dec 01 16:18:26 crc kubenswrapper[4810]: I1201 16:18:26.505690 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb2847c5-3667-4b51-a1b7-784fcea0a897" path="/var/lib/kubelet/pods/fb2847c5-3667-4b51-a1b7-784fcea0a897/volumes" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.426792 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-699f5f4fd7-ck6rq"] Dec 01 16:18:33 crc kubenswrapper[4810]: E1201 16:18:33.427926 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0" containerName="init" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.427944 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0" containerName="init" Dec 01 16:18:33 crc kubenswrapper[4810]: E1201 16:18:33.427960 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb2847c5-3667-4b51-a1b7-784fcea0a897" containerName="octavia-api-provider-agent" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.427967 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb2847c5-3667-4b51-a1b7-784fcea0a897" containerName="octavia-api-provider-agent" Dec 01 16:18:33 crc kubenswrapper[4810]: E1201 16:18:33.427988 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b286d0d-df8f-4501-9fa9-4436cb768aa5" containerName="extract-utilities" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.427997 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b286d0d-df8f-4501-9fa9-4436cb768aa5" containerName="extract-utilities" Dec 01 16:18:33 crc kubenswrapper[4810]: E1201 16:18:33.428020 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b286d0d-df8f-4501-9fa9-4436cb768aa5" containerName="registry-server" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.428028 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b286d0d-df8f-4501-9fa9-4436cb768aa5" containerName="registry-server" Dec 01 16:18:33 crc kubenswrapper[4810]: E1201 16:18:33.428044 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0" containerName="octavia-amphora-httpd" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.428051 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0" containerName="octavia-amphora-httpd" Dec 01 16:18:33 crc kubenswrapper[4810]: E1201 16:18:33.428065 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb2847c5-3667-4b51-a1b7-784fcea0a897" containerName="octavia-api" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.428074 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb2847c5-3667-4b51-a1b7-784fcea0a897" containerName="octavia-api" Dec 01 16:18:33 crc kubenswrapper[4810]: E1201 16:18:33.428090 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b286d0d-df8f-4501-9fa9-4436cb768aa5" containerName="extract-content" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.428097 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b286d0d-df8f-4501-9fa9-4436cb768aa5" containerName="extract-content" Dec 01 16:18:33 crc kubenswrapper[4810]: E1201 16:18:33.428132 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb2847c5-3667-4b51-a1b7-784fcea0a897" containerName="init" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.428140 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb2847c5-3667-4b51-a1b7-784fcea0a897" containerName="init" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.428413 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb2847c5-3667-4b51-a1b7-784fcea0a897" containerName="octavia-api" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.428430 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb2847c5-3667-4b51-a1b7-784fcea0a897" containerName="octavia-api-provider-agent" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.428450 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9b6f063-3e0a-4e8f-bcfe-31034b0f4ac0" containerName="octavia-amphora-httpd" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.428488 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b286d0d-df8f-4501-9fa9-4436cb768aa5" containerName="registry-server" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.429707 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-699f5f4fd7-ck6rq" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.433081 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.437638 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-699f5f4fd7-ck6rq"] Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.570952 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/883efbb3-cd1c-45d5-a17e-482ca1b8eb9e-amphora-image\") pod \"octavia-image-upload-699f5f4fd7-ck6rq\" (UID: \"883efbb3-cd1c-45d5-a17e-482ca1b8eb9e\") " pod="openstack/octavia-image-upload-699f5f4fd7-ck6rq" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.571227 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/883efbb3-cd1c-45d5-a17e-482ca1b8eb9e-httpd-config\") pod \"octavia-image-upload-699f5f4fd7-ck6rq\" (UID: \"883efbb3-cd1c-45d5-a17e-482ca1b8eb9e\") " pod="openstack/octavia-image-upload-699f5f4fd7-ck6rq" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.673003 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/883efbb3-cd1c-45d5-a17e-482ca1b8eb9e-amphora-image\") pod \"octavia-image-upload-699f5f4fd7-ck6rq\" (UID: \"883efbb3-cd1c-45d5-a17e-482ca1b8eb9e\") " pod="openstack/octavia-image-upload-699f5f4fd7-ck6rq" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.673130 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/883efbb3-cd1c-45d5-a17e-482ca1b8eb9e-httpd-config\") pod \"octavia-image-upload-699f5f4fd7-ck6rq\" (UID: \"883efbb3-cd1c-45d5-a17e-482ca1b8eb9e\") " pod="openstack/octavia-image-upload-699f5f4fd7-ck6rq" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.673677 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/883efbb3-cd1c-45d5-a17e-482ca1b8eb9e-amphora-image\") pod \"octavia-image-upload-699f5f4fd7-ck6rq\" (UID: \"883efbb3-cd1c-45d5-a17e-482ca1b8eb9e\") " pod="openstack/octavia-image-upload-699f5f4fd7-ck6rq" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.679615 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/883efbb3-cd1c-45d5-a17e-482ca1b8eb9e-httpd-config\") pod \"octavia-image-upload-699f5f4fd7-ck6rq\" (UID: \"883efbb3-cd1c-45d5-a17e-482ca1b8eb9e\") " pod="openstack/octavia-image-upload-699f5f4fd7-ck6rq" Dec 01 16:18:33 crc kubenswrapper[4810]: I1201 16:18:33.751709 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-699f5f4fd7-ck6rq" Dec 01 16:18:34 crc kubenswrapper[4810]: I1201 16:18:34.239108 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-699f5f4fd7-ck6rq"] Dec 01 16:18:34 crc kubenswrapper[4810]: I1201 16:18:34.464948 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-699f5f4fd7-ck6rq" event={"ID":"883efbb3-cd1c-45d5-a17e-482ca1b8eb9e","Type":"ContainerStarted","Data":"1d078a9d0a32577301fb2df384750e408a3647bc46f94b7d08299e3317c1ea29"} Dec 01 16:18:34 crc kubenswrapper[4810]: I1201 16:18:34.504511 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:18:34 crc kubenswrapper[4810]: E1201 16:18:34.504750 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:18:35 crc kubenswrapper[4810]: I1201 16:18:35.474996 4810 generic.go:334] "Generic (PLEG): container finished" podID="883efbb3-cd1c-45d5-a17e-482ca1b8eb9e" containerID="dd161cd80a48c5cfedd6622d92cceed284f56e56c17b6bdf14294f31fe35ed91" exitCode=0 Dec 01 16:18:35 crc kubenswrapper[4810]: I1201 16:18:35.475366 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-699f5f4fd7-ck6rq" event={"ID":"883efbb3-cd1c-45d5-a17e-482ca1b8eb9e","Type":"ContainerDied","Data":"dd161cd80a48c5cfedd6622d92cceed284f56e56c17b6bdf14294f31fe35ed91"} Dec 01 16:18:36 crc kubenswrapper[4810]: I1201 16:18:36.484632 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-699f5f4fd7-ck6rq" event={"ID":"883efbb3-cd1c-45d5-a17e-482ca1b8eb9e","Type":"ContainerStarted","Data":"9491fa03d0e8f65a9843c0f0d34ca368fe201eb555b38c54dbfc6862d4d1dc6b"} Dec 01 16:18:36 crc kubenswrapper[4810]: I1201 16:18:36.511075 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-699f5f4fd7-ck6rq" podStartSLOduration=2.990735714 podStartE2EDuration="3.51105069s" podCreationTimestamp="2025-12-01 16:18:33 +0000 UTC" firstStartedPulling="2025-12-01 16:18:34.243130832 +0000 UTC m=+6280.006640435" lastFinishedPulling="2025-12-01 16:18:34.763445808 +0000 UTC m=+6280.526955411" observedRunningTime="2025-12-01 16:18:36.506781424 +0000 UTC m=+6282.270291027" watchObservedRunningTime="2025-12-01 16:18:36.51105069 +0000 UTC m=+6282.274560293" Dec 01 16:18:46 crc kubenswrapper[4810]: I1201 16:18:46.491971 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:18:46 crc kubenswrapper[4810]: E1201 16:18:46.492747 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.029567 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-frfm6"] Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.031526 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.035049 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.036206 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.036528 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.042809 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-frfm6"] Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.156117 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b-config-data-merged\") pod \"octavia-healthmanager-frfm6\" (UID: \"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b\") " pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.156172 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b-combined-ca-bundle\") pod \"octavia-healthmanager-frfm6\" (UID: \"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b\") " pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.156192 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b-config-data\") pod \"octavia-healthmanager-frfm6\" (UID: \"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b\") " pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.156213 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b-hm-ports\") pod \"octavia-healthmanager-frfm6\" (UID: \"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b\") " pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.156285 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b-amphora-certs\") pod \"octavia-healthmanager-frfm6\" (UID: \"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b\") " pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.156357 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b-scripts\") pod \"octavia-healthmanager-frfm6\" (UID: \"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b\") " pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.258371 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b-scripts\") pod \"octavia-healthmanager-frfm6\" (UID: \"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b\") " pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.259522 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b-config-data-merged\") pod \"octavia-healthmanager-frfm6\" (UID: \"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b\") " pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.259569 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b-combined-ca-bundle\") pod \"octavia-healthmanager-frfm6\" (UID: \"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b\") " pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.259598 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b-config-data\") pod \"octavia-healthmanager-frfm6\" (UID: \"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b\") " pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.259624 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b-hm-ports\") pod \"octavia-healthmanager-frfm6\" (UID: \"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b\") " pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.259707 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b-amphora-certs\") pod \"octavia-healthmanager-frfm6\" (UID: \"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b\") " pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.260007 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b-config-data-merged\") pod \"octavia-healthmanager-frfm6\" (UID: \"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b\") " pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.261027 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b-hm-ports\") pod \"octavia-healthmanager-frfm6\" (UID: \"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b\") " pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.266333 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b-combined-ca-bundle\") pod \"octavia-healthmanager-frfm6\" (UID: \"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b\") " pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.266889 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b-scripts\") pod \"octavia-healthmanager-frfm6\" (UID: \"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b\") " pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.267836 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b-config-data\") pod \"octavia-healthmanager-frfm6\" (UID: \"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b\") " pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.268438 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b-amphora-certs\") pod \"octavia-healthmanager-frfm6\" (UID: \"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b\") " pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:49 crc kubenswrapper[4810]: I1201 16:18:49.362344 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:50 crc kubenswrapper[4810]: I1201 16:18:50.021346 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-frfm6"] Dec 01 16:18:50 crc kubenswrapper[4810]: I1201 16:18:50.613989 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-frfm6" event={"ID":"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b","Type":"ContainerStarted","Data":"7a2db8c6902f1429a370bd47ce20dae6b0eddc390682a608d37f0a43128232e8"} Dec 01 16:18:50 crc kubenswrapper[4810]: I1201 16:18:50.614030 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-frfm6" event={"ID":"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b","Type":"ContainerStarted","Data":"b49bb10ef5f90c89a4e4e998b3a358ebfa9a2e53d7f6226dc489de4604547e2f"} Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.190490 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-qthjz"] Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.192174 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.194086 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.194395 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.206230 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-qthjz"] Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.300530 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/a88e2d45-97f3-4c58-ade1-9ce92b0897a8-hm-ports\") pod \"octavia-housekeeping-qthjz\" (UID: \"a88e2d45-97f3-4c58-ade1-9ce92b0897a8\") " pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.300899 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a88e2d45-97f3-4c58-ade1-9ce92b0897a8-combined-ca-bundle\") pod \"octavia-housekeeping-qthjz\" (UID: \"a88e2d45-97f3-4c58-ade1-9ce92b0897a8\") " pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.300931 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/a88e2d45-97f3-4c58-ade1-9ce92b0897a8-amphora-certs\") pod \"octavia-housekeeping-qthjz\" (UID: \"a88e2d45-97f3-4c58-ade1-9ce92b0897a8\") " pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.300954 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a88e2d45-97f3-4c58-ade1-9ce92b0897a8-config-data\") pod \"octavia-housekeeping-qthjz\" (UID: \"a88e2d45-97f3-4c58-ade1-9ce92b0897a8\") " pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.301047 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a88e2d45-97f3-4c58-ade1-9ce92b0897a8-scripts\") pod \"octavia-housekeeping-qthjz\" (UID: \"a88e2d45-97f3-4c58-ade1-9ce92b0897a8\") " pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.301155 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/a88e2d45-97f3-4c58-ade1-9ce92b0897a8-config-data-merged\") pod \"octavia-housekeeping-qthjz\" (UID: \"a88e2d45-97f3-4c58-ade1-9ce92b0897a8\") " pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.403381 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a88e2d45-97f3-4c58-ade1-9ce92b0897a8-combined-ca-bundle\") pod \"octavia-housekeeping-qthjz\" (UID: \"a88e2d45-97f3-4c58-ade1-9ce92b0897a8\") " pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.403705 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/a88e2d45-97f3-4c58-ade1-9ce92b0897a8-amphora-certs\") pod \"octavia-housekeeping-qthjz\" (UID: \"a88e2d45-97f3-4c58-ade1-9ce92b0897a8\") " pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.403870 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a88e2d45-97f3-4c58-ade1-9ce92b0897a8-config-data\") pod \"octavia-housekeeping-qthjz\" (UID: \"a88e2d45-97f3-4c58-ade1-9ce92b0897a8\") " pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.403984 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a88e2d45-97f3-4c58-ade1-9ce92b0897a8-scripts\") pod \"octavia-housekeeping-qthjz\" (UID: \"a88e2d45-97f3-4c58-ade1-9ce92b0897a8\") " pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.404118 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/a88e2d45-97f3-4c58-ade1-9ce92b0897a8-config-data-merged\") pod \"octavia-housekeeping-qthjz\" (UID: \"a88e2d45-97f3-4c58-ade1-9ce92b0897a8\") " pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.404315 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/a88e2d45-97f3-4c58-ade1-9ce92b0897a8-hm-ports\") pod \"octavia-housekeeping-qthjz\" (UID: \"a88e2d45-97f3-4c58-ade1-9ce92b0897a8\") " pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.404646 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/a88e2d45-97f3-4c58-ade1-9ce92b0897a8-config-data-merged\") pod \"octavia-housekeeping-qthjz\" (UID: \"a88e2d45-97f3-4c58-ade1-9ce92b0897a8\") " pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.405631 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/a88e2d45-97f3-4c58-ade1-9ce92b0897a8-hm-ports\") pod \"octavia-housekeeping-qthjz\" (UID: \"a88e2d45-97f3-4c58-ade1-9ce92b0897a8\") " pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.410275 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a88e2d45-97f3-4c58-ade1-9ce92b0897a8-combined-ca-bundle\") pod \"octavia-housekeeping-qthjz\" (UID: \"a88e2d45-97f3-4c58-ade1-9ce92b0897a8\") " pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.413424 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a88e2d45-97f3-4c58-ade1-9ce92b0897a8-scripts\") pod \"octavia-housekeeping-qthjz\" (UID: \"a88e2d45-97f3-4c58-ade1-9ce92b0897a8\") " pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.413550 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a88e2d45-97f3-4c58-ade1-9ce92b0897a8-config-data\") pod \"octavia-housekeeping-qthjz\" (UID: \"a88e2d45-97f3-4c58-ade1-9ce92b0897a8\") " pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.417940 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/a88e2d45-97f3-4c58-ade1-9ce92b0897a8-amphora-certs\") pod \"octavia-housekeeping-qthjz\" (UID: \"a88e2d45-97f3-4c58-ade1-9ce92b0897a8\") " pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:51 crc kubenswrapper[4810]: I1201 16:18:51.518228 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:52 crc kubenswrapper[4810]: I1201 16:18:52.212456 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-qthjz"] Dec 01 16:18:52 crc kubenswrapper[4810]: I1201 16:18:52.638778 4810 generic.go:334] "Generic (PLEG): container finished" podID="6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b" containerID="7a2db8c6902f1429a370bd47ce20dae6b0eddc390682a608d37f0a43128232e8" exitCode=0 Dec 01 16:18:52 crc kubenswrapper[4810]: I1201 16:18:52.638876 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-frfm6" event={"ID":"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b","Type":"ContainerDied","Data":"7a2db8c6902f1429a370bd47ce20dae6b0eddc390682a608d37f0a43128232e8"} Dec 01 16:18:52 crc kubenswrapper[4810]: I1201 16:18:52.640665 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-qthjz" event={"ID":"a88e2d45-97f3-4c58-ade1-9ce92b0897a8","Type":"ContainerStarted","Data":"4e7ff49bbf0631613dd641496bdd45031087b752149b46fd85bb959883e10d89"} Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.046301 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-4gnz2"] Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.050122 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.053645 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.053874 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.066881 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-4gnz2"] Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.146353 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/8e768084-da29-4e95-ae6e-feacb0cbfb70-config-data-merged\") pod \"octavia-worker-4gnz2\" (UID: \"8e768084-da29-4e95-ae6e-feacb0cbfb70\") " pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.146447 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e768084-da29-4e95-ae6e-feacb0cbfb70-scripts\") pod \"octavia-worker-4gnz2\" (UID: \"8e768084-da29-4e95-ae6e-feacb0cbfb70\") " pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.146560 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e768084-da29-4e95-ae6e-feacb0cbfb70-config-data\") pod \"octavia-worker-4gnz2\" (UID: \"8e768084-da29-4e95-ae6e-feacb0cbfb70\") " pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.146597 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/8e768084-da29-4e95-ae6e-feacb0cbfb70-amphora-certs\") pod \"octavia-worker-4gnz2\" (UID: \"8e768084-da29-4e95-ae6e-feacb0cbfb70\") " pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.146621 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/8e768084-da29-4e95-ae6e-feacb0cbfb70-hm-ports\") pod \"octavia-worker-4gnz2\" (UID: \"8e768084-da29-4e95-ae6e-feacb0cbfb70\") " pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.146659 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e768084-da29-4e95-ae6e-feacb0cbfb70-combined-ca-bundle\") pod \"octavia-worker-4gnz2\" (UID: \"8e768084-da29-4e95-ae6e-feacb0cbfb70\") " pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.249075 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e768084-da29-4e95-ae6e-feacb0cbfb70-config-data\") pod \"octavia-worker-4gnz2\" (UID: \"8e768084-da29-4e95-ae6e-feacb0cbfb70\") " pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.249126 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/8e768084-da29-4e95-ae6e-feacb0cbfb70-amphora-certs\") pod \"octavia-worker-4gnz2\" (UID: \"8e768084-da29-4e95-ae6e-feacb0cbfb70\") " pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.249150 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/8e768084-da29-4e95-ae6e-feacb0cbfb70-hm-ports\") pod \"octavia-worker-4gnz2\" (UID: \"8e768084-da29-4e95-ae6e-feacb0cbfb70\") " pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.249181 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e768084-da29-4e95-ae6e-feacb0cbfb70-combined-ca-bundle\") pod \"octavia-worker-4gnz2\" (UID: \"8e768084-da29-4e95-ae6e-feacb0cbfb70\") " pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.249253 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/8e768084-da29-4e95-ae6e-feacb0cbfb70-config-data-merged\") pod \"octavia-worker-4gnz2\" (UID: \"8e768084-da29-4e95-ae6e-feacb0cbfb70\") " pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.249305 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e768084-da29-4e95-ae6e-feacb0cbfb70-scripts\") pod \"octavia-worker-4gnz2\" (UID: \"8e768084-da29-4e95-ae6e-feacb0cbfb70\") " pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.250278 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/8e768084-da29-4e95-ae6e-feacb0cbfb70-hm-ports\") pod \"octavia-worker-4gnz2\" (UID: \"8e768084-da29-4e95-ae6e-feacb0cbfb70\") " pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.250311 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/8e768084-da29-4e95-ae6e-feacb0cbfb70-config-data-merged\") pod \"octavia-worker-4gnz2\" (UID: \"8e768084-da29-4e95-ae6e-feacb0cbfb70\") " pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.254903 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/8e768084-da29-4e95-ae6e-feacb0cbfb70-amphora-certs\") pod \"octavia-worker-4gnz2\" (UID: \"8e768084-da29-4e95-ae6e-feacb0cbfb70\") " pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.255056 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e768084-da29-4e95-ae6e-feacb0cbfb70-scripts\") pod \"octavia-worker-4gnz2\" (UID: \"8e768084-da29-4e95-ae6e-feacb0cbfb70\") " pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.255211 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e768084-da29-4e95-ae6e-feacb0cbfb70-combined-ca-bundle\") pod \"octavia-worker-4gnz2\" (UID: \"8e768084-da29-4e95-ae6e-feacb0cbfb70\") " pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.262243 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e768084-da29-4e95-ae6e-feacb0cbfb70-config-data\") pod \"octavia-worker-4gnz2\" (UID: \"8e768084-da29-4e95-ae6e-feacb0cbfb70\") " pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.373822 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.657247 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-frfm6" event={"ID":"6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b","Type":"ContainerStarted","Data":"7a202f8f15019fb81dc6803bc394f89c9cc1ac0511c88c429e886ce401ecc511"} Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.657509 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:18:53 crc kubenswrapper[4810]: I1201 16:18:53.678108 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-frfm6" podStartSLOduration=5.678086042 podStartE2EDuration="5.678086042s" podCreationTimestamp="2025-12-01 16:18:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:18:53.676089427 +0000 UTC m=+6299.439599060" watchObservedRunningTime="2025-12-01 16:18:53.678086042 +0000 UTC m=+6299.441595645" Dec 01 16:18:54 crc kubenswrapper[4810]: I1201 16:18:54.435601 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-4gnz2"] Dec 01 16:18:54 crc kubenswrapper[4810]: I1201 16:18:54.668393 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-qthjz" event={"ID":"a88e2d45-97f3-4c58-ade1-9ce92b0897a8","Type":"ContainerStarted","Data":"bd8183c3a64fea79a07a916721fd17f9b39971fd8d8a9e774ecbfe2fd7d8c9fb"} Dec 01 16:18:54 crc kubenswrapper[4810]: I1201 16:18:54.670533 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-4gnz2" event={"ID":"8e768084-da29-4e95-ae6e-feacb0cbfb70","Type":"ContainerStarted","Data":"e48547bc57d8f529afcd2b6345b31199fb04715dc4c63c83283d460ba4dc1ee9"} Dec 01 16:18:55 crc kubenswrapper[4810]: I1201 16:18:55.698821 4810 generic.go:334] "Generic (PLEG): container finished" podID="a88e2d45-97f3-4c58-ade1-9ce92b0897a8" containerID="bd8183c3a64fea79a07a916721fd17f9b39971fd8d8a9e774ecbfe2fd7d8c9fb" exitCode=0 Dec 01 16:18:55 crc kubenswrapper[4810]: I1201 16:18:55.698924 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-qthjz" event={"ID":"a88e2d45-97f3-4c58-ade1-9ce92b0897a8","Type":"ContainerDied","Data":"bd8183c3a64fea79a07a916721fd17f9b39971fd8d8a9e774ecbfe2fd7d8c9fb"} Dec 01 16:18:56 crc kubenswrapper[4810]: I1201 16:18:56.713710 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-qthjz" event={"ID":"a88e2d45-97f3-4c58-ade1-9ce92b0897a8","Type":"ContainerStarted","Data":"78221eb698543389a60fa06fe38b11c46b79c0a7d006ad501e5a3868b16f3d61"} Dec 01 16:18:56 crc kubenswrapper[4810]: I1201 16:18:56.714318 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:18:56 crc kubenswrapper[4810]: I1201 16:18:56.742938 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-qthjz" podStartSLOduration=4.072036204 podStartE2EDuration="5.742917073s" podCreationTimestamp="2025-12-01 16:18:51 +0000 UTC" firstStartedPulling="2025-12-01 16:18:52.185378329 +0000 UTC m=+6297.948887932" lastFinishedPulling="2025-12-01 16:18:53.856259198 +0000 UTC m=+6299.619768801" observedRunningTime="2025-12-01 16:18:56.737043843 +0000 UTC m=+6302.500553446" watchObservedRunningTime="2025-12-01 16:18:56.742917073 +0000 UTC m=+6302.506426686" Dec 01 16:18:57 crc kubenswrapper[4810]: I1201 16:18:57.724073 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-4gnz2" event={"ID":"8e768084-da29-4e95-ae6e-feacb0cbfb70","Type":"ContainerStarted","Data":"df728e70e0bfa7b885f9383efd0f8ca5b80e51de177005bb4b88c15feaf197d0"} Dec 01 16:18:58 crc kubenswrapper[4810]: I1201 16:18:58.737158 4810 generic.go:334] "Generic (PLEG): container finished" podID="8e768084-da29-4e95-ae6e-feacb0cbfb70" containerID="df728e70e0bfa7b885f9383efd0f8ca5b80e51de177005bb4b88c15feaf197d0" exitCode=0 Dec 01 16:18:58 crc kubenswrapper[4810]: I1201 16:18:58.737203 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-4gnz2" event={"ID":"8e768084-da29-4e95-ae6e-feacb0cbfb70","Type":"ContainerDied","Data":"df728e70e0bfa7b885f9383efd0f8ca5b80e51de177005bb4b88c15feaf197d0"} Dec 01 16:18:59 crc kubenswrapper[4810]: I1201 16:18:59.748934 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-4gnz2" event={"ID":"8e768084-da29-4e95-ae6e-feacb0cbfb70","Type":"ContainerStarted","Data":"1438928a059def3b5dc0ec0af783a4b757128a83e4233e90a6427af61fc10d91"} Dec 01 16:18:59 crc kubenswrapper[4810]: I1201 16:18:59.749777 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-4gnz2" Dec 01 16:18:59 crc kubenswrapper[4810]: I1201 16:18:59.768185 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-4gnz2" podStartSLOduration=5.706497731 podStartE2EDuration="7.768166219s" podCreationTimestamp="2025-12-01 16:18:52 +0000 UTC" firstStartedPulling="2025-12-01 16:18:54.442891884 +0000 UTC m=+6300.206401487" lastFinishedPulling="2025-12-01 16:18:56.504560372 +0000 UTC m=+6302.268069975" observedRunningTime="2025-12-01 16:18:59.764385156 +0000 UTC m=+6305.527894769" watchObservedRunningTime="2025-12-01 16:18:59.768166219 +0000 UTC m=+6305.531675822" Dec 01 16:19:01 crc kubenswrapper[4810]: I1201 16:19:01.491633 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:19:01 crc kubenswrapper[4810]: E1201 16:19:01.492266 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:19:04 crc kubenswrapper[4810]: I1201 16:19:04.393421 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-frfm6" Dec 01 16:19:06 crc kubenswrapper[4810]: I1201 16:19:06.546027 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-qthjz" Dec 01 16:19:08 crc kubenswrapper[4810]: I1201 16:19:08.412086 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-4gnz2" Dec 01 16:19:13 crc kubenswrapper[4810]: I1201 16:19:13.492264 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:19:13 crc kubenswrapper[4810]: I1201 16:19:13.880090 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"13b8d9a5912bf37c10017c37926041fb03894967f1288822928bc2736df5d8df"} Dec 01 16:19:46 crc kubenswrapper[4810]: I1201 16:19:46.038731 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-t57h9"] Dec 01 16:19:46 crc kubenswrapper[4810]: I1201 16:19:46.048190 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-fc84-account-create-update-wgmzd"] Dec 01 16:19:46 crc kubenswrapper[4810]: I1201 16:19:46.058934 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-fc84-account-create-update-wgmzd"] Dec 01 16:19:46 crc kubenswrapper[4810]: I1201 16:19:46.067703 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-t57h9"] Dec 01 16:19:46 crc kubenswrapper[4810]: I1201 16:19:46.502941 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25572bab-6803-44f8-b3d9-98491d59fe7b" path="/var/lib/kubelet/pods/25572bab-6803-44f8-b3d9-98491d59fe7b/volumes" Dec 01 16:19:46 crc kubenswrapper[4810]: I1201 16:19:46.504014 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea8c6b74-11cf-4f7c-be88-5715592abf4b" path="/var/lib/kubelet/pods/ea8c6b74-11cf-4f7c-be88-5715592abf4b/volumes" Dec 01 16:19:53 crc kubenswrapper[4810]: I1201 16:19:53.032076 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-g8lkr"] Dec 01 16:19:53 crc kubenswrapper[4810]: I1201 16:19:53.044458 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-g8lkr"] Dec 01 16:19:54 crc kubenswrapper[4810]: I1201 16:19:54.503122 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dc4b16f-f399-4883-bb29-0ec81865287f" path="/var/lib/kubelet/pods/1dc4b16f-f399-4883-bb29-0ec81865287f/volumes" Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.798208 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6977f7d7f7-vld6p"] Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.801214 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.806927 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6977f7d7f7-vld6p"] Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.807425 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.807445 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.807665 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.807908 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-wb4pb" Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.864798 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.865216 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="76104e9f-958f-4ca1-bc3a-1a15e635ff48" containerName="glance-log" containerID="cri-o://c39c26527d998183b88e7e60f894db477ff11d24b80a4b7263e61d7d2bcec83f" gracePeriod=30 Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.869045 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="76104e9f-958f-4ca1-bc3a-1a15e635ff48" containerName="glance-httpd" containerID="cri-o://9df7f870b2d4de005275809ca768bdb334b257b2b522ea3cda6a960246b34076" gracePeriod=30 Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.890715 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-config-data\") pod \"horizon-6977f7d7f7-vld6p\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.891043 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-scripts\") pod \"horizon-6977f7d7f7-vld6p\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.891117 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9bmn\" (UniqueName: \"kubernetes.io/projected/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-kube-api-access-l9bmn\") pod \"horizon-6977f7d7f7-vld6p\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.891203 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-logs\") pod \"horizon-6977f7d7f7-vld6p\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.891338 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-horizon-secret-key\") pod \"horizon-6977f7d7f7-vld6p\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.918496 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.918769 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d401e073-d123-406b-8282-7e25300eea65" containerName="glance-log" containerID="cri-o://1d7fedbac8c47a4792657340b81ad6aa5cbff1c15d61f7b1aa7699324bc89fce" gracePeriod=30 Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.919281 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d401e073-d123-406b-8282-7e25300eea65" containerName="glance-httpd" containerID="cri-o://7cb96c27e96c8a68f6eeeb30ff69f968200aead5bc35e93c92675395d16ee569" gracePeriod=30 Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.994750 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-horizon-secret-key\") pod \"horizon-6977f7d7f7-vld6p\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.994848 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-config-data\") pod \"horizon-6977f7d7f7-vld6p\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.994882 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-scripts\") pod \"horizon-6977f7d7f7-vld6p\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.994920 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9bmn\" (UniqueName: \"kubernetes.io/projected/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-kube-api-access-l9bmn\") pod \"horizon-6977f7d7f7-vld6p\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.994968 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-logs\") pod \"horizon-6977f7d7f7-vld6p\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.995357 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-logs\") pod \"horizon-6977f7d7f7-vld6p\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.996322 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5f86c44d77-j7mjd"] Dec 01 16:19:59 crc kubenswrapper[4810]: I1201 16:19:59.998392 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.002236 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-config-data\") pod \"horizon-6977f7d7f7-vld6p\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.004285 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-scripts\") pod \"horizon-6977f7d7f7-vld6p\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.013429 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-horizon-secret-key\") pod \"horizon-6977f7d7f7-vld6p\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.026301 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f86c44d77-j7mjd"] Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.031204 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9bmn\" (UniqueName: \"kubernetes.io/projected/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-kube-api-access-l9bmn\") pod \"horizon-6977f7d7f7-vld6p\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.097720 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c482dc55-76a0-4494-8b9c-67e2c1110a54-config-data\") pod \"horizon-5f86c44d77-j7mjd\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.097830 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggbz4\" (UniqueName: \"kubernetes.io/projected/c482dc55-76a0-4494-8b9c-67e2c1110a54-kube-api-access-ggbz4\") pod \"horizon-5f86c44d77-j7mjd\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.097869 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c482dc55-76a0-4494-8b9c-67e2c1110a54-horizon-secret-key\") pod \"horizon-5f86c44d77-j7mjd\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.097910 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c482dc55-76a0-4494-8b9c-67e2c1110a54-logs\") pod \"horizon-5f86c44d77-j7mjd\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.098105 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c482dc55-76a0-4494-8b9c-67e2c1110a54-scripts\") pod \"horizon-5f86c44d77-j7mjd\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.146042 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.200612 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c482dc55-76a0-4494-8b9c-67e2c1110a54-config-data\") pod \"horizon-5f86c44d77-j7mjd\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.201061 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggbz4\" (UniqueName: \"kubernetes.io/projected/c482dc55-76a0-4494-8b9c-67e2c1110a54-kube-api-access-ggbz4\") pod \"horizon-5f86c44d77-j7mjd\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.201104 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c482dc55-76a0-4494-8b9c-67e2c1110a54-horizon-secret-key\") pod \"horizon-5f86c44d77-j7mjd\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.201244 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c482dc55-76a0-4494-8b9c-67e2c1110a54-logs\") pod \"horizon-5f86c44d77-j7mjd\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.201274 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c482dc55-76a0-4494-8b9c-67e2c1110a54-scripts\") pod \"horizon-5f86c44d77-j7mjd\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.201698 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c482dc55-76a0-4494-8b9c-67e2c1110a54-logs\") pod \"horizon-5f86c44d77-j7mjd\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.201970 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c482dc55-76a0-4494-8b9c-67e2c1110a54-scripts\") pod \"horizon-5f86c44d77-j7mjd\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.202090 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c482dc55-76a0-4494-8b9c-67e2c1110a54-config-data\") pod \"horizon-5f86c44d77-j7mjd\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.205303 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c482dc55-76a0-4494-8b9c-67e2c1110a54-horizon-secret-key\") pod \"horizon-5f86c44d77-j7mjd\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.220657 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggbz4\" (UniqueName: \"kubernetes.io/projected/c482dc55-76a0-4494-8b9c-67e2c1110a54-kube-api-access-ggbz4\") pod \"horizon-5f86c44d77-j7mjd\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.346811 4810 generic.go:334] "Generic (PLEG): container finished" podID="76104e9f-958f-4ca1-bc3a-1a15e635ff48" containerID="c39c26527d998183b88e7e60f894db477ff11d24b80a4b7263e61d7d2bcec83f" exitCode=143 Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.346903 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"76104e9f-958f-4ca1-bc3a-1a15e635ff48","Type":"ContainerDied","Data":"c39c26527d998183b88e7e60f894db477ff11d24b80a4b7263e61d7d2bcec83f"} Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.349030 4810 generic.go:334] "Generic (PLEG): container finished" podID="d401e073-d123-406b-8282-7e25300eea65" containerID="1d7fedbac8c47a4792657340b81ad6aa5cbff1c15d61f7b1aa7699324bc89fce" exitCode=143 Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.349060 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d401e073-d123-406b-8282-7e25300eea65","Type":"ContainerDied","Data":"1d7fedbac8c47a4792657340b81ad6aa5cbff1c15d61f7b1aa7699324bc89fce"} Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.429637 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.666204 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6977f7d7f7-vld6p"] Dec 01 16:20:00 crc kubenswrapper[4810]: I1201 16:20:00.913785 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f86c44d77-j7mjd"] Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.357687 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f86c44d77-j7mjd" event={"ID":"c482dc55-76a0-4494-8b9c-67e2c1110a54","Type":"ContainerStarted","Data":"b0e7acb56d1233b3ee148df3ae11602c0f840c5542151f7e2a36493c676c1869"} Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.359462 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6977f7d7f7-vld6p" event={"ID":"9c3d5365-14f8-4de4-8304-9b6a58e32c4a","Type":"ContainerStarted","Data":"7fd301f7c2d05358923f485a0eef4e22bc09730d6d528283430e44b87c9997dc"} Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.564790 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f86c44d77-j7mjd"] Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.596443 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-f4c7cd84d-sm8dj"] Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.605417 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.609446 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.623985 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f4c7cd84d-sm8dj"] Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.631130 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4923848-5f10-4163-96b7-ac8076519136-logs\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.631258 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4923848-5f10-4163-96b7-ac8076519136-horizon-tls-certs\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.631324 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8qlr\" (UniqueName: \"kubernetes.io/projected/a4923848-5f10-4163-96b7-ac8076519136-kube-api-access-d8qlr\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.631364 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4923848-5f10-4163-96b7-ac8076519136-combined-ca-bundle\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.631628 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4923848-5f10-4163-96b7-ac8076519136-config-data\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.631755 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a4923848-5f10-4163-96b7-ac8076519136-horizon-secret-key\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.631882 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4923848-5f10-4163-96b7-ac8076519136-scripts\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.687389 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6977f7d7f7-vld6p"] Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.713900 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-57bd9c5d44-hps7g"] Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.717962 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.733263 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8qlr\" (UniqueName: \"kubernetes.io/projected/a4923848-5f10-4163-96b7-ac8076519136-kube-api-access-d8qlr\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.733326 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4923848-5f10-4163-96b7-ac8076519136-combined-ca-bundle\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.733375 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4923848-5f10-4163-96b7-ac8076519136-config-data\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.733415 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a4923848-5f10-4163-96b7-ac8076519136-horizon-secret-key\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.733452 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4923848-5f10-4163-96b7-ac8076519136-scripts\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.733501 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4923848-5f10-4163-96b7-ac8076519136-logs\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.733563 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4923848-5f10-4163-96b7-ac8076519136-horizon-tls-certs\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.734776 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4923848-5f10-4163-96b7-ac8076519136-scripts\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.735117 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4923848-5f10-4163-96b7-ac8076519136-logs\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.735252 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4923848-5f10-4163-96b7-ac8076519136-config-data\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.741775 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-57bd9c5d44-hps7g"] Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.743717 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a4923848-5f10-4163-96b7-ac8076519136-horizon-secret-key\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.745151 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4923848-5f10-4163-96b7-ac8076519136-combined-ca-bundle\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.759858 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4923848-5f10-4163-96b7-ac8076519136-horizon-tls-certs\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.767176 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8qlr\" (UniqueName: \"kubernetes.io/projected/a4923848-5f10-4163-96b7-ac8076519136-kube-api-access-d8qlr\") pod \"horizon-f4c7cd84d-sm8dj\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.836017 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfad2664-5a86-48b8-b3a5-cad4eec57408-combined-ca-bundle\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.836079 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfad2664-5a86-48b8-b3a5-cad4eec57408-horizon-tls-certs\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.836130 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bfad2664-5a86-48b8-b3a5-cad4eec57408-scripts\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.836218 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bfad2664-5a86-48b8-b3a5-cad4eec57408-config-data\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.836367 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjdpk\" (UniqueName: \"kubernetes.io/projected/bfad2664-5a86-48b8-b3a5-cad4eec57408-kube-api-access-pjdpk\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.836408 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bfad2664-5a86-48b8-b3a5-cad4eec57408-horizon-secret-key\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.836441 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfad2664-5a86-48b8-b3a5-cad4eec57408-logs\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.930168 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.938077 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfad2664-5a86-48b8-b3a5-cad4eec57408-combined-ca-bundle\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.938286 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfad2664-5a86-48b8-b3a5-cad4eec57408-horizon-tls-certs\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.938373 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bfad2664-5a86-48b8-b3a5-cad4eec57408-scripts\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.938539 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bfad2664-5a86-48b8-b3a5-cad4eec57408-config-data\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.938701 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjdpk\" (UniqueName: \"kubernetes.io/projected/bfad2664-5a86-48b8-b3a5-cad4eec57408-kube-api-access-pjdpk\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.938780 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bfad2664-5a86-48b8-b3a5-cad4eec57408-horizon-secret-key\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.938855 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfad2664-5a86-48b8-b3a5-cad4eec57408-logs\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.939276 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfad2664-5a86-48b8-b3a5-cad4eec57408-logs\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.939852 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bfad2664-5a86-48b8-b3a5-cad4eec57408-scripts\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.943674 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bfad2664-5a86-48b8-b3a5-cad4eec57408-horizon-secret-key\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.943951 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfad2664-5a86-48b8-b3a5-cad4eec57408-combined-ca-bundle\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.944087 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfad2664-5a86-48b8-b3a5-cad4eec57408-horizon-tls-certs\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.947215 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bfad2664-5a86-48b8-b3a5-cad4eec57408-config-data\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:01 crc kubenswrapper[4810]: I1201 16:20:01.962355 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjdpk\" (UniqueName: \"kubernetes.io/projected/bfad2664-5a86-48b8-b3a5-cad4eec57408-kube-api-access-pjdpk\") pod \"horizon-57bd9c5d44-hps7g\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:02 crc kubenswrapper[4810]: I1201 16:20:02.154345 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:02 crc kubenswrapper[4810]: I1201 16:20:02.412729 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f4c7cd84d-sm8dj"] Dec 01 16:20:02 crc kubenswrapper[4810]: I1201 16:20:02.703900 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-57bd9c5d44-hps7g"] Dec 01 16:20:02 crc kubenswrapper[4810]: W1201 16:20:02.707339 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfad2664_5a86_48b8_b3a5_cad4eec57408.slice/crio-c76190228c0e362d5d4618bef66941165d9b56ef8ad9459f6d3ff456e2e51536 WatchSource:0}: Error finding container c76190228c0e362d5d4618bef66941165d9b56ef8ad9459f6d3ff456e2e51536: Status 404 returned error can't find the container with id c76190228c0e362d5d4618bef66941165d9b56ef8ad9459f6d3ff456e2e51536 Dec 01 16:20:03 crc kubenswrapper[4810]: I1201 16:20:03.315959 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="76104e9f-958f-4ca1-bc3a-1a15e635ff48" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.1.61:9292/healthcheck\": read tcp 10.217.0.2:34646->10.217.1.61:9292: read: connection reset by peer" Dec 01 16:20:03 crc kubenswrapper[4810]: I1201 16:20:03.316284 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="76104e9f-958f-4ca1-bc3a-1a15e635ff48" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.1.61:9292/healthcheck\": read tcp 10.217.0.2:34662->10.217.1.61:9292: read: connection reset by peer" Dec 01 16:20:03 crc kubenswrapper[4810]: I1201 16:20:03.335348 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="d401e073-d123-406b-8282-7e25300eea65" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.1.60:9292/healthcheck\": read tcp 10.217.0.2:56208->10.217.1.60:9292: read: connection reset by peer" Dec 01 16:20:03 crc kubenswrapper[4810]: I1201 16:20:03.335402 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="d401e073-d123-406b-8282-7e25300eea65" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.1.60:9292/healthcheck\": read tcp 10.217.0.2:56210->10.217.1.60:9292: read: connection reset by peer" Dec 01 16:20:03 crc kubenswrapper[4810]: I1201 16:20:03.390061 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57bd9c5d44-hps7g" event={"ID":"bfad2664-5a86-48b8-b3a5-cad4eec57408","Type":"ContainerStarted","Data":"c76190228c0e362d5d4618bef66941165d9b56ef8ad9459f6d3ff456e2e51536"} Dec 01 16:20:03 crc kubenswrapper[4810]: I1201 16:20:03.392313 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f4c7cd84d-sm8dj" event={"ID":"a4923848-5f10-4163-96b7-ac8076519136","Type":"ContainerStarted","Data":"c646ff07ba41c678b7ffc0cd111078e7a3ad4a4720652779a35ef48edcbb821a"} Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.041062 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.090338 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-combined-ca-bundle\") pod \"d401e073-d123-406b-8282-7e25300eea65\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.090765 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-internal-tls-certs\") pod \"d401e073-d123-406b-8282-7e25300eea65\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.090973 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5dmg\" (UniqueName: \"kubernetes.io/projected/d401e073-d123-406b-8282-7e25300eea65-kube-api-access-v5dmg\") pod \"d401e073-d123-406b-8282-7e25300eea65\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.091142 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d401e073-d123-406b-8282-7e25300eea65-httpd-run\") pod \"d401e073-d123-406b-8282-7e25300eea65\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.091292 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-config-data\") pod \"d401e073-d123-406b-8282-7e25300eea65\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.091414 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d401e073-d123-406b-8282-7e25300eea65-logs\") pod \"d401e073-d123-406b-8282-7e25300eea65\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.091568 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-scripts\") pod \"d401e073-d123-406b-8282-7e25300eea65\" (UID: \"d401e073-d123-406b-8282-7e25300eea65\") " Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.093159 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d401e073-d123-406b-8282-7e25300eea65-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d401e073-d123-406b-8282-7e25300eea65" (UID: "d401e073-d123-406b-8282-7e25300eea65"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.093238 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d401e073-d123-406b-8282-7e25300eea65-logs" (OuterVolumeSpecName: "logs") pod "d401e073-d123-406b-8282-7e25300eea65" (UID: "d401e073-d123-406b-8282-7e25300eea65"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.098665 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-scripts" (OuterVolumeSpecName: "scripts") pod "d401e073-d123-406b-8282-7e25300eea65" (UID: "d401e073-d123-406b-8282-7e25300eea65"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.099044 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d401e073-d123-406b-8282-7e25300eea65-kube-api-access-v5dmg" (OuterVolumeSpecName: "kube-api-access-v5dmg") pod "d401e073-d123-406b-8282-7e25300eea65" (UID: "d401e073-d123-406b-8282-7e25300eea65"). InnerVolumeSpecName "kube-api-access-v5dmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.123230 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d401e073-d123-406b-8282-7e25300eea65" (UID: "d401e073-d123-406b-8282-7e25300eea65"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.150673 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d401e073-d123-406b-8282-7e25300eea65" (UID: "d401e073-d123-406b-8282-7e25300eea65"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.194148 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d401e073-d123-406b-8282-7e25300eea65-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.194503 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d401e073-d123-406b-8282-7e25300eea65-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.194586 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.194743 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.194814 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.194938 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5dmg\" (UniqueName: \"kubernetes.io/projected/d401e073-d123-406b-8282-7e25300eea65-kube-api-access-v5dmg\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.200853 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-config-data" (OuterVolumeSpecName: "config-data") pod "d401e073-d123-406b-8282-7e25300eea65" (UID: "d401e073-d123-406b-8282-7e25300eea65"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.296991 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d401e073-d123-406b-8282-7e25300eea65-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.411734 4810 generic.go:334] "Generic (PLEG): container finished" podID="76104e9f-958f-4ca1-bc3a-1a15e635ff48" containerID="9df7f870b2d4de005275809ca768bdb334b257b2b522ea3cda6a960246b34076" exitCode=0 Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.411930 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"76104e9f-958f-4ca1-bc3a-1a15e635ff48","Type":"ContainerDied","Data":"9df7f870b2d4de005275809ca768bdb334b257b2b522ea3cda6a960246b34076"} Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.415072 4810 generic.go:334] "Generic (PLEG): container finished" podID="d401e073-d123-406b-8282-7e25300eea65" containerID="7cb96c27e96c8a68f6eeeb30ff69f968200aead5bc35e93c92675395d16ee569" exitCode=0 Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.415113 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d401e073-d123-406b-8282-7e25300eea65","Type":"ContainerDied","Data":"7cb96c27e96c8a68f6eeeb30ff69f968200aead5bc35e93c92675395d16ee569"} Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.415141 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d401e073-d123-406b-8282-7e25300eea65","Type":"ContainerDied","Data":"109b8610ffd8f7d330500c7934091e3a004d95ac6cf6961dfa9742a57adec879"} Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.415164 4810 scope.go:117] "RemoveContainer" containerID="7cb96c27e96c8a68f6eeeb30ff69f968200aead5bc35e93c92675395d16ee569" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.415312 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.456957 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.488399 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.542113 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d401e073-d123-406b-8282-7e25300eea65" path="/var/lib/kubelet/pods/d401e073-d123-406b-8282-7e25300eea65/volumes" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.543005 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:20:04 crc kubenswrapper[4810]: E1201 16:20:04.544359 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d401e073-d123-406b-8282-7e25300eea65" containerName="glance-log" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.544391 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d401e073-d123-406b-8282-7e25300eea65" containerName="glance-log" Dec 01 16:20:04 crc kubenswrapper[4810]: E1201 16:20:04.544411 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d401e073-d123-406b-8282-7e25300eea65" containerName="glance-httpd" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.544437 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d401e073-d123-406b-8282-7e25300eea65" containerName="glance-httpd" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.544817 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d401e073-d123-406b-8282-7e25300eea65" containerName="glance-log" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.544838 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d401e073-d123-406b-8282-7e25300eea65" containerName="glance-httpd" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.546172 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.546279 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.548705 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.549175 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.633431 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.633497 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.633607 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.633636 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjgb5\" (UniqueName: \"kubernetes.io/projected/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-kube-api-access-xjgb5\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.633663 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-config-data\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.633690 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-scripts\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.633807 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-logs\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.735877 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.735932 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.736018 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.736211 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjgb5\" (UniqueName: \"kubernetes.io/projected/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-kube-api-access-xjgb5\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.736266 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-config-data\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.736310 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-scripts\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.736350 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-logs\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.736823 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.736887 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-logs\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.744705 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.745379 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.747838 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-scripts\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.762915 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-config-data\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.765845 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjgb5\" (UniqueName: \"kubernetes.io/projected/11ef5ba9-de25-4010-8a26-a8f38b5e7e08-kube-api-access-xjgb5\") pod \"glance-default-internal-api-0\" (UID: \"11ef5ba9-de25-4010-8a26-a8f38b5e7e08\") " pod="openstack/glance-default-internal-api-0" Dec 01 16:20:04 crc kubenswrapper[4810]: I1201 16:20:04.883451 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 16:20:09 crc kubenswrapper[4810]: I1201 16:20:09.673387 4810 scope.go:117] "RemoveContainer" containerID="1d7fedbac8c47a4792657340b81ad6aa5cbff1c15d61f7b1aa7699324bc89fce" Dec 01 16:20:09 crc kubenswrapper[4810]: I1201 16:20:09.856982 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 16:20:09 crc kubenswrapper[4810]: I1201 16:20:09.864015 4810 scope.go:117] "RemoveContainer" containerID="7cb96c27e96c8a68f6eeeb30ff69f968200aead5bc35e93c92675395d16ee569" Dec 01 16:20:09 crc kubenswrapper[4810]: E1201 16:20:09.864600 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cb96c27e96c8a68f6eeeb30ff69f968200aead5bc35e93c92675395d16ee569\": container with ID starting with 7cb96c27e96c8a68f6eeeb30ff69f968200aead5bc35e93c92675395d16ee569 not found: ID does not exist" containerID="7cb96c27e96c8a68f6eeeb30ff69f968200aead5bc35e93c92675395d16ee569" Dec 01 16:20:09 crc kubenswrapper[4810]: I1201 16:20:09.864641 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cb96c27e96c8a68f6eeeb30ff69f968200aead5bc35e93c92675395d16ee569"} err="failed to get container status \"7cb96c27e96c8a68f6eeeb30ff69f968200aead5bc35e93c92675395d16ee569\": rpc error: code = NotFound desc = could not find container \"7cb96c27e96c8a68f6eeeb30ff69f968200aead5bc35e93c92675395d16ee569\": container with ID starting with 7cb96c27e96c8a68f6eeeb30ff69f968200aead5bc35e93c92675395d16ee569 not found: ID does not exist" Dec 01 16:20:09 crc kubenswrapper[4810]: I1201 16:20:09.864664 4810 scope.go:117] "RemoveContainer" containerID="1d7fedbac8c47a4792657340b81ad6aa5cbff1c15d61f7b1aa7699324bc89fce" Dec 01 16:20:09 crc kubenswrapper[4810]: E1201 16:20:09.864952 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d7fedbac8c47a4792657340b81ad6aa5cbff1c15d61f7b1aa7699324bc89fce\": container with ID starting with 1d7fedbac8c47a4792657340b81ad6aa5cbff1c15d61f7b1aa7699324bc89fce not found: ID does not exist" containerID="1d7fedbac8c47a4792657340b81ad6aa5cbff1c15d61f7b1aa7699324bc89fce" Dec 01 16:20:09 crc kubenswrapper[4810]: I1201 16:20:09.864975 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d7fedbac8c47a4792657340b81ad6aa5cbff1c15d61f7b1aa7699324bc89fce"} err="failed to get container status \"1d7fedbac8c47a4792657340b81ad6aa5cbff1c15d61f7b1aa7699324bc89fce\": rpc error: code = NotFound desc = could not find container \"1d7fedbac8c47a4792657340b81ad6aa5cbff1c15d61f7b1aa7699324bc89fce\": container with ID starting with 1d7fedbac8c47a4792657340b81ad6aa5cbff1c15d61f7b1aa7699324bc89fce not found: ID does not exist" Dec 01 16:20:09 crc kubenswrapper[4810]: I1201 16:20:09.955786 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76104e9f-958f-4ca1-bc3a-1a15e635ff48-logs\") pod \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " Dec 01 16:20:09 crc kubenswrapper[4810]: I1201 16:20:09.956413 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-scripts\") pod \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " Dec 01 16:20:09 crc kubenswrapper[4810]: I1201 16:20:09.956511 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24f89\" (UniqueName: \"kubernetes.io/projected/76104e9f-958f-4ca1-bc3a-1a15e635ff48-kube-api-access-24f89\") pod \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " Dec 01 16:20:09 crc kubenswrapper[4810]: I1201 16:20:09.956560 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-combined-ca-bundle\") pod \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " Dec 01 16:20:09 crc kubenswrapper[4810]: I1201 16:20:09.956590 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-public-tls-certs\") pod \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " Dec 01 16:20:09 crc kubenswrapper[4810]: I1201 16:20:09.956588 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76104e9f-958f-4ca1-bc3a-1a15e635ff48-logs" (OuterVolumeSpecName: "logs") pod "76104e9f-958f-4ca1-bc3a-1a15e635ff48" (UID: "76104e9f-958f-4ca1-bc3a-1a15e635ff48"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:20:09 crc kubenswrapper[4810]: I1201 16:20:09.956611 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-config-data\") pod \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " Dec 01 16:20:09 crc kubenswrapper[4810]: I1201 16:20:09.956723 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/76104e9f-958f-4ca1-bc3a-1a15e635ff48-httpd-run\") pod \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\" (UID: \"76104e9f-958f-4ca1-bc3a-1a15e635ff48\") " Dec 01 16:20:09 crc kubenswrapper[4810]: I1201 16:20:09.957315 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76104e9f-958f-4ca1-bc3a-1a15e635ff48-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:09 crc kubenswrapper[4810]: I1201 16:20:09.957442 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76104e9f-958f-4ca1-bc3a-1a15e635ff48-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "76104e9f-958f-4ca1-bc3a-1a15e635ff48" (UID: "76104e9f-958f-4ca1-bc3a-1a15e635ff48"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:20:09 crc kubenswrapper[4810]: I1201 16:20:09.967918 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-scripts" (OuterVolumeSpecName: "scripts") pod "76104e9f-958f-4ca1-bc3a-1a15e635ff48" (UID: "76104e9f-958f-4ca1-bc3a-1a15e635ff48"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:20:09 crc kubenswrapper[4810]: I1201 16:20:09.979649 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76104e9f-958f-4ca1-bc3a-1a15e635ff48-kube-api-access-24f89" (OuterVolumeSpecName: "kube-api-access-24f89") pod "76104e9f-958f-4ca1-bc3a-1a15e635ff48" (UID: "76104e9f-958f-4ca1-bc3a-1a15e635ff48"). InnerVolumeSpecName "kube-api-access-24f89". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.002988 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76104e9f-958f-4ca1-bc3a-1a15e635ff48" (UID: "76104e9f-958f-4ca1-bc3a-1a15e635ff48"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.040029 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-config-data" (OuterVolumeSpecName: "config-data") pod "76104e9f-958f-4ca1-bc3a-1a15e635ff48" (UID: "76104e9f-958f-4ca1-bc3a-1a15e635ff48"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.047275 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "76104e9f-958f-4ca1-bc3a-1a15e635ff48" (UID: "76104e9f-958f-4ca1-bc3a-1a15e635ff48"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.059195 4810 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/76104e9f-958f-4ca1-bc3a-1a15e635ff48-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.059231 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.059245 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24f89\" (UniqueName: \"kubernetes.io/projected/76104e9f-958f-4ca1-bc3a-1a15e635ff48-kube-api-access-24f89\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.059258 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.059270 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.059280 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/76104e9f-958f-4ca1-bc3a-1a15e635ff48-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.313656 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.486621 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f4c7cd84d-sm8dj" event={"ID":"a4923848-5f10-4163-96b7-ac8076519136","Type":"ContainerStarted","Data":"827d5e87019fa7efe0ccacc01296c4671c440e38e50c4aebf6448ac2a41f655f"} Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.488810 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57bd9c5d44-hps7g" event={"ID":"bfad2664-5a86-48b8-b3a5-cad4eec57408","Type":"ContainerStarted","Data":"ff427cf2a241a76a12a13f31fbfd44c3fab7d63ef78a625e7e10a45cd6d0ddf2"} Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.503707 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.510032 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6977f7d7f7-vld6p" event={"ID":"9c3d5365-14f8-4de4-8304-9b6a58e32c4a","Type":"ContainerStarted","Data":"8c5c5c19725165b163a753d4c04b57245928999a3f753cd94cef96c2d05e3c75"} Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.510075 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"76104e9f-958f-4ca1-bc3a-1a15e635ff48","Type":"ContainerDied","Data":"c9746fb3c485f308dc5ce3e783d7f8d5df06eab6e8e078866a0379dca331d8dd"} Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.510101 4810 scope.go:117] "RemoveContainer" containerID="9df7f870b2d4de005275809ca768bdb334b257b2b522ea3cda6a960246b34076" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.513048 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"11ef5ba9-de25-4010-8a26-a8f38b5e7e08","Type":"ContainerStarted","Data":"942df8e8760ada9fb8e84f3875eb32cb2b4f04722fc28940c628a29d3280e250"} Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.516140 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f86c44d77-j7mjd" event={"ID":"c482dc55-76a0-4494-8b9c-67e2c1110a54","Type":"ContainerStarted","Data":"b3ea48b2adcfb96f6d300b23400979e6a92f05040073b1c1b8b131c55e46a961"} Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.562156 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.575559 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.591288 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:20:10 crc kubenswrapper[4810]: E1201 16:20:10.591803 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76104e9f-958f-4ca1-bc3a-1a15e635ff48" containerName="glance-httpd" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.591820 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="76104e9f-958f-4ca1-bc3a-1a15e635ff48" containerName="glance-httpd" Dec 01 16:20:10 crc kubenswrapper[4810]: E1201 16:20:10.591843 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76104e9f-958f-4ca1-bc3a-1a15e635ff48" containerName="glance-log" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.591849 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="76104e9f-958f-4ca1-bc3a-1a15e635ff48" containerName="glance-log" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.592069 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="76104e9f-958f-4ca1-bc3a-1a15e635ff48" containerName="glance-httpd" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.592087 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="76104e9f-958f-4ca1-bc3a-1a15e635ff48" containerName="glance-log" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.600451 4810 scope.go:117] "RemoveContainer" containerID="c39c26527d998183b88e7e60f894db477ff11d24b80a4b7263e61d7d2bcec83f" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.604124 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.604284 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.608013 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.608636 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.675262 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7rwb\" (UniqueName: \"kubernetes.io/projected/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-kube-api-access-g7rwb\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.675356 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.675539 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.675687 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.675809 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-scripts\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.675921 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-config-data\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.675960 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-logs\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.777744 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-logs\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.777824 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7rwb\" (UniqueName: \"kubernetes.io/projected/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-kube-api-access-g7rwb\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.777873 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.777914 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.777956 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.778002 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-scripts\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.778044 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-config-data\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.783111 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.783232 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-config-data\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.783577 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-logs\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.787067 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.788205 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-scripts\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.791502 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.803043 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7rwb\" (UniqueName: \"kubernetes.io/projected/1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe-kube-api-access-g7rwb\") pod \"glance-default-external-api-0\" (UID: \"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe\") " pod="openstack/glance-default-external-api-0" Dec 01 16:20:10 crc kubenswrapper[4810]: I1201 16:20:10.944014 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.200057 4810 scope.go:117] "RemoveContainer" containerID="84f0047271368cf2223c8fbcd2ff1ab071bee50b8138c1d762f09707ef546dc5" Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.246191 4810 scope.go:117] "RemoveContainer" containerID="51ce5819550849f51a2704ce44356a9f9400058434bc9d595bc1d317b876e573" Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.290380 4810 scope.go:117] "RemoveContainer" containerID="7173470a69c19c6fcbcf89a00f13376b1f317809ba6047ff859d127b2a57ad6f" Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.516793 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.535300 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f4c7cd84d-sm8dj" event={"ID":"a4923848-5f10-4163-96b7-ac8076519136","Type":"ContainerStarted","Data":"cf9f0e7882457e67859f245983f2d92db53016f190cce6fe25e0aa6c4fd6c8b1"} Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.543139 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6977f7d7f7-vld6p" event={"ID":"9c3d5365-14f8-4de4-8304-9b6a58e32c4a","Type":"ContainerStarted","Data":"0371452afa751ee77eefa06ce3fdc4ced719891c077ea4423903b40a6d4d7876"} Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.543273 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6977f7d7f7-vld6p" podUID="9c3d5365-14f8-4de4-8304-9b6a58e32c4a" containerName="horizon-log" containerID="cri-o://8c5c5c19725165b163a753d4c04b57245928999a3f753cd94cef96c2d05e3c75" gracePeriod=30 Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.543353 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6977f7d7f7-vld6p" podUID="9c3d5365-14f8-4de4-8304-9b6a58e32c4a" containerName="horizon" containerID="cri-o://0371452afa751ee77eefa06ce3fdc4ced719891c077ea4423903b40a6d4d7876" gracePeriod=30 Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.548857 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f86c44d77-j7mjd" event={"ID":"c482dc55-76a0-4494-8b9c-67e2c1110a54","Type":"ContainerStarted","Data":"727cbb6be2c8356cf2e8ccf64247786e9fae3a1ab2a92df73f6e2b9a1c279795"} Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.549029 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5f86c44d77-j7mjd" podUID="c482dc55-76a0-4494-8b9c-67e2c1110a54" containerName="horizon-log" containerID="cri-o://b3ea48b2adcfb96f6d300b23400979e6a92f05040073b1c1b8b131c55e46a961" gracePeriod=30 Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.549118 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5f86c44d77-j7mjd" podUID="c482dc55-76a0-4494-8b9c-67e2c1110a54" containerName="horizon" containerID="cri-o://727cbb6be2c8356cf2e8ccf64247786e9fae3a1ab2a92df73f6e2b9a1c279795" gracePeriod=30 Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.565880 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-f4c7cd84d-sm8dj" podStartSLOduration=3.135716417 podStartE2EDuration="10.565856653s" podCreationTimestamp="2025-12-01 16:20:01 +0000 UTC" firstStartedPulling="2025-12-01 16:20:02.433833766 +0000 UTC m=+6368.197343369" lastFinishedPulling="2025-12-01 16:20:09.863974002 +0000 UTC m=+6375.627483605" observedRunningTime="2025-12-01 16:20:11.556268472 +0000 UTC m=+6377.319778075" watchObservedRunningTime="2025-12-01 16:20:11.565856653 +0000 UTC m=+6377.329366256" Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.567660 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57bd9c5d44-hps7g" event={"ID":"bfad2664-5a86-48b8-b3a5-cad4eec57408","Type":"ContainerStarted","Data":"d7079fd9f393e10798a10ad299a51d5120a549c4ac48bec74590bab9fbce49f0"} Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.571417 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"11ef5ba9-de25-4010-8a26-a8f38b5e7e08","Type":"ContainerStarted","Data":"9cea66fe70907912bad530b94a753de073ce93ede54837f4c0505b47a4136a1f"} Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.586334 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6977f7d7f7-vld6p" podStartSLOduration=3.471488408 podStartE2EDuration="12.586315878s" podCreationTimestamp="2025-12-01 16:19:59 +0000 UTC" firstStartedPulling="2025-12-01 16:20:00.673623422 +0000 UTC m=+6366.437133025" lastFinishedPulling="2025-12-01 16:20:09.788450892 +0000 UTC m=+6375.551960495" observedRunningTime="2025-12-01 16:20:11.578217188 +0000 UTC m=+6377.341726791" watchObservedRunningTime="2025-12-01 16:20:11.586315878 +0000 UTC m=+6377.349825481" Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.601687 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5f86c44d77-j7mjd" podStartSLOduration=3.653331895 podStartE2EDuration="12.601666745s" podCreationTimestamp="2025-12-01 16:19:59 +0000 UTC" firstStartedPulling="2025-12-01 16:20:00.916347431 +0000 UTC m=+6366.679857044" lastFinishedPulling="2025-12-01 16:20:09.864682291 +0000 UTC m=+6375.628191894" observedRunningTime="2025-12-01 16:20:11.598648273 +0000 UTC m=+6377.362157896" watchObservedRunningTime="2025-12-01 16:20:11.601666745 +0000 UTC m=+6377.365176348" Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.629491 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-57bd9c5d44-hps7g" podStartSLOduration=3.546697705 podStartE2EDuration="10.62945282s" podCreationTimestamp="2025-12-01 16:20:01 +0000 UTC" firstStartedPulling="2025-12-01 16:20:02.709211292 +0000 UTC m=+6368.472720895" lastFinishedPulling="2025-12-01 16:20:09.791966407 +0000 UTC m=+6375.555476010" observedRunningTime="2025-12-01 16:20:11.619640543 +0000 UTC m=+6377.383150146" watchObservedRunningTime="2025-12-01 16:20:11.62945282 +0000 UTC m=+6377.392962423" Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.930272 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:11 crc kubenswrapper[4810]: I1201 16:20:11.930335 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:12 crc kubenswrapper[4810]: I1201 16:20:12.155481 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:12 crc kubenswrapper[4810]: I1201 16:20:12.155774 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:12 crc kubenswrapper[4810]: I1201 16:20:12.502395 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76104e9f-958f-4ca1-bc3a-1a15e635ff48" path="/var/lib/kubelet/pods/76104e9f-958f-4ca1-bc3a-1a15e635ff48/volumes" Dec 01 16:20:12 crc kubenswrapper[4810]: I1201 16:20:12.584274 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"11ef5ba9-de25-4010-8a26-a8f38b5e7e08","Type":"ContainerStarted","Data":"927349fff293c5c54292cd0eba81ed9101758e4bc90842e6a7be972550b8c86e"} Dec 01 16:20:12 crc kubenswrapper[4810]: I1201 16:20:12.587016 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe","Type":"ContainerStarted","Data":"81a25d4b1449d1864f246981aa8a2c3e15e3460288fd4adf76ad3b62e80fe60a"} Dec 01 16:20:12 crc kubenswrapper[4810]: I1201 16:20:12.587086 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe","Type":"ContainerStarted","Data":"2be2d0e8e7efa9e301af2340f64802ec3bd957a51acee0ef5f11af0509f5ab74"} Dec 01 16:20:12 crc kubenswrapper[4810]: I1201 16:20:12.620269 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.620240016 podStartE2EDuration="8.620240016s" podCreationTimestamp="2025-12-01 16:20:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:20:12.60270519 +0000 UTC m=+6378.366214803" watchObservedRunningTime="2025-12-01 16:20:12.620240016 +0000 UTC m=+6378.383749639" Dec 01 16:20:12 crc kubenswrapper[4810]: I1201 16:20:12.644066 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.644038432 podStartE2EDuration="2.644038432s" podCreationTimestamp="2025-12-01 16:20:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:20:12.638654966 +0000 UTC m=+6378.402164559" watchObservedRunningTime="2025-12-01 16:20:12.644038432 +0000 UTC m=+6378.407548035" Dec 01 16:20:13 crc kubenswrapper[4810]: I1201 16:20:13.600497 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe","Type":"ContainerStarted","Data":"ae3668fbafbfc31cdaf69b45c6bc71afc61827ac04ac220a3a4492a451aded59"} Dec 01 16:20:14 crc kubenswrapper[4810]: I1201 16:20:14.885096 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 16:20:14 crc kubenswrapper[4810]: I1201 16:20:14.885423 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 16:20:14 crc kubenswrapper[4810]: I1201 16:20:14.919994 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 16:20:14 crc kubenswrapper[4810]: I1201 16:20:14.930632 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 16:20:15 crc kubenswrapper[4810]: I1201 16:20:15.617831 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 16:20:15 crc kubenswrapper[4810]: I1201 16:20:15.617890 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 16:20:18 crc kubenswrapper[4810]: I1201 16:20:18.934127 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 16:20:19 crc kubenswrapper[4810]: I1201 16:20:19.052495 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 16:20:20 crc kubenswrapper[4810]: I1201 16:20:20.049499 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-s6472"] Dec 01 16:20:20 crc kubenswrapper[4810]: I1201 16:20:20.062223 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-2611-account-create-update-b2jpl"] Dec 01 16:20:20 crc kubenswrapper[4810]: I1201 16:20:20.070451 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-s6472"] Dec 01 16:20:20 crc kubenswrapper[4810]: I1201 16:20:20.078371 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-2611-account-create-update-b2jpl"] Dec 01 16:20:20 crc kubenswrapper[4810]: I1201 16:20:20.147079 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:20:20 crc kubenswrapper[4810]: I1201 16:20:20.430585 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:20 crc kubenswrapper[4810]: I1201 16:20:20.503457 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87906600-d845-4859-9875-4137a77092f9" path="/var/lib/kubelet/pods/87906600-d845-4859-9875-4137a77092f9/volumes" Dec 01 16:20:20 crc kubenswrapper[4810]: I1201 16:20:20.504439 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbd26113-dc91-4686-8f81-93eea1d5eedc" path="/var/lib/kubelet/pods/cbd26113-dc91-4686-8f81-93eea1d5eedc/volumes" Dec 01 16:20:20 crc kubenswrapper[4810]: I1201 16:20:20.944287 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 16:20:20 crc kubenswrapper[4810]: I1201 16:20:20.944337 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 16:20:20 crc kubenswrapper[4810]: I1201 16:20:20.987991 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 16:20:20 crc kubenswrapper[4810]: I1201 16:20:20.990947 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 16:20:21 crc kubenswrapper[4810]: I1201 16:20:21.671322 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 16:20:21 crc kubenswrapper[4810]: I1201 16:20:21.671361 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 16:20:21 crc kubenswrapper[4810]: I1201 16:20:21.933149 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-f4c7cd84d-sm8dj" podUID="a4923848-5f10-4163-96b7-ac8076519136" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.117:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.117:8443: connect: connection refused" Dec 01 16:20:22 crc kubenswrapper[4810]: I1201 16:20:22.156873 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-57bd9c5d44-hps7g" podUID="bfad2664-5a86-48b8-b3a5-cad4eec57408" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.118:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.118:8443: connect: connection refused" Dec 01 16:20:23 crc kubenswrapper[4810]: I1201 16:20:23.639428 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 16:20:23 crc kubenswrapper[4810]: I1201 16:20:23.639519 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 16:20:28 crc kubenswrapper[4810]: I1201 16:20:28.047487 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-djlnt"] Dec 01 16:20:28 crc kubenswrapper[4810]: I1201 16:20:28.061189 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-djlnt"] Dec 01 16:20:28 crc kubenswrapper[4810]: I1201 16:20:28.502054 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caaddf08-f0f1-40d1-ba3a-88c187a1f08b" path="/var/lib/kubelet/pods/caaddf08-f0f1-40d1-ba3a-88c187a1f08b/volumes" Dec 01 16:20:33 crc kubenswrapper[4810]: I1201 16:20:33.721197 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:33 crc kubenswrapper[4810]: I1201 16:20:33.949825 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:35 crc kubenswrapper[4810]: I1201 16:20:35.621303 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:20:35 crc kubenswrapper[4810]: I1201 16:20:35.636213 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:20:35 crc kubenswrapper[4810]: I1201 16:20:35.733329 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-f4c7cd84d-sm8dj"] Dec 01 16:20:35 crc kubenswrapper[4810]: I1201 16:20:35.883515 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-f4c7cd84d-sm8dj" podUID="a4923848-5f10-4163-96b7-ac8076519136" containerName="horizon-log" containerID="cri-o://827d5e87019fa7efe0ccacc01296c4671c440e38e50c4aebf6448ac2a41f655f" gracePeriod=30 Dec 01 16:20:35 crc kubenswrapper[4810]: I1201 16:20:35.883607 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-f4c7cd84d-sm8dj" podUID="a4923848-5f10-4163-96b7-ac8076519136" containerName="horizon" containerID="cri-o://cf9f0e7882457e67859f245983f2d92db53016f190cce6fe25e0aa6c4fd6c8b1" gracePeriod=30 Dec 01 16:20:39 crc kubenswrapper[4810]: I1201 16:20:39.920712 4810 generic.go:334] "Generic (PLEG): container finished" podID="a4923848-5f10-4163-96b7-ac8076519136" containerID="cf9f0e7882457e67859f245983f2d92db53016f190cce6fe25e0aa6c4fd6c8b1" exitCode=0 Dec 01 16:20:39 crc kubenswrapper[4810]: I1201 16:20:39.920826 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f4c7cd84d-sm8dj" event={"ID":"a4923848-5f10-4163-96b7-ac8076519136","Type":"ContainerDied","Data":"cf9f0e7882457e67859f245983f2d92db53016f190cce6fe25e0aa6c4fd6c8b1"} Dec 01 16:20:41 crc kubenswrapper[4810]: I1201 16:20:41.931057 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-f4c7cd84d-sm8dj" podUID="a4923848-5f10-4163-96b7-ac8076519136" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.117:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.117:8443: connect: connection refused" Dec 01 16:20:41 crc kubenswrapper[4810]: I1201 16:20:41.951030 4810 generic.go:334] "Generic (PLEG): container finished" podID="9c3d5365-14f8-4de4-8304-9b6a58e32c4a" containerID="0371452afa751ee77eefa06ce3fdc4ced719891c077ea4423903b40a6d4d7876" exitCode=137 Dec 01 16:20:41 crc kubenswrapper[4810]: I1201 16:20:41.951062 4810 generic.go:334] "Generic (PLEG): container finished" podID="9c3d5365-14f8-4de4-8304-9b6a58e32c4a" containerID="8c5c5c19725165b163a753d4c04b57245928999a3f753cd94cef96c2d05e3c75" exitCode=137 Dec 01 16:20:41 crc kubenswrapper[4810]: I1201 16:20:41.951110 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6977f7d7f7-vld6p" event={"ID":"9c3d5365-14f8-4de4-8304-9b6a58e32c4a","Type":"ContainerDied","Data":"0371452afa751ee77eefa06ce3fdc4ced719891c077ea4423903b40a6d4d7876"} Dec 01 16:20:41 crc kubenswrapper[4810]: I1201 16:20:41.951159 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6977f7d7f7-vld6p" event={"ID":"9c3d5365-14f8-4de4-8304-9b6a58e32c4a","Type":"ContainerDied","Data":"8c5c5c19725165b163a753d4c04b57245928999a3f753cd94cef96c2d05e3c75"} Dec 01 16:20:41 crc kubenswrapper[4810]: I1201 16:20:41.953533 4810 generic.go:334] "Generic (PLEG): container finished" podID="c482dc55-76a0-4494-8b9c-67e2c1110a54" containerID="727cbb6be2c8356cf2e8ccf64247786e9fae3a1ab2a92df73f6e2b9a1c279795" exitCode=137 Dec 01 16:20:41 crc kubenswrapper[4810]: I1201 16:20:41.953686 4810 generic.go:334] "Generic (PLEG): container finished" podID="c482dc55-76a0-4494-8b9c-67e2c1110a54" containerID="b3ea48b2adcfb96f6d300b23400979e6a92f05040073b1c1b8b131c55e46a961" exitCode=137 Dec 01 16:20:41 crc kubenswrapper[4810]: I1201 16:20:41.953579 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f86c44d77-j7mjd" event={"ID":"c482dc55-76a0-4494-8b9c-67e2c1110a54","Type":"ContainerDied","Data":"727cbb6be2c8356cf2e8ccf64247786e9fae3a1ab2a92df73f6e2b9a1c279795"} Dec 01 16:20:41 crc kubenswrapper[4810]: I1201 16:20:41.953921 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f86c44d77-j7mjd" event={"ID":"c482dc55-76a0-4494-8b9c-67e2c1110a54","Type":"ContainerDied","Data":"b3ea48b2adcfb96f6d300b23400979e6a92f05040073b1c1b8b131c55e46a961"} Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.093538 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.100620 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.218999 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c482dc55-76a0-4494-8b9c-67e2c1110a54-horizon-secret-key\") pod \"c482dc55-76a0-4494-8b9c-67e2c1110a54\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.219377 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c482dc55-76a0-4494-8b9c-67e2c1110a54-config-data\") pod \"c482dc55-76a0-4494-8b9c-67e2c1110a54\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.219438 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-scripts\") pod \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.219460 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-logs\") pod \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.219522 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9bmn\" (UniqueName: \"kubernetes.io/projected/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-kube-api-access-l9bmn\") pod \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.219567 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c482dc55-76a0-4494-8b9c-67e2c1110a54-scripts\") pod \"c482dc55-76a0-4494-8b9c-67e2c1110a54\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.219611 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c482dc55-76a0-4494-8b9c-67e2c1110a54-logs\") pod \"c482dc55-76a0-4494-8b9c-67e2c1110a54\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.219661 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggbz4\" (UniqueName: \"kubernetes.io/projected/c482dc55-76a0-4494-8b9c-67e2c1110a54-kube-api-access-ggbz4\") pod \"c482dc55-76a0-4494-8b9c-67e2c1110a54\" (UID: \"c482dc55-76a0-4494-8b9c-67e2c1110a54\") " Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.219693 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-config-data\") pod \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.219838 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-horizon-secret-key\") pod \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\" (UID: \"9c3d5365-14f8-4de4-8304-9b6a58e32c4a\") " Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.221893 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c482dc55-76a0-4494-8b9c-67e2c1110a54-logs" (OuterVolumeSpecName: "logs") pod "c482dc55-76a0-4494-8b9c-67e2c1110a54" (UID: "c482dc55-76a0-4494-8b9c-67e2c1110a54"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.222122 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-logs" (OuterVolumeSpecName: "logs") pod "9c3d5365-14f8-4de4-8304-9b6a58e32c4a" (UID: "9c3d5365-14f8-4de4-8304-9b6a58e32c4a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.226047 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-kube-api-access-l9bmn" (OuterVolumeSpecName: "kube-api-access-l9bmn") pod "9c3d5365-14f8-4de4-8304-9b6a58e32c4a" (UID: "9c3d5365-14f8-4de4-8304-9b6a58e32c4a"). InnerVolumeSpecName "kube-api-access-l9bmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.226592 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "9c3d5365-14f8-4de4-8304-9b6a58e32c4a" (UID: "9c3d5365-14f8-4de4-8304-9b6a58e32c4a"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.227510 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c482dc55-76a0-4494-8b9c-67e2c1110a54-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c482dc55-76a0-4494-8b9c-67e2c1110a54" (UID: "c482dc55-76a0-4494-8b9c-67e2c1110a54"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.228335 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c482dc55-76a0-4494-8b9c-67e2c1110a54-kube-api-access-ggbz4" (OuterVolumeSpecName: "kube-api-access-ggbz4") pod "c482dc55-76a0-4494-8b9c-67e2c1110a54" (UID: "c482dc55-76a0-4494-8b9c-67e2c1110a54"). InnerVolumeSpecName "kube-api-access-ggbz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.247280 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c482dc55-76a0-4494-8b9c-67e2c1110a54-config-data" (OuterVolumeSpecName: "config-data") pod "c482dc55-76a0-4494-8b9c-67e2c1110a54" (UID: "c482dc55-76a0-4494-8b9c-67e2c1110a54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.248368 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c482dc55-76a0-4494-8b9c-67e2c1110a54-scripts" (OuterVolumeSpecName: "scripts") pod "c482dc55-76a0-4494-8b9c-67e2c1110a54" (UID: "c482dc55-76a0-4494-8b9c-67e2c1110a54"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.248576 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-config-data" (OuterVolumeSpecName: "config-data") pod "9c3d5365-14f8-4de4-8304-9b6a58e32c4a" (UID: "9c3d5365-14f8-4de4-8304-9b6a58e32c4a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.249215 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-scripts" (OuterVolumeSpecName: "scripts") pod "9c3d5365-14f8-4de4-8304-9b6a58e32c4a" (UID: "9c3d5365-14f8-4de4-8304-9b6a58e32c4a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.321895 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c482dc55-76a0-4494-8b9c-67e2c1110a54-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.322119 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggbz4\" (UniqueName: \"kubernetes.io/projected/c482dc55-76a0-4494-8b9c-67e2c1110a54-kube-api-access-ggbz4\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.322191 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.322248 4810 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.322312 4810 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c482dc55-76a0-4494-8b9c-67e2c1110a54-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.322403 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c482dc55-76a0-4494-8b9c-67e2c1110a54-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.322489 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.322557 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.322618 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9bmn\" (UniqueName: \"kubernetes.io/projected/9c3d5365-14f8-4de4-8304-9b6a58e32c4a-kube-api-access-l9bmn\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.322675 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c482dc55-76a0-4494-8b9c-67e2c1110a54-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.966737 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6977f7d7f7-vld6p" event={"ID":"9c3d5365-14f8-4de4-8304-9b6a58e32c4a","Type":"ContainerDied","Data":"7fd301f7c2d05358923f485a0eef4e22bc09730d6d528283430e44b87c9997dc"} Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.966749 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6977f7d7f7-vld6p" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.968003 4810 scope.go:117] "RemoveContainer" containerID="0371452afa751ee77eefa06ce3fdc4ced719891c077ea4423903b40a6d4d7876" Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.969208 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f86c44d77-j7mjd" event={"ID":"c482dc55-76a0-4494-8b9c-67e2c1110a54","Type":"ContainerDied","Data":"b0e7acb56d1233b3ee148df3ae11602c0f840c5542151f7e2a36493c676c1869"} Dec 01 16:20:42 crc kubenswrapper[4810]: I1201 16:20:42.969250 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f86c44d77-j7mjd" Dec 01 16:20:43 crc kubenswrapper[4810]: I1201 16:20:43.014530 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f86c44d77-j7mjd"] Dec 01 16:20:43 crc kubenswrapper[4810]: I1201 16:20:43.026915 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5f86c44d77-j7mjd"] Dec 01 16:20:43 crc kubenswrapper[4810]: I1201 16:20:43.035614 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6977f7d7f7-vld6p"] Dec 01 16:20:43 crc kubenswrapper[4810]: I1201 16:20:43.044435 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6977f7d7f7-vld6p"] Dec 01 16:20:43 crc kubenswrapper[4810]: I1201 16:20:43.145426 4810 scope.go:117] "RemoveContainer" containerID="8c5c5c19725165b163a753d4c04b57245928999a3f753cd94cef96c2d05e3c75" Dec 01 16:20:43 crc kubenswrapper[4810]: I1201 16:20:43.167408 4810 scope.go:117] "RemoveContainer" containerID="727cbb6be2c8356cf2e8ccf64247786e9fae3a1ab2a92df73f6e2b9a1c279795" Dec 01 16:20:43 crc kubenswrapper[4810]: I1201 16:20:43.334336 4810 scope.go:117] "RemoveContainer" containerID="b3ea48b2adcfb96f6d300b23400979e6a92f05040073b1c1b8b131c55e46a961" Dec 01 16:20:44 crc kubenswrapper[4810]: I1201 16:20:44.506329 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c3d5365-14f8-4de4-8304-9b6a58e32c4a" path="/var/lib/kubelet/pods/9c3d5365-14f8-4de4-8304-9b6a58e32c4a/volumes" Dec 01 16:20:44 crc kubenswrapper[4810]: I1201 16:20:44.508325 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c482dc55-76a0-4494-8b9c-67e2c1110a54" path="/var/lib/kubelet/pods/c482dc55-76a0-4494-8b9c-67e2c1110a54/volumes" Dec 01 16:20:51 crc kubenswrapper[4810]: I1201 16:20:51.931630 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-f4c7cd84d-sm8dj" podUID="a4923848-5f10-4163-96b7-ac8076519136" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.117:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.117:8443: connect: connection refused" Dec 01 16:21:01 crc kubenswrapper[4810]: I1201 16:21:01.931679 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-f4c7cd84d-sm8dj" podUID="a4923848-5f10-4163-96b7-ac8076519136" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.117:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.117:8443: connect: connection refused" Dec 01 16:21:01 crc kubenswrapper[4810]: I1201 16:21:01.932247 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.225946 4810 generic.go:334] "Generic (PLEG): container finished" podID="a4923848-5f10-4163-96b7-ac8076519136" containerID="827d5e87019fa7efe0ccacc01296c4671c440e38e50c4aebf6448ac2a41f655f" exitCode=137 Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.226586 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f4c7cd84d-sm8dj" event={"ID":"a4923848-5f10-4163-96b7-ac8076519136","Type":"ContainerDied","Data":"827d5e87019fa7efe0ccacc01296c4671c440e38e50c4aebf6448ac2a41f655f"} Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.395924 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.518257 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8qlr\" (UniqueName: \"kubernetes.io/projected/a4923848-5f10-4163-96b7-ac8076519136-kube-api-access-d8qlr\") pod \"a4923848-5f10-4163-96b7-ac8076519136\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.518339 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a4923848-5f10-4163-96b7-ac8076519136-horizon-secret-key\") pod \"a4923848-5f10-4163-96b7-ac8076519136\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.518395 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4923848-5f10-4163-96b7-ac8076519136-config-data\") pod \"a4923848-5f10-4163-96b7-ac8076519136\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.518499 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4923848-5f10-4163-96b7-ac8076519136-scripts\") pod \"a4923848-5f10-4163-96b7-ac8076519136\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.518558 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4923848-5f10-4163-96b7-ac8076519136-logs\") pod \"a4923848-5f10-4163-96b7-ac8076519136\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.518704 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4923848-5f10-4163-96b7-ac8076519136-combined-ca-bundle\") pod \"a4923848-5f10-4163-96b7-ac8076519136\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.518744 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4923848-5f10-4163-96b7-ac8076519136-horizon-tls-certs\") pod \"a4923848-5f10-4163-96b7-ac8076519136\" (UID: \"a4923848-5f10-4163-96b7-ac8076519136\") " Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.519314 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4923848-5f10-4163-96b7-ac8076519136-logs" (OuterVolumeSpecName: "logs") pod "a4923848-5f10-4163-96b7-ac8076519136" (UID: "a4923848-5f10-4163-96b7-ac8076519136"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.519724 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4923848-5f10-4163-96b7-ac8076519136-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.526096 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4923848-5f10-4163-96b7-ac8076519136-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a4923848-5f10-4163-96b7-ac8076519136" (UID: "a4923848-5f10-4163-96b7-ac8076519136"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.527034 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4923848-5f10-4163-96b7-ac8076519136-kube-api-access-d8qlr" (OuterVolumeSpecName: "kube-api-access-d8qlr") pod "a4923848-5f10-4163-96b7-ac8076519136" (UID: "a4923848-5f10-4163-96b7-ac8076519136"). InnerVolumeSpecName "kube-api-access-d8qlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.552337 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4923848-5f10-4163-96b7-ac8076519136-config-data" (OuterVolumeSpecName: "config-data") pod "a4923848-5f10-4163-96b7-ac8076519136" (UID: "a4923848-5f10-4163-96b7-ac8076519136"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.562527 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4923848-5f10-4163-96b7-ac8076519136-scripts" (OuterVolumeSpecName: "scripts") pod "a4923848-5f10-4163-96b7-ac8076519136" (UID: "a4923848-5f10-4163-96b7-ac8076519136"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.568794 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4923848-5f10-4163-96b7-ac8076519136-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4923848-5f10-4163-96b7-ac8076519136" (UID: "a4923848-5f10-4163-96b7-ac8076519136"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.583908 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4923848-5f10-4163-96b7-ac8076519136-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "a4923848-5f10-4163-96b7-ac8076519136" (UID: "a4923848-5f10-4163-96b7-ac8076519136"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.621707 4810 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4923848-5f10-4163-96b7-ac8076519136-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.621756 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8qlr\" (UniqueName: \"kubernetes.io/projected/a4923848-5f10-4163-96b7-ac8076519136-kube-api-access-d8qlr\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.621770 4810 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a4923848-5f10-4163-96b7-ac8076519136-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.621781 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4923848-5f10-4163-96b7-ac8076519136-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.621793 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4923848-5f10-4163-96b7-ac8076519136-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:06 crc kubenswrapper[4810]: I1201 16:21:06.621804 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4923848-5f10-4163-96b7-ac8076519136-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:07 crc kubenswrapper[4810]: I1201 16:21:07.241961 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f4c7cd84d-sm8dj" event={"ID":"a4923848-5f10-4163-96b7-ac8076519136","Type":"ContainerDied","Data":"c646ff07ba41c678b7ffc0cd111078e7a3ad4a4720652779a35ef48edcbb821a"} Dec 01 16:21:07 crc kubenswrapper[4810]: I1201 16:21:07.242045 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f4c7cd84d-sm8dj" Dec 01 16:21:07 crc kubenswrapper[4810]: I1201 16:21:07.242456 4810 scope.go:117] "RemoveContainer" containerID="cf9f0e7882457e67859f245983f2d92db53016f190cce6fe25e0aa6c4fd6c8b1" Dec 01 16:21:07 crc kubenswrapper[4810]: I1201 16:21:07.287052 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-f4c7cd84d-sm8dj"] Dec 01 16:21:07 crc kubenswrapper[4810]: I1201 16:21:07.295190 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-f4c7cd84d-sm8dj"] Dec 01 16:21:07 crc kubenswrapper[4810]: I1201 16:21:07.454778 4810 scope.go:117] "RemoveContainer" containerID="827d5e87019fa7efe0ccacc01296c4671c440e38e50c4aebf6448ac2a41f655f" Dec 01 16:21:08 crc kubenswrapper[4810]: I1201 16:21:08.502160 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4923848-5f10-4163-96b7-ac8076519136" path="/var/lib/kubelet/pods/a4923848-5f10-4163-96b7-ac8076519136/volumes" Dec 01 16:21:11 crc kubenswrapper[4810]: I1201 16:21:11.467261 4810 scope.go:117] "RemoveContainer" containerID="488324fcd51e26039702d3aa0feaa773895d50aef3a27513f8a21b002eb7d114" Dec 01 16:21:11 crc kubenswrapper[4810]: I1201 16:21:11.512324 4810 scope.go:117] "RemoveContainer" containerID="bfbbd76e2e527ce521228e1438ead5d83f6f8dced1636912a78a58ded9f2b6d6" Dec 01 16:21:11 crc kubenswrapper[4810]: I1201 16:21:11.571826 4810 scope.go:117] "RemoveContainer" containerID="e7c3e1ede2fceeee5ec3674a8a1085fdf71cdb92211e67d47e9bcfeffbd1c695" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.438200 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-574946f46-cwn7p"] Dec 01 16:21:17 crc kubenswrapper[4810]: E1201 16:21:17.438994 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c3d5365-14f8-4de4-8304-9b6a58e32c4a" containerName="horizon-log" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.439011 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c3d5365-14f8-4de4-8304-9b6a58e32c4a" containerName="horizon-log" Dec 01 16:21:17 crc kubenswrapper[4810]: E1201 16:21:17.439035 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4923848-5f10-4163-96b7-ac8076519136" containerName="horizon-log" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.439042 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4923848-5f10-4163-96b7-ac8076519136" containerName="horizon-log" Dec 01 16:21:17 crc kubenswrapper[4810]: E1201 16:21:17.439056 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c482dc55-76a0-4494-8b9c-67e2c1110a54" containerName="horizon" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.439063 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c482dc55-76a0-4494-8b9c-67e2c1110a54" containerName="horizon" Dec 01 16:21:17 crc kubenswrapper[4810]: E1201 16:21:17.439077 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c482dc55-76a0-4494-8b9c-67e2c1110a54" containerName="horizon-log" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.439082 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c482dc55-76a0-4494-8b9c-67e2c1110a54" containerName="horizon-log" Dec 01 16:21:17 crc kubenswrapper[4810]: E1201 16:21:17.439092 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c3d5365-14f8-4de4-8304-9b6a58e32c4a" containerName="horizon" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.439098 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c3d5365-14f8-4de4-8304-9b6a58e32c4a" containerName="horizon" Dec 01 16:21:17 crc kubenswrapper[4810]: E1201 16:21:17.439110 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4923848-5f10-4163-96b7-ac8076519136" containerName="horizon" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.439118 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4923848-5f10-4163-96b7-ac8076519136" containerName="horizon" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.439316 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4923848-5f10-4163-96b7-ac8076519136" containerName="horizon" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.439329 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4923848-5f10-4163-96b7-ac8076519136" containerName="horizon-log" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.439338 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c482dc55-76a0-4494-8b9c-67e2c1110a54" containerName="horizon" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.439347 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c3d5365-14f8-4de4-8304-9b6a58e32c4a" containerName="horizon" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.439359 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c3d5365-14f8-4de4-8304-9b6a58e32c4a" containerName="horizon-log" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.439374 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c482dc55-76a0-4494-8b9c-67e2c1110a54" containerName="horizon-log" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.440434 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.478291 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-574946f46-cwn7p"] Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.553534 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73a62ae2-3e32-424b-b181-da93d1c88f35-logs\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.553604 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73a62ae2-3e32-424b-b181-da93d1c88f35-horizon-secret-key\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.553638 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwfq8\" (UniqueName: \"kubernetes.io/projected/73a62ae2-3e32-424b-b181-da93d1c88f35-kube-api-access-mwfq8\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.553653 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/73a62ae2-3e32-424b-b181-da93d1c88f35-horizon-tls-certs\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.553963 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73a62ae2-3e32-424b-b181-da93d1c88f35-config-data\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.554040 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73a62ae2-3e32-424b-b181-da93d1c88f35-combined-ca-bundle\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.554207 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73a62ae2-3e32-424b-b181-da93d1c88f35-scripts\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.655645 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73a62ae2-3e32-424b-b181-da93d1c88f35-scripts\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.655794 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73a62ae2-3e32-424b-b181-da93d1c88f35-logs\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.655847 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73a62ae2-3e32-424b-b181-da93d1c88f35-horizon-secret-key\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.655892 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwfq8\" (UniqueName: \"kubernetes.io/projected/73a62ae2-3e32-424b-b181-da93d1c88f35-kube-api-access-mwfq8\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.655916 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/73a62ae2-3e32-424b-b181-da93d1c88f35-horizon-tls-certs\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.655971 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73a62ae2-3e32-424b-b181-da93d1c88f35-config-data\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.655994 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73a62ae2-3e32-424b-b181-da93d1c88f35-combined-ca-bundle\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.656561 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73a62ae2-3e32-424b-b181-da93d1c88f35-scripts\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.656867 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73a62ae2-3e32-424b-b181-da93d1c88f35-logs\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.658971 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73a62ae2-3e32-424b-b181-da93d1c88f35-config-data\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.672343 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73a62ae2-3e32-424b-b181-da93d1c88f35-horizon-secret-key\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.672420 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/73a62ae2-3e32-424b-b181-da93d1c88f35-horizon-tls-certs\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.672629 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73a62ae2-3e32-424b-b181-da93d1c88f35-combined-ca-bundle\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.677322 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwfq8\" (UniqueName: \"kubernetes.io/projected/73a62ae2-3e32-424b-b181-da93d1c88f35-kube-api-access-mwfq8\") pod \"horizon-574946f46-cwn7p\" (UID: \"73a62ae2-3e32-424b-b181-da93d1c88f35\") " pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:17 crc kubenswrapper[4810]: I1201 16:21:17.760191 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:18 crc kubenswrapper[4810]: I1201 16:21:18.226696 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-574946f46-cwn7p"] Dec 01 16:21:18 crc kubenswrapper[4810]: I1201 16:21:18.381239 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574946f46-cwn7p" event={"ID":"73a62ae2-3e32-424b-b181-da93d1c88f35","Type":"ContainerStarted","Data":"cf02898427e8902d2559f7214c3a6d5b5e896d5cce08b933568955f243d9792b"} Dec 01 16:21:18 crc kubenswrapper[4810]: I1201 16:21:18.381289 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574946f46-cwn7p" event={"ID":"73a62ae2-3e32-424b-b181-da93d1c88f35","Type":"ContainerStarted","Data":"38c93fa4afe1b4258b11ac71befdd9ea4439c62652374cee809269473bc0a008"} Dec 01 16:21:18 crc kubenswrapper[4810]: I1201 16:21:18.847347 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-4f6vj"] Dec 01 16:21:18 crc kubenswrapper[4810]: I1201 16:21:18.849135 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-4f6vj" Dec 01 16:21:18 crc kubenswrapper[4810]: I1201 16:21:18.858240 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-4f6vj"] Dec 01 16:21:18 crc kubenswrapper[4810]: I1201 16:21:18.945805 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-4dbf-account-create-update-l4q8c"] Dec 01 16:21:18 crc kubenswrapper[4810]: I1201 16:21:18.948590 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-4dbf-account-create-update-l4q8c" Dec 01 16:21:18 crc kubenswrapper[4810]: I1201 16:21:18.952890 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 01 16:21:18 crc kubenswrapper[4810]: I1201 16:21:18.964774 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-4dbf-account-create-update-l4q8c"] Dec 01 16:21:18 crc kubenswrapper[4810]: I1201 16:21:18.992840 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f36f4f1-7565-45e8-b96d-5b52449a4bb4-operator-scripts\") pod \"heat-db-create-4f6vj\" (UID: \"6f36f4f1-7565-45e8-b96d-5b52449a4bb4\") " pod="openstack/heat-db-create-4f6vj" Dec 01 16:21:18 crc kubenswrapper[4810]: I1201 16:21:18.993109 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrm7b\" (UniqueName: \"kubernetes.io/projected/6f36f4f1-7565-45e8-b96d-5b52449a4bb4-kube-api-access-mrm7b\") pod \"heat-db-create-4f6vj\" (UID: \"6f36f4f1-7565-45e8-b96d-5b52449a4bb4\") " pod="openstack/heat-db-create-4f6vj" Dec 01 16:21:19 crc kubenswrapper[4810]: I1201 16:21:19.094449 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4ded436-383b-493b-a03b-c65f14063d69-operator-scripts\") pod \"heat-4dbf-account-create-update-l4q8c\" (UID: \"d4ded436-383b-493b-a03b-c65f14063d69\") " pod="openstack/heat-4dbf-account-create-update-l4q8c" Dec 01 16:21:19 crc kubenswrapper[4810]: I1201 16:21:19.095275 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4wd5\" (UniqueName: \"kubernetes.io/projected/d4ded436-383b-493b-a03b-c65f14063d69-kube-api-access-r4wd5\") pod \"heat-4dbf-account-create-update-l4q8c\" (UID: \"d4ded436-383b-493b-a03b-c65f14063d69\") " pod="openstack/heat-4dbf-account-create-update-l4q8c" Dec 01 16:21:19 crc kubenswrapper[4810]: I1201 16:21:19.095565 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrm7b\" (UniqueName: \"kubernetes.io/projected/6f36f4f1-7565-45e8-b96d-5b52449a4bb4-kube-api-access-mrm7b\") pod \"heat-db-create-4f6vj\" (UID: \"6f36f4f1-7565-45e8-b96d-5b52449a4bb4\") " pod="openstack/heat-db-create-4f6vj" Dec 01 16:21:19 crc kubenswrapper[4810]: I1201 16:21:19.095623 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f36f4f1-7565-45e8-b96d-5b52449a4bb4-operator-scripts\") pod \"heat-db-create-4f6vj\" (UID: \"6f36f4f1-7565-45e8-b96d-5b52449a4bb4\") " pod="openstack/heat-db-create-4f6vj" Dec 01 16:21:19 crc kubenswrapper[4810]: I1201 16:21:19.096326 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f36f4f1-7565-45e8-b96d-5b52449a4bb4-operator-scripts\") pod \"heat-db-create-4f6vj\" (UID: \"6f36f4f1-7565-45e8-b96d-5b52449a4bb4\") " pod="openstack/heat-db-create-4f6vj" Dec 01 16:21:19 crc kubenswrapper[4810]: I1201 16:21:19.112065 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrm7b\" (UniqueName: \"kubernetes.io/projected/6f36f4f1-7565-45e8-b96d-5b52449a4bb4-kube-api-access-mrm7b\") pod \"heat-db-create-4f6vj\" (UID: \"6f36f4f1-7565-45e8-b96d-5b52449a4bb4\") " pod="openstack/heat-db-create-4f6vj" Dec 01 16:21:19 crc kubenswrapper[4810]: I1201 16:21:19.197737 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4ded436-383b-493b-a03b-c65f14063d69-operator-scripts\") pod \"heat-4dbf-account-create-update-l4q8c\" (UID: \"d4ded436-383b-493b-a03b-c65f14063d69\") " pod="openstack/heat-4dbf-account-create-update-l4q8c" Dec 01 16:21:19 crc kubenswrapper[4810]: I1201 16:21:19.197785 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4wd5\" (UniqueName: \"kubernetes.io/projected/d4ded436-383b-493b-a03b-c65f14063d69-kube-api-access-r4wd5\") pod \"heat-4dbf-account-create-update-l4q8c\" (UID: \"d4ded436-383b-493b-a03b-c65f14063d69\") " pod="openstack/heat-4dbf-account-create-update-l4q8c" Dec 01 16:21:19 crc kubenswrapper[4810]: I1201 16:21:19.198676 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4ded436-383b-493b-a03b-c65f14063d69-operator-scripts\") pod \"heat-4dbf-account-create-update-l4q8c\" (UID: \"d4ded436-383b-493b-a03b-c65f14063d69\") " pod="openstack/heat-4dbf-account-create-update-l4q8c" Dec 01 16:21:19 crc kubenswrapper[4810]: I1201 16:21:19.207678 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-4f6vj" Dec 01 16:21:19 crc kubenswrapper[4810]: I1201 16:21:19.218645 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4wd5\" (UniqueName: \"kubernetes.io/projected/d4ded436-383b-493b-a03b-c65f14063d69-kube-api-access-r4wd5\") pod \"heat-4dbf-account-create-update-l4q8c\" (UID: \"d4ded436-383b-493b-a03b-c65f14063d69\") " pod="openstack/heat-4dbf-account-create-update-l4q8c" Dec 01 16:21:19 crc kubenswrapper[4810]: I1201 16:21:19.265701 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-4dbf-account-create-update-l4q8c" Dec 01 16:21:19 crc kubenswrapper[4810]: I1201 16:21:19.420947 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-574946f46-cwn7p" event={"ID":"73a62ae2-3e32-424b-b181-da93d1c88f35","Type":"ContainerStarted","Data":"fc25af516f084e28e5a07e84bd555eb73bb9f2034bb92283d3bb43cb9ff15dba"} Dec 01 16:21:19 crc kubenswrapper[4810]: I1201 16:21:19.459324 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-574946f46-cwn7p" podStartSLOduration=2.4593065689999998 podStartE2EDuration="2.459306569s" podCreationTimestamp="2025-12-01 16:21:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:21:19.44500648 +0000 UTC m=+6445.208516103" watchObservedRunningTime="2025-12-01 16:21:19.459306569 +0000 UTC m=+6445.222816172" Dec 01 16:21:19 crc kubenswrapper[4810]: I1201 16:21:19.775647 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-4f6vj"] Dec 01 16:21:19 crc kubenswrapper[4810]: W1201 16:21:19.783669 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f36f4f1_7565_45e8_b96d_5b52449a4bb4.slice/crio-7f26405ec608ec171036f52b4bb68b7822400dc44512376c99e1abfd76895b7c WatchSource:0}: Error finding container 7f26405ec608ec171036f52b4bb68b7822400dc44512376c99e1abfd76895b7c: Status 404 returned error can't find the container with id 7f26405ec608ec171036f52b4bb68b7822400dc44512376c99e1abfd76895b7c Dec 01 16:21:19 crc kubenswrapper[4810]: I1201 16:21:19.790553 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-4dbf-account-create-update-l4q8c"] Dec 01 16:21:20 crc kubenswrapper[4810]: I1201 16:21:20.431531 4810 generic.go:334] "Generic (PLEG): container finished" podID="d4ded436-383b-493b-a03b-c65f14063d69" containerID="83803efd30b35fd5969eb155df297e46804d7964eb683c52cea0d0fa6962f5c1" exitCode=0 Dec 01 16:21:20 crc kubenswrapper[4810]: I1201 16:21:20.431718 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-4dbf-account-create-update-l4q8c" event={"ID":"d4ded436-383b-493b-a03b-c65f14063d69","Type":"ContainerDied","Data":"83803efd30b35fd5969eb155df297e46804d7964eb683c52cea0d0fa6962f5c1"} Dec 01 16:21:20 crc kubenswrapper[4810]: I1201 16:21:20.431830 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-4dbf-account-create-update-l4q8c" event={"ID":"d4ded436-383b-493b-a03b-c65f14063d69","Type":"ContainerStarted","Data":"2f04ffa2ce0dcc2cb24d26fd8b423185bce1c817ef9b448e2706fa2dd731281a"} Dec 01 16:21:20 crc kubenswrapper[4810]: I1201 16:21:20.433660 4810 generic.go:334] "Generic (PLEG): container finished" podID="6f36f4f1-7565-45e8-b96d-5b52449a4bb4" containerID="5a341e04cd35522af14e63af6a7a386eba582fd84fe4bf09171e73d6c627bd0e" exitCode=0 Dec 01 16:21:20 crc kubenswrapper[4810]: I1201 16:21:20.433759 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-4f6vj" event={"ID":"6f36f4f1-7565-45e8-b96d-5b52449a4bb4","Type":"ContainerDied","Data":"5a341e04cd35522af14e63af6a7a386eba582fd84fe4bf09171e73d6c627bd0e"} Dec 01 16:21:20 crc kubenswrapper[4810]: I1201 16:21:20.433785 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-4f6vj" event={"ID":"6f36f4f1-7565-45e8-b96d-5b52449a4bb4","Type":"ContainerStarted","Data":"7f26405ec608ec171036f52b4bb68b7822400dc44512376c99e1abfd76895b7c"} Dec 01 16:21:21 crc kubenswrapper[4810]: I1201 16:21:21.779828 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-4dbf-account-create-update-l4q8c" Dec 01 16:21:21 crc kubenswrapper[4810]: I1201 16:21:21.882949 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-4f6vj" Dec 01 16:21:21 crc kubenswrapper[4810]: I1201 16:21:21.973844 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4wd5\" (UniqueName: \"kubernetes.io/projected/d4ded436-383b-493b-a03b-c65f14063d69-kube-api-access-r4wd5\") pod \"d4ded436-383b-493b-a03b-c65f14063d69\" (UID: \"d4ded436-383b-493b-a03b-c65f14063d69\") " Dec 01 16:21:21 crc kubenswrapper[4810]: I1201 16:21:21.974332 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4ded436-383b-493b-a03b-c65f14063d69-operator-scripts\") pod \"d4ded436-383b-493b-a03b-c65f14063d69\" (UID: \"d4ded436-383b-493b-a03b-c65f14063d69\") " Dec 01 16:21:21 crc kubenswrapper[4810]: I1201 16:21:21.974654 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f36f4f1-7565-45e8-b96d-5b52449a4bb4-operator-scripts\") pod \"6f36f4f1-7565-45e8-b96d-5b52449a4bb4\" (UID: \"6f36f4f1-7565-45e8-b96d-5b52449a4bb4\") " Dec 01 16:21:21 crc kubenswrapper[4810]: I1201 16:21:21.975406 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f36f4f1-7565-45e8-b96d-5b52449a4bb4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6f36f4f1-7565-45e8-b96d-5b52449a4bb4" (UID: "6f36f4f1-7565-45e8-b96d-5b52449a4bb4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:21:21 crc kubenswrapper[4810]: I1201 16:21:21.975502 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4ded436-383b-493b-a03b-c65f14063d69-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d4ded436-383b-493b-a03b-c65f14063d69" (UID: "d4ded436-383b-493b-a03b-c65f14063d69"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:21:21 crc kubenswrapper[4810]: I1201 16:21:21.976515 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4ded436-383b-493b-a03b-c65f14063d69-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:21 crc kubenswrapper[4810]: I1201 16:21:21.976537 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f36f4f1-7565-45e8-b96d-5b52449a4bb4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:21 crc kubenswrapper[4810]: I1201 16:21:21.981991 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4ded436-383b-493b-a03b-c65f14063d69-kube-api-access-r4wd5" (OuterVolumeSpecName: "kube-api-access-r4wd5") pod "d4ded436-383b-493b-a03b-c65f14063d69" (UID: "d4ded436-383b-493b-a03b-c65f14063d69"). InnerVolumeSpecName "kube-api-access-r4wd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:21:22 crc kubenswrapper[4810]: I1201 16:21:22.077113 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrm7b\" (UniqueName: \"kubernetes.io/projected/6f36f4f1-7565-45e8-b96d-5b52449a4bb4-kube-api-access-mrm7b\") pod \"6f36f4f1-7565-45e8-b96d-5b52449a4bb4\" (UID: \"6f36f4f1-7565-45e8-b96d-5b52449a4bb4\") " Dec 01 16:21:22 crc kubenswrapper[4810]: I1201 16:21:22.078432 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4wd5\" (UniqueName: \"kubernetes.io/projected/d4ded436-383b-493b-a03b-c65f14063d69-kube-api-access-r4wd5\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:22 crc kubenswrapper[4810]: I1201 16:21:22.080203 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f36f4f1-7565-45e8-b96d-5b52449a4bb4-kube-api-access-mrm7b" (OuterVolumeSpecName: "kube-api-access-mrm7b") pod "6f36f4f1-7565-45e8-b96d-5b52449a4bb4" (UID: "6f36f4f1-7565-45e8-b96d-5b52449a4bb4"). InnerVolumeSpecName "kube-api-access-mrm7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:21:22 crc kubenswrapper[4810]: I1201 16:21:22.181028 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrm7b\" (UniqueName: \"kubernetes.io/projected/6f36f4f1-7565-45e8-b96d-5b52449a4bb4-kube-api-access-mrm7b\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:22 crc kubenswrapper[4810]: I1201 16:21:22.454541 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-4f6vj" event={"ID":"6f36f4f1-7565-45e8-b96d-5b52449a4bb4","Type":"ContainerDied","Data":"7f26405ec608ec171036f52b4bb68b7822400dc44512376c99e1abfd76895b7c"} Dec 01 16:21:22 crc kubenswrapper[4810]: I1201 16:21:22.454597 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f26405ec608ec171036f52b4bb68b7822400dc44512376c99e1abfd76895b7c" Dec 01 16:21:22 crc kubenswrapper[4810]: I1201 16:21:22.454567 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-4f6vj" Dec 01 16:21:22 crc kubenswrapper[4810]: I1201 16:21:22.456735 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-4dbf-account-create-update-l4q8c" event={"ID":"d4ded436-383b-493b-a03b-c65f14063d69","Type":"ContainerDied","Data":"2f04ffa2ce0dcc2cb24d26fd8b423185bce1c817ef9b448e2706fa2dd731281a"} Dec 01 16:21:22 crc kubenswrapper[4810]: I1201 16:21:22.456812 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f04ffa2ce0dcc2cb24d26fd8b423185bce1c817ef9b448e2706fa2dd731281a" Dec 01 16:21:22 crc kubenswrapper[4810]: I1201 16:21:22.456761 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-4dbf-account-create-update-l4q8c" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.095431 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-z84cx"] Dec 01 16:21:24 crc kubenswrapper[4810]: E1201 16:21:24.096273 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f36f4f1-7565-45e8-b96d-5b52449a4bb4" containerName="mariadb-database-create" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.096291 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f36f4f1-7565-45e8-b96d-5b52449a4bb4" containerName="mariadb-database-create" Dec 01 16:21:24 crc kubenswrapper[4810]: E1201 16:21:24.096315 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4ded436-383b-493b-a03b-c65f14063d69" containerName="mariadb-account-create-update" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.096323 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4ded436-383b-493b-a03b-c65f14063d69" containerName="mariadb-account-create-update" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.096582 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f36f4f1-7565-45e8-b96d-5b52449a4bb4" containerName="mariadb-database-create" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.096609 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4ded436-383b-493b-a03b-c65f14063d69" containerName="mariadb-account-create-update" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.098817 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-z84cx" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.101898 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.107999 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-gpxv8" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.109086 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-z84cx"] Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.224166 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd707047-2f0e-4029-846b-9b5a4de8cad7-combined-ca-bundle\") pod \"heat-db-sync-z84cx\" (UID: \"bd707047-2f0e-4029-846b-9b5a4de8cad7\") " pod="openstack/heat-db-sync-z84cx" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.224292 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9dmz\" (UniqueName: \"kubernetes.io/projected/bd707047-2f0e-4029-846b-9b5a4de8cad7-kube-api-access-n9dmz\") pod \"heat-db-sync-z84cx\" (UID: \"bd707047-2f0e-4029-846b-9b5a4de8cad7\") " pod="openstack/heat-db-sync-z84cx" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.224555 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd707047-2f0e-4029-846b-9b5a4de8cad7-config-data\") pod \"heat-db-sync-z84cx\" (UID: \"bd707047-2f0e-4029-846b-9b5a4de8cad7\") " pod="openstack/heat-db-sync-z84cx" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.326151 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd707047-2f0e-4029-846b-9b5a4de8cad7-combined-ca-bundle\") pod \"heat-db-sync-z84cx\" (UID: \"bd707047-2f0e-4029-846b-9b5a4de8cad7\") " pod="openstack/heat-db-sync-z84cx" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.326264 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9dmz\" (UniqueName: \"kubernetes.io/projected/bd707047-2f0e-4029-846b-9b5a4de8cad7-kube-api-access-n9dmz\") pod \"heat-db-sync-z84cx\" (UID: \"bd707047-2f0e-4029-846b-9b5a4de8cad7\") " pod="openstack/heat-db-sync-z84cx" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.326345 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd707047-2f0e-4029-846b-9b5a4de8cad7-config-data\") pod \"heat-db-sync-z84cx\" (UID: \"bd707047-2f0e-4029-846b-9b5a4de8cad7\") " pod="openstack/heat-db-sync-z84cx" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.331463 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd707047-2f0e-4029-846b-9b5a4de8cad7-config-data\") pod \"heat-db-sync-z84cx\" (UID: \"bd707047-2f0e-4029-846b-9b5a4de8cad7\") " pod="openstack/heat-db-sync-z84cx" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.331753 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd707047-2f0e-4029-846b-9b5a4de8cad7-combined-ca-bundle\") pod \"heat-db-sync-z84cx\" (UID: \"bd707047-2f0e-4029-846b-9b5a4de8cad7\") " pod="openstack/heat-db-sync-z84cx" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.350115 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9dmz\" (UniqueName: \"kubernetes.io/projected/bd707047-2f0e-4029-846b-9b5a4de8cad7-kube-api-access-n9dmz\") pod \"heat-db-sync-z84cx\" (UID: \"bd707047-2f0e-4029-846b-9b5a4de8cad7\") " pod="openstack/heat-db-sync-z84cx" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.419572 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-z84cx" Dec 01 16:21:24 crc kubenswrapper[4810]: I1201 16:21:24.894897 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-z84cx"] Dec 01 16:21:24 crc kubenswrapper[4810]: W1201 16:21:24.900712 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd707047_2f0e_4029_846b_9b5a4de8cad7.slice/crio-eafb9fde47b2de0a639cec62ac63bd9a92398ed04be93caba1c46c11fa48a939 WatchSource:0}: Error finding container eafb9fde47b2de0a639cec62ac63bd9a92398ed04be93caba1c46c11fa48a939: Status 404 returned error can't find the container with id eafb9fde47b2de0a639cec62ac63bd9a92398ed04be93caba1c46c11fa48a939 Dec 01 16:21:25 crc kubenswrapper[4810]: I1201 16:21:25.486561 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-z84cx" event={"ID":"bd707047-2f0e-4029-846b-9b5a4de8cad7","Type":"ContainerStarted","Data":"eafb9fde47b2de0a639cec62ac63bd9a92398ed04be93caba1c46c11fa48a939"} Dec 01 16:21:27 crc kubenswrapper[4810]: I1201 16:21:27.761445 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:27 crc kubenswrapper[4810]: I1201 16:21:27.764860 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:32 crc kubenswrapper[4810]: I1201 16:21:32.545834 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-z84cx" event={"ID":"bd707047-2f0e-4029-846b-9b5a4de8cad7","Type":"ContainerStarted","Data":"1ff4a064bacfef6ce9cf508b9b97be081917f14446589ff03b147b9176491865"} Dec 01 16:21:32 crc kubenswrapper[4810]: I1201 16:21:32.595215 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-z84cx" podStartSLOduration=1.848890446 podStartE2EDuration="8.595185837s" podCreationTimestamp="2025-12-01 16:21:24 +0000 UTC" firstStartedPulling="2025-12-01 16:21:24.903677156 +0000 UTC m=+6450.667186759" lastFinishedPulling="2025-12-01 16:21:31.649972547 +0000 UTC m=+6457.413482150" observedRunningTime="2025-12-01 16:21:32.566149709 +0000 UTC m=+6458.329659352" watchObservedRunningTime="2025-12-01 16:21:32.595185837 +0000 UTC m=+6458.358695500" Dec 01 16:21:32 crc kubenswrapper[4810]: I1201 16:21:32.972429 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:21:32 crc kubenswrapper[4810]: I1201 16:21:32.972893 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:21:33 crc kubenswrapper[4810]: I1201 16:21:33.578691 4810 generic.go:334] "Generic (PLEG): container finished" podID="bd707047-2f0e-4029-846b-9b5a4de8cad7" containerID="1ff4a064bacfef6ce9cf508b9b97be081917f14446589ff03b147b9176491865" exitCode=0 Dec 01 16:21:33 crc kubenswrapper[4810]: I1201 16:21:33.578757 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-z84cx" event={"ID":"bd707047-2f0e-4029-846b-9b5a4de8cad7","Type":"ContainerDied","Data":"1ff4a064bacfef6ce9cf508b9b97be081917f14446589ff03b147b9176491865"} Dec 01 16:21:35 crc kubenswrapper[4810]: I1201 16:21:34.926273 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-z84cx" Dec 01 16:21:35 crc kubenswrapper[4810]: I1201 16:21:35.093708 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9dmz\" (UniqueName: \"kubernetes.io/projected/bd707047-2f0e-4029-846b-9b5a4de8cad7-kube-api-access-n9dmz\") pod \"bd707047-2f0e-4029-846b-9b5a4de8cad7\" (UID: \"bd707047-2f0e-4029-846b-9b5a4de8cad7\") " Dec 01 16:21:35 crc kubenswrapper[4810]: I1201 16:21:35.094030 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd707047-2f0e-4029-846b-9b5a4de8cad7-combined-ca-bundle\") pod \"bd707047-2f0e-4029-846b-9b5a4de8cad7\" (UID: \"bd707047-2f0e-4029-846b-9b5a4de8cad7\") " Dec 01 16:21:35 crc kubenswrapper[4810]: I1201 16:21:35.094058 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd707047-2f0e-4029-846b-9b5a4de8cad7-config-data\") pod \"bd707047-2f0e-4029-846b-9b5a4de8cad7\" (UID: \"bd707047-2f0e-4029-846b-9b5a4de8cad7\") " Dec 01 16:21:35 crc kubenswrapper[4810]: I1201 16:21:35.105728 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd707047-2f0e-4029-846b-9b5a4de8cad7-kube-api-access-n9dmz" (OuterVolumeSpecName: "kube-api-access-n9dmz") pod "bd707047-2f0e-4029-846b-9b5a4de8cad7" (UID: "bd707047-2f0e-4029-846b-9b5a4de8cad7"). InnerVolumeSpecName "kube-api-access-n9dmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:21:35 crc kubenswrapper[4810]: I1201 16:21:35.171068 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd707047-2f0e-4029-846b-9b5a4de8cad7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd707047-2f0e-4029-846b-9b5a4de8cad7" (UID: "bd707047-2f0e-4029-846b-9b5a4de8cad7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:21:35 crc kubenswrapper[4810]: I1201 16:21:35.196567 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9dmz\" (UniqueName: \"kubernetes.io/projected/bd707047-2f0e-4029-846b-9b5a4de8cad7-kube-api-access-n9dmz\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:35 crc kubenswrapper[4810]: I1201 16:21:35.196589 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd707047-2f0e-4029-846b-9b5a4de8cad7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:35 crc kubenswrapper[4810]: I1201 16:21:35.205151 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd707047-2f0e-4029-846b-9b5a4de8cad7-config-data" (OuterVolumeSpecName: "config-data") pod "bd707047-2f0e-4029-846b-9b5a4de8cad7" (UID: "bd707047-2f0e-4029-846b-9b5a4de8cad7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:21:35 crc kubenswrapper[4810]: I1201 16:21:35.298710 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd707047-2f0e-4029-846b-9b5a4de8cad7-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:35 crc kubenswrapper[4810]: I1201 16:21:35.621981 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-z84cx" event={"ID":"bd707047-2f0e-4029-846b-9b5a4de8cad7","Type":"ContainerDied","Data":"eafb9fde47b2de0a639cec62ac63bd9a92398ed04be93caba1c46c11fa48a939"} Dec 01 16:21:35 crc kubenswrapper[4810]: I1201 16:21:35.622019 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eafb9fde47b2de0a639cec62ac63bd9a92398ed04be93caba1c46c11fa48a939" Dec 01 16:21:35 crc kubenswrapper[4810]: I1201 16:21:35.622068 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-z84cx" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.622661 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7ddd6f685-vw58v"] Dec 01 16:21:36 crc kubenswrapper[4810]: E1201 16:21:36.623403 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd707047-2f0e-4029-846b-9b5a4de8cad7" containerName="heat-db-sync" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.623418 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd707047-2f0e-4029-846b-9b5a4de8cad7" containerName="heat-db-sync" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.623680 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd707047-2f0e-4029-846b-9b5a4de8cad7" containerName="heat-db-sync" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.624535 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7ddd6f685-vw58v" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.627821 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.628101 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-gpxv8" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.628263 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.645895 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7ddd6f685-vw58v"] Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.732184 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58744d74-c602-4e76-9f82-509d711591dc-config-data-custom\") pod \"heat-engine-7ddd6f685-vw58v\" (UID: \"58744d74-c602-4e76-9f82-509d711591dc\") " pod="openstack/heat-engine-7ddd6f685-vw58v" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.732232 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58744d74-c602-4e76-9f82-509d711591dc-combined-ca-bundle\") pod \"heat-engine-7ddd6f685-vw58v\" (UID: \"58744d74-c602-4e76-9f82-509d711591dc\") " pod="openstack/heat-engine-7ddd6f685-vw58v" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.732258 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qj2gj\" (UniqueName: \"kubernetes.io/projected/58744d74-c602-4e76-9f82-509d711591dc-kube-api-access-qj2gj\") pod \"heat-engine-7ddd6f685-vw58v\" (UID: \"58744d74-c602-4e76-9f82-509d711591dc\") " pod="openstack/heat-engine-7ddd6f685-vw58v" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.732682 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58744d74-c602-4e76-9f82-509d711591dc-config-data\") pod \"heat-engine-7ddd6f685-vw58v\" (UID: \"58744d74-c602-4e76-9f82-509d711591dc\") " pod="openstack/heat-engine-7ddd6f685-vw58v" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.798593 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6b989ffc48-bsdpj"] Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.799785 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6b989ffc48-bsdpj" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.804085 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.814031 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6b989ffc48-bsdpj"] Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.835994 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58744d74-c602-4e76-9f82-509d711591dc-config-data\") pod \"heat-engine-7ddd6f685-vw58v\" (UID: \"58744d74-c602-4e76-9f82-509d711591dc\") " pod="openstack/heat-engine-7ddd6f685-vw58v" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.836174 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58744d74-c602-4e76-9f82-509d711591dc-config-data-custom\") pod \"heat-engine-7ddd6f685-vw58v\" (UID: \"58744d74-c602-4e76-9f82-509d711591dc\") " pod="openstack/heat-engine-7ddd6f685-vw58v" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.836200 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58744d74-c602-4e76-9f82-509d711591dc-combined-ca-bundle\") pod \"heat-engine-7ddd6f685-vw58v\" (UID: \"58744d74-c602-4e76-9f82-509d711591dc\") " pod="openstack/heat-engine-7ddd6f685-vw58v" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.836220 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj2gj\" (UniqueName: \"kubernetes.io/projected/58744d74-c602-4e76-9f82-509d711591dc-kube-api-access-qj2gj\") pod \"heat-engine-7ddd6f685-vw58v\" (UID: \"58744d74-c602-4e76-9f82-509d711591dc\") " pod="openstack/heat-engine-7ddd6f685-vw58v" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.837300 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5ffcf8767d-6597k"] Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.838520 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5ffcf8767d-6597k" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.841728 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.842224 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58744d74-c602-4e76-9f82-509d711591dc-config-data-custom\") pod \"heat-engine-7ddd6f685-vw58v\" (UID: \"58744d74-c602-4e76-9f82-509d711591dc\") " pod="openstack/heat-engine-7ddd6f685-vw58v" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.842544 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58744d74-c602-4e76-9f82-509d711591dc-combined-ca-bundle\") pod \"heat-engine-7ddd6f685-vw58v\" (UID: \"58744d74-c602-4e76-9f82-509d711591dc\") " pod="openstack/heat-engine-7ddd6f685-vw58v" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.844314 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58744d74-c602-4e76-9f82-509d711591dc-config-data\") pod \"heat-engine-7ddd6f685-vw58v\" (UID: \"58744d74-c602-4e76-9f82-509d711591dc\") " pod="openstack/heat-engine-7ddd6f685-vw58v" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.868114 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qj2gj\" (UniqueName: \"kubernetes.io/projected/58744d74-c602-4e76-9f82-509d711591dc-kube-api-access-qj2gj\") pod \"heat-engine-7ddd6f685-vw58v\" (UID: \"58744d74-c602-4e76-9f82-509d711591dc\") " pod="openstack/heat-engine-7ddd6f685-vw58v" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.875250 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5ffcf8767d-6597k"] Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.937482 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d61bd8a-151d-474a-ac48-f332b48e519f-config-data\") pod \"heat-cfnapi-5ffcf8767d-6597k\" (UID: \"3d61bd8a-151d-474a-ac48-f332b48e519f\") " pod="openstack/heat-cfnapi-5ffcf8767d-6597k" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.937659 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-combined-ca-bundle\") pod \"heat-api-6b989ffc48-bsdpj\" (UID: \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\") " pod="openstack/heat-api-6b989ffc48-bsdpj" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.937890 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-config-data\") pod \"heat-api-6b989ffc48-bsdpj\" (UID: \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\") " pod="openstack/heat-api-6b989ffc48-bsdpj" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.937943 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbvd6\" (UniqueName: \"kubernetes.io/projected/3d61bd8a-151d-474a-ac48-f332b48e519f-kube-api-access-cbvd6\") pod \"heat-cfnapi-5ffcf8767d-6597k\" (UID: \"3d61bd8a-151d-474a-ac48-f332b48e519f\") " pod="openstack/heat-cfnapi-5ffcf8767d-6597k" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.938013 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d61bd8a-151d-474a-ac48-f332b48e519f-combined-ca-bundle\") pod \"heat-cfnapi-5ffcf8767d-6597k\" (UID: \"3d61bd8a-151d-474a-ac48-f332b48e519f\") " pod="openstack/heat-cfnapi-5ffcf8767d-6597k" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.938075 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qprvh\" (UniqueName: \"kubernetes.io/projected/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-kube-api-access-qprvh\") pod \"heat-api-6b989ffc48-bsdpj\" (UID: \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\") " pod="openstack/heat-api-6b989ffc48-bsdpj" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.938153 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-config-data-custom\") pod \"heat-api-6b989ffc48-bsdpj\" (UID: \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\") " pod="openstack/heat-api-6b989ffc48-bsdpj" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.938230 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d61bd8a-151d-474a-ac48-f332b48e519f-config-data-custom\") pod \"heat-cfnapi-5ffcf8767d-6597k\" (UID: \"3d61bd8a-151d-474a-ac48-f332b48e519f\") " pod="openstack/heat-cfnapi-5ffcf8767d-6597k" Dec 01 16:21:36 crc kubenswrapper[4810]: I1201 16:21:36.982531 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7ddd6f685-vw58v" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.040605 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d61bd8a-151d-474a-ac48-f332b48e519f-config-data-custom\") pod \"heat-cfnapi-5ffcf8767d-6597k\" (UID: \"3d61bd8a-151d-474a-ac48-f332b48e519f\") " pod="openstack/heat-cfnapi-5ffcf8767d-6597k" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.041885 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d61bd8a-151d-474a-ac48-f332b48e519f-config-data\") pod \"heat-cfnapi-5ffcf8767d-6597k\" (UID: \"3d61bd8a-151d-474a-ac48-f332b48e519f\") " pod="openstack/heat-cfnapi-5ffcf8767d-6597k" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.042218 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-combined-ca-bundle\") pod \"heat-api-6b989ffc48-bsdpj\" (UID: \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\") " pod="openstack/heat-api-6b989ffc48-bsdpj" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.043046 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-config-data\") pod \"heat-api-6b989ffc48-bsdpj\" (UID: \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\") " pod="openstack/heat-api-6b989ffc48-bsdpj" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.044543 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbvd6\" (UniqueName: \"kubernetes.io/projected/3d61bd8a-151d-474a-ac48-f332b48e519f-kube-api-access-cbvd6\") pod \"heat-cfnapi-5ffcf8767d-6597k\" (UID: \"3d61bd8a-151d-474a-ac48-f332b48e519f\") " pod="openstack/heat-cfnapi-5ffcf8767d-6597k" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.044790 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d61bd8a-151d-474a-ac48-f332b48e519f-combined-ca-bundle\") pod \"heat-cfnapi-5ffcf8767d-6597k\" (UID: \"3d61bd8a-151d-474a-ac48-f332b48e519f\") " pod="openstack/heat-cfnapi-5ffcf8767d-6597k" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.045042 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qprvh\" (UniqueName: \"kubernetes.io/projected/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-kube-api-access-qprvh\") pod \"heat-api-6b989ffc48-bsdpj\" (UID: \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\") " pod="openstack/heat-api-6b989ffc48-bsdpj" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.045804 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-config-data-custom\") pod \"heat-api-6b989ffc48-bsdpj\" (UID: \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\") " pod="openstack/heat-api-6b989ffc48-bsdpj" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.048233 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-combined-ca-bundle\") pod \"heat-api-6b989ffc48-bsdpj\" (UID: \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\") " pod="openstack/heat-api-6b989ffc48-bsdpj" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.049552 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d61bd8a-151d-474a-ac48-f332b48e519f-config-data\") pod \"heat-cfnapi-5ffcf8767d-6597k\" (UID: \"3d61bd8a-151d-474a-ac48-f332b48e519f\") " pod="openstack/heat-cfnapi-5ffcf8767d-6597k" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.050185 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d61bd8a-151d-474a-ac48-f332b48e519f-config-data-custom\") pod \"heat-cfnapi-5ffcf8767d-6597k\" (UID: \"3d61bd8a-151d-474a-ac48-f332b48e519f\") " pod="openstack/heat-cfnapi-5ffcf8767d-6597k" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.052996 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-config-data-custom\") pod \"heat-api-6b989ffc48-bsdpj\" (UID: \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\") " pod="openstack/heat-api-6b989ffc48-bsdpj" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.060081 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-config-data\") pod \"heat-api-6b989ffc48-bsdpj\" (UID: \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\") " pod="openstack/heat-api-6b989ffc48-bsdpj" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.062176 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d61bd8a-151d-474a-ac48-f332b48e519f-combined-ca-bundle\") pod \"heat-cfnapi-5ffcf8767d-6597k\" (UID: \"3d61bd8a-151d-474a-ac48-f332b48e519f\") " pod="openstack/heat-cfnapi-5ffcf8767d-6597k" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.065085 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qprvh\" (UniqueName: \"kubernetes.io/projected/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-kube-api-access-qprvh\") pod \"heat-api-6b989ffc48-bsdpj\" (UID: \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\") " pod="openstack/heat-api-6b989ffc48-bsdpj" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.070245 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbvd6\" (UniqueName: \"kubernetes.io/projected/3d61bd8a-151d-474a-ac48-f332b48e519f-kube-api-access-cbvd6\") pod \"heat-cfnapi-5ffcf8767d-6597k\" (UID: \"3d61bd8a-151d-474a-ac48-f332b48e519f\") " pod="openstack/heat-cfnapi-5ffcf8767d-6597k" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.127345 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6b989ffc48-bsdpj" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.263723 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5ffcf8767d-6597k" Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.589862 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7ddd6f685-vw58v"] Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.659412 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7ddd6f685-vw58v" event={"ID":"58744d74-c602-4e76-9f82-509d711591dc","Type":"ContainerStarted","Data":"7e3f7b5375638281b5a791838a4a769a034de5b76f353ccaa48bd7c29e8a3b14"} Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.743087 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6b989ffc48-bsdpj"] Dec 01 16:21:37 crc kubenswrapper[4810]: I1201 16:21:37.836713 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5ffcf8767d-6597k"] Dec 01 16:21:38 crc kubenswrapper[4810]: I1201 16:21:38.690313 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5ffcf8767d-6597k" event={"ID":"3d61bd8a-151d-474a-ac48-f332b48e519f","Type":"ContainerStarted","Data":"5972d65af7df91edf5610fb5105c0bcc27c1640cbf037ed832f466df26885da0"} Dec 01 16:21:38 crc kubenswrapper[4810]: I1201 16:21:38.693494 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6b989ffc48-bsdpj" event={"ID":"7d8e7e3c-e906-4197-9631-ae27a3ed10b1","Type":"ContainerStarted","Data":"de0ed53ecd16473dc15bf532b625bca2319484bc1f7896c9a3c50aaf2e50f4c2"} Dec 01 16:21:38 crc kubenswrapper[4810]: I1201 16:21:38.696621 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7ddd6f685-vw58v" event={"ID":"58744d74-c602-4e76-9f82-509d711591dc","Type":"ContainerStarted","Data":"1b0c435a2665fd022d98359bc9ffae1a94fa1e75a34267d1703fbdabb8588d1c"} Dec 01 16:21:38 crc kubenswrapper[4810]: I1201 16:21:38.696906 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7ddd6f685-vw58v" Dec 01 16:21:38 crc kubenswrapper[4810]: I1201 16:21:38.719980 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7ddd6f685-vw58v" podStartSLOduration=2.719960957 podStartE2EDuration="2.719960957s" podCreationTimestamp="2025-12-01 16:21:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:21:38.713634954 +0000 UTC m=+6464.477144577" watchObservedRunningTime="2025-12-01 16:21:38.719960957 +0000 UTC m=+6464.483470560" Dec 01 16:21:39 crc kubenswrapper[4810]: I1201 16:21:39.889078 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:40 crc kubenswrapper[4810]: I1201 16:21:40.742010 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5ffcf8767d-6597k" event={"ID":"3d61bd8a-151d-474a-ac48-f332b48e519f","Type":"ContainerStarted","Data":"a1fea62c437047870ed241cef5dd614ed1e0c60ac60756d969ca7e3f46c870cb"} Dec 01 16:21:40 crc kubenswrapper[4810]: I1201 16:21:40.742676 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5ffcf8767d-6597k" Dec 01 16:21:40 crc kubenswrapper[4810]: I1201 16:21:40.764236 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5ffcf8767d-6597k" podStartSLOduration=2.867921593 podStartE2EDuration="4.7642203s" podCreationTimestamp="2025-12-01 16:21:36 +0000 UTC" firstStartedPulling="2025-12-01 16:21:37.833688977 +0000 UTC m=+6463.597198580" lastFinishedPulling="2025-12-01 16:21:39.729987654 +0000 UTC m=+6465.493497287" observedRunningTime="2025-12-01 16:21:40.7609063 +0000 UTC m=+6466.524415903" watchObservedRunningTime="2025-12-01 16:21:40.7642203 +0000 UTC m=+6466.527729903" Dec 01 16:21:41 crc kubenswrapper[4810]: I1201 16:21:41.753531 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6b989ffc48-bsdpj" event={"ID":"7d8e7e3c-e906-4197-9631-ae27a3ed10b1","Type":"ContainerStarted","Data":"eafae05c44b1790037d1812792777a021056624f4de00bba02e1b0999e67b168"} Dec 01 16:21:41 crc kubenswrapper[4810]: I1201 16:21:41.754917 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6b989ffc48-bsdpj" Dec 01 16:21:41 crc kubenswrapper[4810]: I1201 16:21:41.781367 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6b989ffc48-bsdpj" podStartSLOduration=2.438530816 podStartE2EDuration="5.781337371s" podCreationTimestamp="2025-12-01 16:21:36 +0000 UTC" firstStartedPulling="2025-12-01 16:21:37.745729329 +0000 UTC m=+6463.509238932" lastFinishedPulling="2025-12-01 16:21:41.088535884 +0000 UTC m=+6466.852045487" observedRunningTime="2025-12-01 16:21:41.776124631 +0000 UTC m=+6467.539634224" watchObservedRunningTime="2025-12-01 16:21:41.781337371 +0000 UTC m=+6467.544846974" Dec 01 16:21:41 crc kubenswrapper[4810]: I1201 16:21:41.924860 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-574946f46-cwn7p" Dec 01 16:21:41 crc kubenswrapper[4810]: I1201 16:21:41.999649 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-57bd9c5d44-hps7g"] Dec 01 16:21:41 crc kubenswrapper[4810]: I1201 16:21:41.999884 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-57bd9c5d44-hps7g" podUID="bfad2664-5a86-48b8-b3a5-cad4eec57408" containerName="horizon-log" containerID="cri-o://ff427cf2a241a76a12a13f31fbfd44c3fab7d63ef78a625e7e10a45cd6d0ddf2" gracePeriod=30 Dec 01 16:21:42 crc kubenswrapper[4810]: I1201 16:21:42.000003 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-57bd9c5d44-hps7g" podUID="bfad2664-5a86-48b8-b3a5-cad4eec57408" containerName="horizon" containerID="cri-o://d7079fd9f393e10798a10ad299a51d5120a549c4ac48bec74590bab9fbce49f0" gracePeriod=30 Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.731172 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-656cb86f76-jpbk2"] Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.733275 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-656cb86f76-jpbk2" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.747642 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-7c9cc4f87-5xtr9"] Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.749048 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7c9cc4f87-5xtr9" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.773440 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-656cb86f76-jpbk2"] Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.797354 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-74bcc475c7-2mx5s"] Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.816193 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.819657 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-combined-ca-bundle\") pod \"heat-api-7c9cc4f87-5xtr9\" (UID: \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\") " pod="openstack/heat-api-7c9cc4f87-5xtr9" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.819701 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-config-data\") pod \"heat-api-7c9cc4f87-5xtr9\" (UID: \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\") " pod="openstack/heat-api-7c9cc4f87-5xtr9" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.819725 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc84fc7d-8fdc-405a-a45f-d80723049d83-config-data\") pod \"heat-cfnapi-74bcc475c7-2mx5s\" (UID: \"fc84fc7d-8fdc-405a-a45f-d80723049d83\") " pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.819764 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-config-data-custom\") pod \"heat-api-7c9cc4f87-5xtr9\" (UID: \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\") " pod="openstack/heat-api-7c9cc4f87-5xtr9" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.819802 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9890a860-9f91-4a78-b26b-23c6c97a130d-config-data-custom\") pod \"heat-engine-656cb86f76-jpbk2\" (UID: \"9890a860-9f91-4a78-b26b-23c6c97a130d\") " pod="openstack/heat-engine-656cb86f76-jpbk2" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.819827 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ldlp\" (UniqueName: \"kubernetes.io/projected/fc84fc7d-8fdc-405a-a45f-d80723049d83-kube-api-access-7ldlp\") pod \"heat-cfnapi-74bcc475c7-2mx5s\" (UID: \"fc84fc7d-8fdc-405a-a45f-d80723049d83\") " pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.819871 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9890a860-9f91-4a78-b26b-23c6c97a130d-combined-ca-bundle\") pod \"heat-engine-656cb86f76-jpbk2\" (UID: \"9890a860-9f91-4a78-b26b-23c6c97a130d\") " pod="openstack/heat-engine-656cb86f76-jpbk2" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.819902 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc84fc7d-8fdc-405a-a45f-d80723049d83-combined-ca-bundle\") pod \"heat-cfnapi-74bcc475c7-2mx5s\" (UID: \"fc84fc7d-8fdc-405a-a45f-d80723049d83\") " pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.819931 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42gwg\" (UniqueName: \"kubernetes.io/projected/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-kube-api-access-42gwg\") pod \"heat-api-7c9cc4f87-5xtr9\" (UID: \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\") " pod="openstack/heat-api-7c9cc4f87-5xtr9" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.819957 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc84fc7d-8fdc-405a-a45f-d80723049d83-config-data-custom\") pod \"heat-cfnapi-74bcc475c7-2mx5s\" (UID: \"fc84fc7d-8fdc-405a-a45f-d80723049d83\") " pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.820022 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbms4\" (UniqueName: \"kubernetes.io/projected/9890a860-9f91-4a78-b26b-23c6c97a130d-kube-api-access-dbms4\") pod \"heat-engine-656cb86f76-jpbk2\" (UID: \"9890a860-9f91-4a78-b26b-23c6c97a130d\") " pod="openstack/heat-engine-656cb86f76-jpbk2" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.820053 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9890a860-9f91-4a78-b26b-23c6c97a130d-config-data\") pod \"heat-engine-656cb86f76-jpbk2\" (UID: \"9890a860-9f91-4a78-b26b-23c6c97a130d\") " pod="openstack/heat-engine-656cb86f76-jpbk2" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.825499 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7c9cc4f87-5xtr9"] Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.835636 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-74bcc475c7-2mx5s"] Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.921081 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9890a860-9f91-4a78-b26b-23c6c97a130d-config-data-custom\") pod \"heat-engine-656cb86f76-jpbk2\" (UID: \"9890a860-9f91-4a78-b26b-23c6c97a130d\") " pod="openstack/heat-engine-656cb86f76-jpbk2" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.921260 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ldlp\" (UniqueName: \"kubernetes.io/projected/fc84fc7d-8fdc-405a-a45f-d80723049d83-kube-api-access-7ldlp\") pod \"heat-cfnapi-74bcc475c7-2mx5s\" (UID: \"fc84fc7d-8fdc-405a-a45f-d80723049d83\") " pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.921320 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9890a860-9f91-4a78-b26b-23c6c97a130d-combined-ca-bundle\") pod \"heat-engine-656cb86f76-jpbk2\" (UID: \"9890a860-9f91-4a78-b26b-23c6c97a130d\") " pod="openstack/heat-engine-656cb86f76-jpbk2" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.921355 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc84fc7d-8fdc-405a-a45f-d80723049d83-combined-ca-bundle\") pod \"heat-cfnapi-74bcc475c7-2mx5s\" (UID: \"fc84fc7d-8fdc-405a-a45f-d80723049d83\") " pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.921388 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42gwg\" (UniqueName: \"kubernetes.io/projected/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-kube-api-access-42gwg\") pod \"heat-api-7c9cc4f87-5xtr9\" (UID: \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\") " pod="openstack/heat-api-7c9cc4f87-5xtr9" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.921421 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc84fc7d-8fdc-405a-a45f-d80723049d83-config-data-custom\") pod \"heat-cfnapi-74bcc475c7-2mx5s\" (UID: \"fc84fc7d-8fdc-405a-a45f-d80723049d83\") " pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.921536 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbms4\" (UniqueName: \"kubernetes.io/projected/9890a860-9f91-4a78-b26b-23c6c97a130d-kube-api-access-dbms4\") pod \"heat-engine-656cb86f76-jpbk2\" (UID: \"9890a860-9f91-4a78-b26b-23c6c97a130d\") " pod="openstack/heat-engine-656cb86f76-jpbk2" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.921564 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9890a860-9f91-4a78-b26b-23c6c97a130d-config-data\") pod \"heat-engine-656cb86f76-jpbk2\" (UID: \"9890a860-9f91-4a78-b26b-23c6c97a130d\") " pod="openstack/heat-engine-656cb86f76-jpbk2" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.921587 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-combined-ca-bundle\") pod \"heat-api-7c9cc4f87-5xtr9\" (UID: \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\") " pod="openstack/heat-api-7c9cc4f87-5xtr9" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.921605 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-config-data\") pod \"heat-api-7c9cc4f87-5xtr9\" (UID: \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\") " pod="openstack/heat-api-7c9cc4f87-5xtr9" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.921722 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc84fc7d-8fdc-405a-a45f-d80723049d83-config-data\") pod \"heat-cfnapi-74bcc475c7-2mx5s\" (UID: \"fc84fc7d-8fdc-405a-a45f-d80723049d83\") " pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.921789 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-config-data-custom\") pod \"heat-api-7c9cc4f87-5xtr9\" (UID: \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\") " pod="openstack/heat-api-7c9cc4f87-5xtr9" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.928937 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc84fc7d-8fdc-405a-a45f-d80723049d83-config-data-custom\") pod \"heat-cfnapi-74bcc475c7-2mx5s\" (UID: \"fc84fc7d-8fdc-405a-a45f-d80723049d83\") " pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.930124 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9890a860-9f91-4a78-b26b-23c6c97a130d-config-data-custom\") pod \"heat-engine-656cb86f76-jpbk2\" (UID: \"9890a860-9f91-4a78-b26b-23c6c97a130d\") " pod="openstack/heat-engine-656cb86f76-jpbk2" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.930486 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9890a860-9f91-4a78-b26b-23c6c97a130d-config-data\") pod \"heat-engine-656cb86f76-jpbk2\" (UID: \"9890a860-9f91-4a78-b26b-23c6c97a130d\") " pod="openstack/heat-engine-656cb86f76-jpbk2" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.930524 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc84fc7d-8fdc-405a-a45f-d80723049d83-config-data\") pod \"heat-cfnapi-74bcc475c7-2mx5s\" (UID: \"fc84fc7d-8fdc-405a-a45f-d80723049d83\") " pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.930709 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9890a860-9f91-4a78-b26b-23c6c97a130d-combined-ca-bundle\") pod \"heat-engine-656cb86f76-jpbk2\" (UID: \"9890a860-9f91-4a78-b26b-23c6c97a130d\") " pod="openstack/heat-engine-656cb86f76-jpbk2" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.931656 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-config-data-custom\") pod \"heat-api-7c9cc4f87-5xtr9\" (UID: \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\") " pod="openstack/heat-api-7c9cc4f87-5xtr9" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.937822 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc84fc7d-8fdc-405a-a45f-d80723049d83-combined-ca-bundle\") pod \"heat-cfnapi-74bcc475c7-2mx5s\" (UID: \"fc84fc7d-8fdc-405a-a45f-d80723049d83\") " pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.938940 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-combined-ca-bundle\") pod \"heat-api-7c9cc4f87-5xtr9\" (UID: \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\") " pod="openstack/heat-api-7c9cc4f87-5xtr9" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.943625 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-config-data\") pod \"heat-api-7c9cc4f87-5xtr9\" (UID: \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\") " pod="openstack/heat-api-7c9cc4f87-5xtr9" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.950505 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42gwg\" (UniqueName: \"kubernetes.io/projected/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-kube-api-access-42gwg\") pod \"heat-api-7c9cc4f87-5xtr9\" (UID: \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\") " pod="openstack/heat-api-7c9cc4f87-5xtr9" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.950844 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ldlp\" (UniqueName: \"kubernetes.io/projected/fc84fc7d-8fdc-405a-a45f-d80723049d83-kube-api-access-7ldlp\") pod \"heat-cfnapi-74bcc475c7-2mx5s\" (UID: \"fc84fc7d-8fdc-405a-a45f-d80723049d83\") " pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" Dec 01 16:21:43 crc kubenswrapper[4810]: I1201 16:21:43.958044 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbms4\" (UniqueName: \"kubernetes.io/projected/9890a860-9f91-4a78-b26b-23c6c97a130d-kube-api-access-dbms4\") pod \"heat-engine-656cb86f76-jpbk2\" (UID: \"9890a860-9f91-4a78-b26b-23c6c97a130d\") " pod="openstack/heat-engine-656cb86f76-jpbk2" Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.054003 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-656cb86f76-jpbk2" Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.074938 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7c9cc4f87-5xtr9" Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.133944 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.572109 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-656cb86f76-jpbk2"] Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.664811 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7c9cc4f87-5xtr9"] Dec 01 16:21:44 crc kubenswrapper[4810]: W1201 16:21:44.665297 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c60ab7f_e4c6_4ab5_ae10_03fdbeacda7a.slice/crio-b947963ac56b211ca5e6035107ed1c0c8c00d25c79ecc1cd313bbaf3c73d2957 WatchSource:0}: Error finding container b947963ac56b211ca5e6035107ed1c0c8c00d25c79ecc1cd313bbaf3c73d2957: Status 404 returned error can't find the container with id b947963ac56b211ca5e6035107ed1c0c8c00d25c79ecc1cd313bbaf3c73d2957 Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.769170 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-74bcc475c7-2mx5s"] Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.795793 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" event={"ID":"fc84fc7d-8fdc-405a-a45f-d80723049d83","Type":"ContainerStarted","Data":"5eab2870035785206105d2070c9117adc4fe15ab23cce598067c9125fc867f90"} Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.800501 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-656cb86f76-jpbk2" event={"ID":"9890a860-9f91-4a78-b26b-23c6c97a130d","Type":"ContainerStarted","Data":"3f90b451768bab4c3988e08ba5bb19087333855a60df1d14dccbd52e03ac8b91"} Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.800538 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-656cb86f76-jpbk2" event={"ID":"9890a860-9f91-4a78-b26b-23c6c97a130d","Type":"ContainerStarted","Data":"3a5bd12b76b552c977cbfdd5cfcf8f93f41938e73542c44eace3fe7f8b78627c"} Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.801309 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-656cb86f76-jpbk2" Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.802760 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7c9cc4f87-5xtr9" event={"ID":"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a","Type":"ContainerStarted","Data":"b947963ac56b211ca5e6035107ed1c0c8c00d25c79ecc1cd313bbaf3c73d2957"} Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.824099 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-656cb86f76-jpbk2" podStartSLOduration=1.824079282 podStartE2EDuration="1.824079282s" podCreationTimestamp="2025-12-01 16:21:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:21:44.81583897 +0000 UTC m=+6470.579348573" watchObservedRunningTime="2025-12-01 16:21:44.824079282 +0000 UTC m=+6470.587588885" Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.915402 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6b989ffc48-bsdpj"] Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.915621 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-6b989ffc48-bsdpj" podUID="7d8e7e3c-e906-4197-9631-ae27a3ed10b1" containerName="heat-api" containerID="cri-o://eafae05c44b1790037d1812792777a021056624f4de00bba02e1b0999e67b168" gracePeriod=60 Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.926929 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-5ffcf8767d-6597k"] Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.927111 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-5ffcf8767d-6597k" podUID="3d61bd8a-151d-474a-ac48-f332b48e519f" containerName="heat-cfnapi" containerID="cri-o://a1fea62c437047870ed241cef5dd614ed1e0c60ac60756d969ca7e3f46c870cb" gracePeriod=60 Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.955046 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-db946669-jbbqh"] Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.956521 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.967758 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.969885 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7d4fd6fc6c-6wbqj"] Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.970072 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.971288 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.976608 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.977829 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Dec 01 16:21:44 crc kubenswrapper[4810]: I1201 16:21:44.984054 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-db946669-jbbqh"] Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.032952 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7d4fd6fc6c-6wbqj"] Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.066702 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-5ffcf8767d-6597k" podUID="3d61bd8a-151d-474a-ac48-f332b48e519f" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.1.127:8000/healthcheck\": EOF" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.153376 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/24dcd733-d9bb-4def-9325-f1dd2611d1f1-internal-tls-certs\") pod \"heat-cfnapi-7d4fd6fc6c-6wbqj\" (UID: \"24dcd733-d9bb-4def-9325-f1dd2611d1f1\") " pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.153427 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/24dcd733-d9bb-4def-9325-f1dd2611d1f1-public-tls-certs\") pod \"heat-cfnapi-7d4fd6fc6c-6wbqj\" (UID: \"24dcd733-d9bb-4def-9325-f1dd2611d1f1\") " pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.153574 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b1083d7-8870-4fbb-b4aa-a25dfaef9b14-config-data-custom\") pod \"heat-api-db946669-jbbqh\" (UID: \"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14\") " pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.153637 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5wvm\" (UniqueName: \"kubernetes.io/projected/0b1083d7-8870-4fbb-b4aa-a25dfaef9b14-kube-api-access-k5wvm\") pod \"heat-api-db946669-jbbqh\" (UID: \"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14\") " pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.153796 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24dcd733-d9bb-4def-9325-f1dd2611d1f1-config-data\") pod \"heat-cfnapi-7d4fd6fc6c-6wbqj\" (UID: \"24dcd733-d9bb-4def-9325-f1dd2611d1f1\") " pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.153908 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2drn\" (UniqueName: \"kubernetes.io/projected/24dcd733-d9bb-4def-9325-f1dd2611d1f1-kube-api-access-r2drn\") pod \"heat-cfnapi-7d4fd6fc6c-6wbqj\" (UID: \"24dcd733-d9bb-4def-9325-f1dd2611d1f1\") " pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.153964 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b1083d7-8870-4fbb-b4aa-a25dfaef9b14-public-tls-certs\") pod \"heat-api-db946669-jbbqh\" (UID: \"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14\") " pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.153995 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24dcd733-d9bb-4def-9325-f1dd2611d1f1-config-data-custom\") pod \"heat-cfnapi-7d4fd6fc6c-6wbqj\" (UID: \"24dcd733-d9bb-4def-9325-f1dd2611d1f1\") " pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.154087 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b1083d7-8870-4fbb-b4aa-a25dfaef9b14-internal-tls-certs\") pod \"heat-api-db946669-jbbqh\" (UID: \"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14\") " pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.154106 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24dcd733-d9bb-4def-9325-f1dd2611d1f1-combined-ca-bundle\") pod \"heat-cfnapi-7d4fd6fc6c-6wbqj\" (UID: \"24dcd733-d9bb-4def-9325-f1dd2611d1f1\") " pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.154185 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b1083d7-8870-4fbb-b4aa-a25dfaef9b14-combined-ca-bundle\") pod \"heat-api-db946669-jbbqh\" (UID: \"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14\") " pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.154225 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b1083d7-8870-4fbb-b4aa-a25dfaef9b14-config-data\") pod \"heat-api-db946669-jbbqh\" (UID: \"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14\") " pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.255831 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/24dcd733-d9bb-4def-9325-f1dd2611d1f1-internal-tls-certs\") pod \"heat-cfnapi-7d4fd6fc6c-6wbqj\" (UID: \"24dcd733-d9bb-4def-9325-f1dd2611d1f1\") " pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.255910 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/24dcd733-d9bb-4def-9325-f1dd2611d1f1-public-tls-certs\") pod \"heat-cfnapi-7d4fd6fc6c-6wbqj\" (UID: \"24dcd733-d9bb-4def-9325-f1dd2611d1f1\") " pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.255960 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b1083d7-8870-4fbb-b4aa-a25dfaef9b14-config-data-custom\") pod \"heat-api-db946669-jbbqh\" (UID: \"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14\") " pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.255988 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5wvm\" (UniqueName: \"kubernetes.io/projected/0b1083d7-8870-4fbb-b4aa-a25dfaef9b14-kube-api-access-k5wvm\") pod \"heat-api-db946669-jbbqh\" (UID: \"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14\") " pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.256073 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24dcd733-d9bb-4def-9325-f1dd2611d1f1-config-data\") pod \"heat-cfnapi-7d4fd6fc6c-6wbqj\" (UID: \"24dcd733-d9bb-4def-9325-f1dd2611d1f1\") " pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.256110 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2drn\" (UniqueName: \"kubernetes.io/projected/24dcd733-d9bb-4def-9325-f1dd2611d1f1-kube-api-access-r2drn\") pod \"heat-cfnapi-7d4fd6fc6c-6wbqj\" (UID: \"24dcd733-d9bb-4def-9325-f1dd2611d1f1\") " pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.256139 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b1083d7-8870-4fbb-b4aa-a25dfaef9b14-public-tls-certs\") pod \"heat-api-db946669-jbbqh\" (UID: \"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14\") " pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.256160 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24dcd733-d9bb-4def-9325-f1dd2611d1f1-config-data-custom\") pod \"heat-cfnapi-7d4fd6fc6c-6wbqj\" (UID: \"24dcd733-d9bb-4def-9325-f1dd2611d1f1\") " pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.256193 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b1083d7-8870-4fbb-b4aa-a25dfaef9b14-internal-tls-certs\") pod \"heat-api-db946669-jbbqh\" (UID: \"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14\") " pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.256212 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24dcd733-d9bb-4def-9325-f1dd2611d1f1-combined-ca-bundle\") pod \"heat-cfnapi-7d4fd6fc6c-6wbqj\" (UID: \"24dcd733-d9bb-4def-9325-f1dd2611d1f1\") " pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.256252 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b1083d7-8870-4fbb-b4aa-a25dfaef9b14-combined-ca-bundle\") pod \"heat-api-db946669-jbbqh\" (UID: \"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14\") " pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.256273 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b1083d7-8870-4fbb-b4aa-a25dfaef9b14-config-data\") pod \"heat-api-db946669-jbbqh\" (UID: \"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14\") " pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.262373 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b1083d7-8870-4fbb-b4aa-a25dfaef9b14-internal-tls-certs\") pod \"heat-api-db946669-jbbqh\" (UID: \"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14\") " pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.262647 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24dcd733-d9bb-4def-9325-f1dd2611d1f1-combined-ca-bundle\") pod \"heat-cfnapi-7d4fd6fc6c-6wbqj\" (UID: \"24dcd733-d9bb-4def-9325-f1dd2611d1f1\") " pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.262890 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b1083d7-8870-4fbb-b4aa-a25dfaef9b14-config-data\") pod \"heat-api-db946669-jbbqh\" (UID: \"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14\") " pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.263150 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/24dcd733-d9bb-4def-9325-f1dd2611d1f1-public-tls-certs\") pod \"heat-cfnapi-7d4fd6fc6c-6wbqj\" (UID: \"24dcd733-d9bb-4def-9325-f1dd2611d1f1\") " pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.263455 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24dcd733-d9bb-4def-9325-f1dd2611d1f1-config-data\") pod \"heat-cfnapi-7d4fd6fc6c-6wbqj\" (UID: \"24dcd733-d9bb-4def-9325-f1dd2611d1f1\") " pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.264005 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24dcd733-d9bb-4def-9325-f1dd2611d1f1-config-data-custom\") pod \"heat-cfnapi-7d4fd6fc6c-6wbqj\" (UID: \"24dcd733-d9bb-4def-9325-f1dd2611d1f1\") " pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.264219 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/24dcd733-d9bb-4def-9325-f1dd2611d1f1-internal-tls-certs\") pod \"heat-cfnapi-7d4fd6fc6c-6wbqj\" (UID: \"24dcd733-d9bb-4def-9325-f1dd2611d1f1\") " pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.264303 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b1083d7-8870-4fbb-b4aa-a25dfaef9b14-combined-ca-bundle\") pod \"heat-api-db946669-jbbqh\" (UID: \"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14\") " pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.267641 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0b1083d7-8870-4fbb-b4aa-a25dfaef9b14-config-data-custom\") pod \"heat-api-db946669-jbbqh\" (UID: \"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14\") " pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.272134 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b1083d7-8870-4fbb-b4aa-a25dfaef9b14-public-tls-certs\") pod \"heat-api-db946669-jbbqh\" (UID: \"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14\") " pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.273112 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2drn\" (UniqueName: \"kubernetes.io/projected/24dcd733-d9bb-4def-9325-f1dd2611d1f1-kube-api-access-r2drn\") pod \"heat-cfnapi-7d4fd6fc6c-6wbqj\" (UID: \"24dcd733-d9bb-4def-9325-f1dd2611d1f1\") " pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.277742 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-57bd9c5d44-hps7g" podUID="bfad2664-5a86-48b8-b3a5-cad4eec57408" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.118:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:32886->10.217.1.118:8443: read: connection reset by peer" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.278321 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5wvm\" (UniqueName: \"kubernetes.io/projected/0b1083d7-8870-4fbb-b4aa-a25dfaef9b14-kube-api-access-k5wvm\") pod \"heat-api-db946669-jbbqh\" (UID: \"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14\") " pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.342979 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.360888 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.813762 4810 generic.go:334] "Generic (PLEG): container finished" podID="fc84fc7d-8fdc-405a-a45f-d80723049d83" containerID="b5311f1891a721050d98a01c298de6fc5649ab078169f2a68924b05175d268ed" exitCode=1 Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.813989 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" event={"ID":"fc84fc7d-8fdc-405a-a45f-d80723049d83","Type":"ContainerDied","Data":"b5311f1891a721050d98a01c298de6fc5649ab078169f2a68924b05175d268ed"} Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.814572 4810 scope.go:117] "RemoveContainer" containerID="b5311f1891a721050d98a01c298de6fc5649ab078169f2a68924b05175d268ed" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.818901 4810 generic.go:334] "Generic (PLEG): container finished" podID="3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a" containerID="8c22b4488e6b3869e8d1b75a7a572db0cb6a67b204cf9027932b2b5c00e77620" exitCode=1 Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.818962 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7c9cc4f87-5xtr9" event={"ID":"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a","Type":"ContainerDied","Data":"8c22b4488e6b3869e8d1b75a7a572db0cb6a67b204cf9027932b2b5c00e77620"} Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.819491 4810 scope.go:117] "RemoveContainer" containerID="8c22b4488e6b3869e8d1b75a7a572db0cb6a67b204cf9027932b2b5c00e77620" Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.834048 4810 generic.go:334] "Generic (PLEG): container finished" podID="7d8e7e3c-e906-4197-9631-ae27a3ed10b1" containerID="eafae05c44b1790037d1812792777a021056624f4de00bba02e1b0999e67b168" exitCode=0 Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.834124 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6b989ffc48-bsdpj" event={"ID":"7d8e7e3c-e906-4197-9631-ae27a3ed10b1","Type":"ContainerDied","Data":"eafae05c44b1790037d1812792777a021056624f4de00bba02e1b0999e67b168"} Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.845140 4810 generic.go:334] "Generic (PLEG): container finished" podID="bfad2664-5a86-48b8-b3a5-cad4eec57408" containerID="d7079fd9f393e10798a10ad299a51d5120a549c4ac48bec74590bab9fbce49f0" exitCode=0 Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.845280 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57bd9c5d44-hps7g" event={"ID":"bfad2664-5a86-48b8-b3a5-cad4eec57408","Type":"ContainerDied","Data":"d7079fd9f393e10798a10ad299a51d5120a549c4ac48bec74590bab9fbce49f0"} Dec 01 16:21:45 crc kubenswrapper[4810]: I1201 16:21:45.906130 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-db946669-jbbqh"] Dec 01 16:21:45 crc kubenswrapper[4810]: W1201 16:21:45.919578 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b1083d7_8870_4fbb_b4aa_a25dfaef9b14.slice/crio-777c68802b199c819d0d6bda821e6783163f8673d65a49ea49d5135bd860eb98 WatchSource:0}: Error finding container 777c68802b199c819d0d6bda821e6783163f8673d65a49ea49d5135bd860eb98: Status 404 returned error can't find the container with id 777c68802b199c819d0d6bda821e6783163f8673d65a49ea49d5135bd860eb98 Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.004547 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7d4fd6fc6c-6wbqj"] Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.282051 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6b989ffc48-bsdpj" Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.385592 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-combined-ca-bundle\") pod \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\" (UID: \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\") " Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.385672 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-config-data\") pod \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\" (UID: \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\") " Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.385847 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qprvh\" (UniqueName: \"kubernetes.io/projected/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-kube-api-access-qprvh\") pod \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\" (UID: \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\") " Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.385938 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-config-data-custom\") pod \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\" (UID: \"7d8e7e3c-e906-4197-9631-ae27a3ed10b1\") " Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.400346 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7d8e7e3c-e906-4197-9631-ae27a3ed10b1" (UID: "7d8e7e3c-e906-4197-9631-ae27a3ed10b1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.402810 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-kube-api-access-qprvh" (OuterVolumeSpecName: "kube-api-access-qprvh") pod "7d8e7e3c-e906-4197-9631-ae27a3ed10b1" (UID: "7d8e7e3c-e906-4197-9631-ae27a3ed10b1"). InnerVolumeSpecName "kube-api-access-qprvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.490342 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qprvh\" (UniqueName: \"kubernetes.io/projected/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-kube-api-access-qprvh\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.490393 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.600611 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d8e7e3c-e906-4197-9631-ae27a3ed10b1" (UID: "7d8e7e3c-e906-4197-9631-ae27a3ed10b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.622045 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-config-data" (OuterVolumeSpecName: "config-data") pod "7d8e7e3c-e906-4197-9631-ae27a3ed10b1" (UID: "7d8e7e3c-e906-4197-9631-ae27a3ed10b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.695045 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.695080 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d8e7e3c-e906-4197-9631-ae27a3ed10b1-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.854446 4810 generic.go:334] "Generic (PLEG): container finished" podID="fc84fc7d-8fdc-405a-a45f-d80723049d83" containerID="25cd6056021ea13d819cc9134bfa5a9e2d72ad41a1bc0131f388c479cb5ebe56" exitCode=1 Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.854515 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" event={"ID":"fc84fc7d-8fdc-405a-a45f-d80723049d83","Type":"ContainerDied","Data":"25cd6056021ea13d819cc9134bfa5a9e2d72ad41a1bc0131f388c479cb5ebe56"} Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.854595 4810 scope.go:117] "RemoveContainer" containerID="b5311f1891a721050d98a01c298de6fc5649ab078169f2a68924b05175d268ed" Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.855183 4810 scope.go:117] "RemoveContainer" containerID="25cd6056021ea13d819cc9134bfa5a9e2d72ad41a1bc0131f388c479cb5ebe56" Dec 01 16:21:46 crc kubenswrapper[4810]: E1201 16:21:46.855553 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-74bcc475c7-2mx5s_openstack(fc84fc7d-8fdc-405a-a45f-d80723049d83)\"" pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" podUID="fc84fc7d-8fdc-405a-a45f-d80723049d83" Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.855930 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-db946669-jbbqh" event={"ID":"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14","Type":"ContainerStarted","Data":"8e430636bbab82225f20d75e92e958b8f852ab985bb86c7ac2706e589466d61a"} Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.855971 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-db946669-jbbqh" event={"ID":"0b1083d7-8870-4fbb-b4aa-a25dfaef9b14","Type":"ContainerStarted","Data":"777c68802b199c819d0d6bda821e6783163f8673d65a49ea49d5135bd860eb98"} Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.856004 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.858202 4810 generic.go:334] "Generic (PLEG): container finished" podID="3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a" containerID="48dcee92d39418ef3ffbdb040d9d70a3912dab02ce1ae9bec09174f3f098d8a9" exitCode=1 Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.858234 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7c9cc4f87-5xtr9" event={"ID":"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a","Type":"ContainerDied","Data":"48dcee92d39418ef3ffbdb040d9d70a3912dab02ce1ae9bec09174f3f098d8a9"} Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.858859 4810 scope.go:117] "RemoveContainer" containerID="48dcee92d39418ef3ffbdb040d9d70a3912dab02ce1ae9bec09174f3f098d8a9" Dec 01 16:21:46 crc kubenswrapper[4810]: E1201 16:21:46.859132 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-7c9cc4f87-5xtr9_openstack(3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a)\"" pod="openstack/heat-api-7c9cc4f87-5xtr9" podUID="3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a" Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.859690 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" event={"ID":"24dcd733-d9bb-4def-9325-f1dd2611d1f1","Type":"ContainerStarted","Data":"95878d3d3b5a78c08cbd30665b5441f3c7ed372a7ed0ac4460607b10334fd8b2"} Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.859718 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" event={"ID":"24dcd733-d9bb-4def-9325-f1dd2611d1f1","Type":"ContainerStarted","Data":"f811b7227b55e79c2e73ae343a14936442cb777dfa12e1f99a1e09bf044cc675"} Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.859836 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.861827 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6b989ffc48-bsdpj" event={"ID":"7d8e7e3c-e906-4197-9631-ae27a3ed10b1","Type":"ContainerDied","Data":"de0ed53ecd16473dc15bf532b625bca2319484bc1f7896c9a3c50aaf2e50f4c2"} Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.861921 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6b989ffc48-bsdpj" Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.902391 4810 scope.go:117] "RemoveContainer" containerID="8c22b4488e6b3869e8d1b75a7a572db0cb6a67b204cf9027932b2b5c00e77620" Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.909920 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-db946669-jbbqh" podStartSLOduration=2.909905297 podStartE2EDuration="2.909905297s" podCreationTimestamp="2025-12-01 16:21:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:21:46.902914797 +0000 UTC m=+6472.666424400" watchObservedRunningTime="2025-12-01 16:21:46.909905297 +0000 UTC m=+6472.673414900" Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.965588 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" podStartSLOduration=2.965571518 podStartE2EDuration="2.965571518s" podCreationTimestamp="2025-12-01 16:21:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:21:46.947731763 +0000 UTC m=+6472.711241356" watchObservedRunningTime="2025-12-01 16:21:46.965571518 +0000 UTC m=+6472.729081121" Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.978062 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6b989ffc48-bsdpj"] Dec 01 16:21:46 crc kubenswrapper[4810]: I1201 16:21:46.988360 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-6b989ffc48-bsdpj"] Dec 01 16:21:47 crc kubenswrapper[4810]: I1201 16:21:47.008539 4810 scope.go:117] "RemoveContainer" containerID="eafae05c44b1790037d1812792777a021056624f4de00bba02e1b0999e67b168" Dec 01 16:21:47 crc kubenswrapper[4810]: I1201 16:21:47.871886 4810 scope.go:117] "RemoveContainer" containerID="25cd6056021ea13d819cc9134bfa5a9e2d72ad41a1bc0131f388c479cb5ebe56" Dec 01 16:21:47 crc kubenswrapper[4810]: E1201 16:21:47.872221 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-74bcc475c7-2mx5s_openstack(fc84fc7d-8fdc-405a-a45f-d80723049d83)\"" pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" podUID="fc84fc7d-8fdc-405a-a45f-d80723049d83" Dec 01 16:21:47 crc kubenswrapper[4810]: I1201 16:21:47.876260 4810 scope.go:117] "RemoveContainer" containerID="48dcee92d39418ef3ffbdb040d9d70a3912dab02ce1ae9bec09174f3f098d8a9" Dec 01 16:21:47 crc kubenswrapper[4810]: E1201 16:21:47.876561 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-7c9cc4f87-5xtr9_openstack(3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a)\"" pod="openstack/heat-api-7c9cc4f87-5xtr9" podUID="3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a" Dec 01 16:21:48 crc kubenswrapper[4810]: I1201 16:21:48.501854 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d8e7e3c-e906-4197-9631-ae27a3ed10b1" path="/var/lib/kubelet/pods/7d8e7e3c-e906-4197-9631-ae27a3ed10b1/volumes" Dec 01 16:21:49 crc kubenswrapper[4810]: I1201 16:21:49.075449 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-7c9cc4f87-5xtr9" Dec 01 16:21:49 crc kubenswrapper[4810]: I1201 16:21:49.075523 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-7c9cc4f87-5xtr9" Dec 01 16:21:49 crc kubenswrapper[4810]: I1201 16:21:49.076256 4810 scope.go:117] "RemoveContainer" containerID="48dcee92d39418ef3ffbdb040d9d70a3912dab02ce1ae9bec09174f3f098d8a9" Dec 01 16:21:49 crc kubenswrapper[4810]: E1201 16:21:49.076581 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-7c9cc4f87-5xtr9_openstack(3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a)\"" pod="openstack/heat-api-7c9cc4f87-5xtr9" podUID="3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a" Dec 01 16:21:49 crc kubenswrapper[4810]: I1201 16:21:49.134357 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" Dec 01 16:21:49 crc kubenswrapper[4810]: I1201 16:21:49.134404 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" Dec 01 16:21:49 crc kubenswrapper[4810]: I1201 16:21:49.135062 4810 scope.go:117] "RemoveContainer" containerID="25cd6056021ea13d819cc9134bfa5a9e2d72ad41a1bc0131f388c479cb5ebe56" Dec 01 16:21:49 crc kubenswrapper[4810]: E1201 16:21:49.135292 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-74bcc475c7-2mx5s_openstack(fc84fc7d-8fdc-405a-a45f-d80723049d83)\"" pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" podUID="fc84fc7d-8fdc-405a-a45f-d80723049d83" Dec 01 16:21:50 crc kubenswrapper[4810]: I1201 16:21:50.607218 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-5ffcf8767d-6597k" podUID="3d61bd8a-151d-474a-ac48-f332b48e519f" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.1.127:8000/healthcheck\": read tcp 10.217.0.2:59348->10.217.1.127:8000: read: connection reset by peer" Dec 01 16:21:50 crc kubenswrapper[4810]: I1201 16:21:50.912447 4810 generic.go:334] "Generic (PLEG): container finished" podID="3d61bd8a-151d-474a-ac48-f332b48e519f" containerID="a1fea62c437047870ed241cef5dd614ed1e0c60ac60756d969ca7e3f46c870cb" exitCode=0 Dec 01 16:21:50 crc kubenswrapper[4810]: I1201 16:21:50.912601 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5ffcf8767d-6597k" event={"ID":"3d61bd8a-151d-474a-ac48-f332b48e519f","Type":"ContainerDied","Data":"a1fea62c437047870ed241cef5dd614ed1e0c60ac60756d969ca7e3f46c870cb"} Dec 01 16:21:51 crc kubenswrapper[4810]: I1201 16:21:51.086754 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5ffcf8767d-6597k" Dec 01 16:21:51 crc kubenswrapper[4810]: I1201 16:21:51.201567 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d61bd8a-151d-474a-ac48-f332b48e519f-config-data-custom\") pod \"3d61bd8a-151d-474a-ac48-f332b48e519f\" (UID: \"3d61bd8a-151d-474a-ac48-f332b48e519f\") " Dec 01 16:21:51 crc kubenswrapper[4810]: I1201 16:21:51.201707 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbvd6\" (UniqueName: \"kubernetes.io/projected/3d61bd8a-151d-474a-ac48-f332b48e519f-kube-api-access-cbvd6\") pod \"3d61bd8a-151d-474a-ac48-f332b48e519f\" (UID: \"3d61bd8a-151d-474a-ac48-f332b48e519f\") " Dec 01 16:21:51 crc kubenswrapper[4810]: I1201 16:21:51.201828 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d61bd8a-151d-474a-ac48-f332b48e519f-config-data\") pod \"3d61bd8a-151d-474a-ac48-f332b48e519f\" (UID: \"3d61bd8a-151d-474a-ac48-f332b48e519f\") " Dec 01 16:21:51 crc kubenswrapper[4810]: I1201 16:21:51.201945 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d61bd8a-151d-474a-ac48-f332b48e519f-combined-ca-bundle\") pod \"3d61bd8a-151d-474a-ac48-f332b48e519f\" (UID: \"3d61bd8a-151d-474a-ac48-f332b48e519f\") " Dec 01 16:21:51 crc kubenswrapper[4810]: I1201 16:21:51.207329 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d61bd8a-151d-474a-ac48-f332b48e519f-kube-api-access-cbvd6" (OuterVolumeSpecName: "kube-api-access-cbvd6") pod "3d61bd8a-151d-474a-ac48-f332b48e519f" (UID: "3d61bd8a-151d-474a-ac48-f332b48e519f"). InnerVolumeSpecName "kube-api-access-cbvd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:21:51 crc kubenswrapper[4810]: I1201 16:21:51.210612 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d61bd8a-151d-474a-ac48-f332b48e519f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3d61bd8a-151d-474a-ac48-f332b48e519f" (UID: "3d61bd8a-151d-474a-ac48-f332b48e519f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:21:51 crc kubenswrapper[4810]: I1201 16:21:51.231993 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d61bd8a-151d-474a-ac48-f332b48e519f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d61bd8a-151d-474a-ac48-f332b48e519f" (UID: "3d61bd8a-151d-474a-ac48-f332b48e519f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:21:51 crc kubenswrapper[4810]: I1201 16:21:51.260231 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d61bd8a-151d-474a-ac48-f332b48e519f-config-data" (OuterVolumeSpecName: "config-data") pod "3d61bd8a-151d-474a-ac48-f332b48e519f" (UID: "3d61bd8a-151d-474a-ac48-f332b48e519f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:21:51 crc kubenswrapper[4810]: I1201 16:21:51.304850 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbvd6\" (UniqueName: \"kubernetes.io/projected/3d61bd8a-151d-474a-ac48-f332b48e519f-kube-api-access-cbvd6\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:51 crc kubenswrapper[4810]: I1201 16:21:51.305040 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d61bd8a-151d-474a-ac48-f332b48e519f-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:51 crc kubenswrapper[4810]: I1201 16:21:51.305117 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d61bd8a-151d-474a-ac48-f332b48e519f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:51 crc kubenswrapper[4810]: I1201 16:21:51.305183 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d61bd8a-151d-474a-ac48-f332b48e519f-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:51 crc kubenswrapper[4810]: I1201 16:21:51.928460 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5ffcf8767d-6597k" event={"ID":"3d61bd8a-151d-474a-ac48-f332b48e519f","Type":"ContainerDied","Data":"5972d65af7df91edf5610fb5105c0bcc27c1640cbf037ed832f466df26885da0"} Dec 01 16:21:51 crc kubenswrapper[4810]: I1201 16:21:51.928833 4810 scope.go:117] "RemoveContainer" containerID="a1fea62c437047870ed241cef5dd614ed1e0c60ac60756d969ca7e3f46c870cb" Dec 01 16:21:51 crc kubenswrapper[4810]: I1201 16:21:51.928535 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5ffcf8767d-6597k" Dec 01 16:21:51 crc kubenswrapper[4810]: I1201 16:21:51.968005 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-5ffcf8767d-6597k"] Dec 01 16:21:51 crc kubenswrapper[4810]: I1201 16:21:51.975922 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-5ffcf8767d-6597k"] Dec 01 16:21:52 crc kubenswrapper[4810]: I1201 16:21:52.155343 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-57bd9c5d44-hps7g" podUID="bfad2664-5a86-48b8-b3a5-cad4eec57408" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.118:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.118:8443: connect: connection refused" Dec 01 16:21:52 crc kubenswrapper[4810]: I1201 16:21:52.508079 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d61bd8a-151d-474a-ac48-f332b48e519f" path="/var/lib/kubelet/pods/3d61bd8a-151d-474a-ac48-f332b48e519f/volumes" Dec 01 16:21:54 crc kubenswrapper[4810]: I1201 16:21:54.101101 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-656cb86f76-jpbk2" Dec 01 16:21:54 crc kubenswrapper[4810]: I1201 16:21:54.164175 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-7ddd6f685-vw58v"] Dec 01 16:21:54 crc kubenswrapper[4810]: I1201 16:21:54.164422 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-7ddd6f685-vw58v" podUID="58744d74-c602-4e76-9f82-509d711591dc" containerName="heat-engine" containerID="cri-o://1b0c435a2665fd022d98359bc9ffae1a94fa1e75a34267d1703fbdabb8588d1c" gracePeriod=60 Dec 01 16:21:54 crc kubenswrapper[4810]: I1201 16:21:54.207753 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7ddd6f685-vw58v" Dec 01 16:21:56 crc kubenswrapper[4810]: I1201 16:21:56.670546 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-db946669-jbbqh" Dec 01 16:21:56 crc kubenswrapper[4810]: I1201 16:21:56.678968 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-7d4fd6fc6c-6wbqj" Dec 01 16:21:56 crc kubenswrapper[4810]: I1201 16:21:56.732027 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-7c9cc4f87-5xtr9"] Dec 01 16:21:56 crc kubenswrapper[4810]: I1201 16:21:56.787367 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-74bcc475c7-2mx5s"] Dec 01 16:21:56 crc kubenswrapper[4810]: E1201 16:21:56.985799 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1b0c435a2665fd022d98359bc9ffae1a94fa1e75a34267d1703fbdabb8588d1c" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 01 16:21:56 crc kubenswrapper[4810]: E1201 16:21:56.987048 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1b0c435a2665fd022d98359bc9ffae1a94fa1e75a34267d1703fbdabb8588d1c" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 01 16:21:56 crc kubenswrapper[4810]: E1201 16:21:56.988083 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1b0c435a2665fd022d98359bc9ffae1a94fa1e75a34267d1703fbdabb8588d1c" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 01 16:21:56 crc kubenswrapper[4810]: E1201 16:21:56.988119 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-7ddd6f685-vw58v" podUID="58744d74-c602-4e76-9f82-509d711591dc" containerName="heat-engine" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.237694 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7c9cc4f87-5xtr9" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.242405 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.256250 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-config-data-custom\") pod \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\" (UID: \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\") " Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.256324 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42gwg\" (UniqueName: \"kubernetes.io/projected/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-kube-api-access-42gwg\") pod \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\" (UID: \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\") " Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.262977 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-kube-api-access-42gwg" (OuterVolumeSpecName: "kube-api-access-42gwg") pod "3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a" (UID: "3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a"). InnerVolumeSpecName "kube-api-access-42gwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.264865 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a" (UID: "3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.265007 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42gwg\" (UniqueName: \"kubernetes.io/projected/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-kube-api-access-42gwg\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.365762 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-combined-ca-bundle\") pod \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\" (UID: \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\") " Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.365863 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc84fc7d-8fdc-405a-a45f-d80723049d83-config-data\") pod \"fc84fc7d-8fdc-405a-a45f-d80723049d83\" (UID: \"fc84fc7d-8fdc-405a-a45f-d80723049d83\") " Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.366017 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc84fc7d-8fdc-405a-a45f-d80723049d83-combined-ca-bundle\") pod \"fc84fc7d-8fdc-405a-a45f-d80723049d83\" (UID: \"fc84fc7d-8fdc-405a-a45f-d80723049d83\") " Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.366146 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc84fc7d-8fdc-405a-a45f-d80723049d83-config-data-custom\") pod \"fc84fc7d-8fdc-405a-a45f-d80723049d83\" (UID: \"fc84fc7d-8fdc-405a-a45f-d80723049d83\") " Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.366195 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ldlp\" (UniqueName: \"kubernetes.io/projected/fc84fc7d-8fdc-405a-a45f-d80723049d83-kube-api-access-7ldlp\") pod \"fc84fc7d-8fdc-405a-a45f-d80723049d83\" (UID: \"fc84fc7d-8fdc-405a-a45f-d80723049d83\") " Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.366267 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-config-data\") pod \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\" (UID: \"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a\") " Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.366931 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.370199 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc84fc7d-8fdc-405a-a45f-d80723049d83-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fc84fc7d-8fdc-405a-a45f-d80723049d83" (UID: "fc84fc7d-8fdc-405a-a45f-d80723049d83"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.370275 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc84fc7d-8fdc-405a-a45f-d80723049d83-kube-api-access-7ldlp" (OuterVolumeSpecName: "kube-api-access-7ldlp") pod "fc84fc7d-8fdc-405a-a45f-d80723049d83" (UID: "fc84fc7d-8fdc-405a-a45f-d80723049d83"). InnerVolumeSpecName "kube-api-access-7ldlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.399880 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc84fc7d-8fdc-405a-a45f-d80723049d83-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc84fc7d-8fdc-405a-a45f-d80723049d83" (UID: "fc84fc7d-8fdc-405a-a45f-d80723049d83"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.416678 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a" (UID: "3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.421963 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc84fc7d-8fdc-405a-a45f-d80723049d83-config-data" (OuterVolumeSpecName: "config-data") pod "fc84fc7d-8fdc-405a-a45f-d80723049d83" (UID: "fc84fc7d-8fdc-405a-a45f-d80723049d83"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.422738 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-config-data" (OuterVolumeSpecName: "config-data") pod "3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a" (UID: "3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.468596 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc84fc7d-8fdc-405a-a45f-d80723049d83-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.468630 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ldlp\" (UniqueName: \"kubernetes.io/projected/fc84fc7d-8fdc-405a-a45f-d80723049d83-kube-api-access-7ldlp\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.468641 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.468649 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.468659 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc84fc7d-8fdc-405a-a45f-d80723049d83-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.468667 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc84fc7d-8fdc-405a-a45f-d80723049d83-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.992836 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" event={"ID":"fc84fc7d-8fdc-405a-a45f-d80723049d83","Type":"ContainerDied","Data":"5eab2870035785206105d2070c9117adc4fe15ab23cce598067c9125fc867f90"} Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.992887 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-74bcc475c7-2mx5s" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.992906 4810 scope.go:117] "RemoveContainer" containerID="25cd6056021ea13d819cc9134bfa5a9e2d72ad41a1bc0131f388c479cb5ebe56" Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.995758 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7c9cc4f87-5xtr9" event={"ID":"3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a","Type":"ContainerDied","Data":"b947963ac56b211ca5e6035107ed1c0c8c00d25c79ecc1cd313bbaf3c73d2957"} Dec 01 16:21:57 crc kubenswrapper[4810]: I1201 16:21:57.995853 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7c9cc4f87-5xtr9" Dec 01 16:21:58 crc kubenswrapper[4810]: I1201 16:21:58.041046 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-7c9cc4f87-5xtr9"] Dec 01 16:21:58 crc kubenswrapper[4810]: I1201 16:21:58.042399 4810 scope.go:117] "RemoveContainer" containerID="48dcee92d39418ef3ffbdb040d9d70a3912dab02ce1ae9bec09174f3f098d8a9" Dec 01 16:21:58 crc kubenswrapper[4810]: I1201 16:21:58.052487 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-7c9cc4f87-5xtr9"] Dec 01 16:21:58 crc kubenswrapper[4810]: I1201 16:21:58.069572 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-74bcc475c7-2mx5s"] Dec 01 16:21:58 crc kubenswrapper[4810]: I1201 16:21:58.078355 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-74bcc475c7-2mx5s"] Dec 01 16:21:58 crc kubenswrapper[4810]: E1201 16:21:58.166102 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c60ab7f_e4c6_4ab5_ae10_03fdbeacda7a.slice/crio-b947963ac56b211ca5e6035107ed1c0c8c00d25c79ecc1cd313bbaf3c73d2957\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc84fc7d_8fdc_405a_a45f_d80723049d83.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc84fc7d_8fdc_405a_a45f_d80723049d83.slice/crio-5eab2870035785206105d2070c9117adc4fe15ab23cce598067c9125fc867f90\": RecentStats: unable to find data in memory cache]" Dec 01 16:21:58 crc kubenswrapper[4810]: I1201 16:21:58.511952 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a" path="/var/lib/kubelet/pods/3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a/volumes" Dec 01 16:21:58 crc kubenswrapper[4810]: I1201 16:21:58.512523 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc84fc7d-8fdc-405a-a45f-d80723049d83" path="/var/lib/kubelet/pods/fc84fc7d-8fdc-405a-a45f-d80723049d83/volumes" Dec 01 16:22:02 crc kubenswrapper[4810]: I1201 16:22:02.156271 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-57bd9c5d44-hps7g" podUID="bfad2664-5a86-48b8-b3a5-cad4eec57408" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.118:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.118:8443: connect: connection refused" Dec 01 16:22:02 crc kubenswrapper[4810]: I1201 16:22:02.156925 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:22:02 crc kubenswrapper[4810]: I1201 16:22:02.972022 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:22:02 crc kubenswrapper[4810]: I1201 16:22:02.972465 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:22:03 crc kubenswrapper[4810]: I1201 16:22:03.062645 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-4cz4n"] Dec 01 16:22:03 crc kubenswrapper[4810]: I1201 16:22:03.086307 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-c3d8-account-create-update-zw6k8"] Dec 01 16:22:03 crc kubenswrapper[4810]: I1201 16:22:03.105254 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-c3d8-account-create-update-zw6k8"] Dec 01 16:22:03 crc kubenswrapper[4810]: I1201 16:22:03.116766 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-4cz4n"] Dec 01 16:22:04 crc kubenswrapper[4810]: I1201 16:22:04.509032 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32c0cc0a-d9d0-4874-9e5c-bd5890560011" path="/var/lib/kubelet/pods/32c0cc0a-d9d0-4874-9e5c-bd5890560011/volumes" Dec 01 16:22:04 crc kubenswrapper[4810]: I1201 16:22:04.511121 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aaa2816-bc3b-44e6-98a0-f16da83a36f8" path="/var/lib/kubelet/pods/9aaa2816-bc3b-44e6-98a0-f16da83a36f8/volumes" Dec 01 16:22:04 crc kubenswrapper[4810]: I1201 16:22:04.597119 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7ddd6f685-vw58v" Dec 01 16:22:04 crc kubenswrapper[4810]: I1201 16:22:04.646120 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qj2gj\" (UniqueName: \"kubernetes.io/projected/58744d74-c602-4e76-9f82-509d711591dc-kube-api-access-qj2gj\") pod \"58744d74-c602-4e76-9f82-509d711591dc\" (UID: \"58744d74-c602-4e76-9f82-509d711591dc\") " Dec 01 16:22:04 crc kubenswrapper[4810]: I1201 16:22:04.646170 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58744d74-c602-4e76-9f82-509d711591dc-combined-ca-bundle\") pod \"58744d74-c602-4e76-9f82-509d711591dc\" (UID: \"58744d74-c602-4e76-9f82-509d711591dc\") " Dec 01 16:22:04 crc kubenswrapper[4810]: I1201 16:22:04.646218 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58744d74-c602-4e76-9f82-509d711591dc-config-data-custom\") pod \"58744d74-c602-4e76-9f82-509d711591dc\" (UID: \"58744d74-c602-4e76-9f82-509d711591dc\") " Dec 01 16:22:04 crc kubenswrapper[4810]: I1201 16:22:04.646250 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58744d74-c602-4e76-9f82-509d711591dc-config-data\") pod \"58744d74-c602-4e76-9f82-509d711591dc\" (UID: \"58744d74-c602-4e76-9f82-509d711591dc\") " Dec 01 16:22:04 crc kubenswrapper[4810]: I1201 16:22:04.654787 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58744d74-c602-4e76-9f82-509d711591dc-kube-api-access-qj2gj" (OuterVolumeSpecName: "kube-api-access-qj2gj") pod "58744d74-c602-4e76-9f82-509d711591dc" (UID: "58744d74-c602-4e76-9f82-509d711591dc"). InnerVolumeSpecName "kube-api-access-qj2gj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:22:04 crc kubenswrapper[4810]: I1201 16:22:04.654805 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58744d74-c602-4e76-9f82-509d711591dc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "58744d74-c602-4e76-9f82-509d711591dc" (UID: "58744d74-c602-4e76-9f82-509d711591dc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:22:04 crc kubenswrapper[4810]: I1201 16:22:04.683448 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58744d74-c602-4e76-9f82-509d711591dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58744d74-c602-4e76-9f82-509d711591dc" (UID: "58744d74-c602-4e76-9f82-509d711591dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:22:04 crc kubenswrapper[4810]: I1201 16:22:04.705529 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58744d74-c602-4e76-9f82-509d711591dc-config-data" (OuterVolumeSpecName: "config-data") pod "58744d74-c602-4e76-9f82-509d711591dc" (UID: "58744d74-c602-4e76-9f82-509d711591dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:22:04 crc kubenswrapper[4810]: I1201 16:22:04.748639 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qj2gj\" (UniqueName: \"kubernetes.io/projected/58744d74-c602-4e76-9f82-509d711591dc-kube-api-access-qj2gj\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:04 crc kubenswrapper[4810]: I1201 16:22:04.748672 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58744d74-c602-4e76-9f82-509d711591dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:04 crc kubenswrapper[4810]: I1201 16:22:04.748682 4810 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58744d74-c602-4e76-9f82-509d711591dc-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:04 crc kubenswrapper[4810]: I1201 16:22:04.748692 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58744d74-c602-4e76-9f82-509d711591dc-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:05 crc kubenswrapper[4810]: I1201 16:22:05.074331 4810 generic.go:334] "Generic (PLEG): container finished" podID="58744d74-c602-4e76-9f82-509d711591dc" containerID="1b0c435a2665fd022d98359bc9ffae1a94fa1e75a34267d1703fbdabb8588d1c" exitCode=0 Dec 01 16:22:05 crc kubenswrapper[4810]: I1201 16:22:05.074387 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7ddd6f685-vw58v" Dec 01 16:22:05 crc kubenswrapper[4810]: I1201 16:22:05.074388 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7ddd6f685-vw58v" event={"ID":"58744d74-c602-4e76-9f82-509d711591dc","Type":"ContainerDied","Data":"1b0c435a2665fd022d98359bc9ffae1a94fa1e75a34267d1703fbdabb8588d1c"} Dec 01 16:22:05 crc kubenswrapper[4810]: I1201 16:22:05.074454 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7ddd6f685-vw58v" event={"ID":"58744d74-c602-4e76-9f82-509d711591dc","Type":"ContainerDied","Data":"7e3f7b5375638281b5a791838a4a769a034de5b76f353ccaa48bd7c29e8a3b14"} Dec 01 16:22:05 crc kubenswrapper[4810]: I1201 16:22:05.074514 4810 scope.go:117] "RemoveContainer" containerID="1b0c435a2665fd022d98359bc9ffae1a94fa1e75a34267d1703fbdabb8588d1c" Dec 01 16:22:05 crc kubenswrapper[4810]: I1201 16:22:05.101507 4810 scope.go:117] "RemoveContainer" containerID="1b0c435a2665fd022d98359bc9ffae1a94fa1e75a34267d1703fbdabb8588d1c" Dec 01 16:22:05 crc kubenswrapper[4810]: E1201 16:22:05.102135 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b0c435a2665fd022d98359bc9ffae1a94fa1e75a34267d1703fbdabb8588d1c\": container with ID starting with 1b0c435a2665fd022d98359bc9ffae1a94fa1e75a34267d1703fbdabb8588d1c not found: ID does not exist" containerID="1b0c435a2665fd022d98359bc9ffae1a94fa1e75a34267d1703fbdabb8588d1c" Dec 01 16:22:05 crc kubenswrapper[4810]: I1201 16:22:05.102170 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b0c435a2665fd022d98359bc9ffae1a94fa1e75a34267d1703fbdabb8588d1c"} err="failed to get container status \"1b0c435a2665fd022d98359bc9ffae1a94fa1e75a34267d1703fbdabb8588d1c\": rpc error: code = NotFound desc = could not find container \"1b0c435a2665fd022d98359bc9ffae1a94fa1e75a34267d1703fbdabb8588d1c\": container with ID starting with 1b0c435a2665fd022d98359bc9ffae1a94fa1e75a34267d1703fbdabb8588d1c not found: ID does not exist" Dec 01 16:22:05 crc kubenswrapper[4810]: I1201 16:22:05.118663 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-7ddd6f685-vw58v"] Dec 01 16:22:05 crc kubenswrapper[4810]: I1201 16:22:05.127670 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-7ddd6f685-vw58v"] Dec 01 16:22:06 crc kubenswrapper[4810]: I1201 16:22:06.505066 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58744d74-c602-4e76-9f82-509d711591dc" path="/var/lib/kubelet/pods/58744d74-c602-4e76-9f82-509d711591dc/volumes" Dec 01 16:22:11 crc kubenswrapper[4810]: I1201 16:22:11.038688 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-45dk8"] Dec 01 16:22:11 crc kubenswrapper[4810]: I1201 16:22:11.054269 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-45dk8"] Dec 01 16:22:11 crc kubenswrapper[4810]: I1201 16:22:11.759215 4810 scope.go:117] "RemoveContainer" containerID="946060ab18c96a9ed5cc43bf6050a5344505fca618bd8e710c6b6ca16190677b" Dec 01 16:22:11 crc kubenswrapper[4810]: I1201 16:22:11.800593 4810 scope.go:117] "RemoveContainer" containerID="6f6b96a27c5e6f72fafb31afd0e93a344d4f49c098de4539403d56ca5eafcf66" Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.150077 4810 generic.go:334] "Generic (PLEG): container finished" podID="bfad2664-5a86-48b8-b3a5-cad4eec57408" containerID="ff427cf2a241a76a12a13f31fbfd44c3fab7d63ef78a625e7e10a45cd6d0ddf2" exitCode=137 Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.150509 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57bd9c5d44-hps7g" event={"ID":"bfad2664-5a86-48b8-b3a5-cad4eec57408","Type":"ContainerDied","Data":"ff427cf2a241a76a12a13f31fbfd44c3fab7d63ef78a625e7e10a45cd6d0ddf2"} Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.156043 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-57bd9c5d44-hps7g" podUID="bfad2664-5a86-48b8-b3a5-cad4eec57408" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.118:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.118:8443: connect: connection refused" Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.462918 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.507750 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eac202a1-bd74-43cf-841c-9b63602c1c24" path="/var/lib/kubelet/pods/eac202a1-bd74-43cf-841c-9b63602c1c24/volumes" Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.543448 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfad2664-5a86-48b8-b3a5-cad4eec57408-horizon-tls-certs\") pod \"bfad2664-5a86-48b8-b3a5-cad4eec57408\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.543615 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bfad2664-5a86-48b8-b3a5-cad4eec57408-config-data\") pod \"bfad2664-5a86-48b8-b3a5-cad4eec57408\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.543668 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bfad2664-5a86-48b8-b3a5-cad4eec57408-horizon-secret-key\") pod \"bfad2664-5a86-48b8-b3a5-cad4eec57408\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.543768 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfad2664-5a86-48b8-b3a5-cad4eec57408-combined-ca-bundle\") pod \"bfad2664-5a86-48b8-b3a5-cad4eec57408\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.543882 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjdpk\" (UniqueName: \"kubernetes.io/projected/bfad2664-5a86-48b8-b3a5-cad4eec57408-kube-api-access-pjdpk\") pod \"bfad2664-5a86-48b8-b3a5-cad4eec57408\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.543983 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfad2664-5a86-48b8-b3a5-cad4eec57408-logs\") pod \"bfad2664-5a86-48b8-b3a5-cad4eec57408\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.544041 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bfad2664-5a86-48b8-b3a5-cad4eec57408-scripts\") pod \"bfad2664-5a86-48b8-b3a5-cad4eec57408\" (UID: \"bfad2664-5a86-48b8-b3a5-cad4eec57408\") " Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.544768 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfad2664-5a86-48b8-b3a5-cad4eec57408-logs" (OuterVolumeSpecName: "logs") pod "bfad2664-5a86-48b8-b3a5-cad4eec57408" (UID: "bfad2664-5a86-48b8-b3a5-cad4eec57408"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.549726 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfad2664-5a86-48b8-b3a5-cad4eec57408-kube-api-access-pjdpk" (OuterVolumeSpecName: "kube-api-access-pjdpk") pod "bfad2664-5a86-48b8-b3a5-cad4eec57408" (UID: "bfad2664-5a86-48b8-b3a5-cad4eec57408"). InnerVolumeSpecName "kube-api-access-pjdpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.549726 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfad2664-5a86-48b8-b3a5-cad4eec57408-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "bfad2664-5a86-48b8-b3a5-cad4eec57408" (UID: "bfad2664-5a86-48b8-b3a5-cad4eec57408"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.570687 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfad2664-5a86-48b8-b3a5-cad4eec57408-config-data" (OuterVolumeSpecName: "config-data") pod "bfad2664-5a86-48b8-b3a5-cad4eec57408" (UID: "bfad2664-5a86-48b8-b3a5-cad4eec57408"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.573268 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfad2664-5a86-48b8-b3a5-cad4eec57408-scripts" (OuterVolumeSpecName: "scripts") pod "bfad2664-5a86-48b8-b3a5-cad4eec57408" (UID: "bfad2664-5a86-48b8-b3a5-cad4eec57408"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.585275 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfad2664-5a86-48b8-b3a5-cad4eec57408-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfad2664-5a86-48b8-b3a5-cad4eec57408" (UID: "bfad2664-5a86-48b8-b3a5-cad4eec57408"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.624788 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfad2664-5a86-48b8-b3a5-cad4eec57408-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "bfad2664-5a86-48b8-b3a5-cad4eec57408" (UID: "bfad2664-5a86-48b8-b3a5-cad4eec57408"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.646552 4810 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfad2664-5a86-48b8-b3a5-cad4eec57408-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.646755 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bfad2664-5a86-48b8-b3a5-cad4eec57408-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.646814 4810 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bfad2664-5a86-48b8-b3a5-cad4eec57408-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.646868 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfad2664-5a86-48b8-b3a5-cad4eec57408-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.646920 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjdpk\" (UniqueName: \"kubernetes.io/projected/bfad2664-5a86-48b8-b3a5-cad4eec57408-kube-api-access-pjdpk\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.646978 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfad2664-5a86-48b8-b3a5-cad4eec57408-logs\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:12 crc kubenswrapper[4810]: I1201 16:22:12.647083 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bfad2664-5a86-48b8-b3a5-cad4eec57408-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:13 crc kubenswrapper[4810]: I1201 16:22:13.166552 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57bd9c5d44-hps7g" event={"ID":"bfad2664-5a86-48b8-b3a5-cad4eec57408","Type":"ContainerDied","Data":"c76190228c0e362d5d4618bef66941165d9b56ef8ad9459f6d3ff456e2e51536"} Dec 01 16:22:13 crc kubenswrapper[4810]: I1201 16:22:13.167679 4810 scope.go:117] "RemoveContainer" containerID="d7079fd9f393e10798a10ad299a51d5120a549c4ac48bec74590bab9fbce49f0" Dec 01 16:22:13 crc kubenswrapper[4810]: I1201 16:22:13.166660 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57bd9c5d44-hps7g" Dec 01 16:22:13 crc kubenswrapper[4810]: I1201 16:22:13.224393 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-57bd9c5d44-hps7g"] Dec 01 16:22:13 crc kubenswrapper[4810]: I1201 16:22:13.233800 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-57bd9c5d44-hps7g"] Dec 01 16:22:13 crc kubenswrapper[4810]: I1201 16:22:13.404597 4810 scope.go:117] "RemoveContainer" containerID="ff427cf2a241a76a12a13f31fbfd44c3fab7d63ef78a625e7e10a45cd6d0ddf2" Dec 01 16:22:14 crc kubenswrapper[4810]: I1201 16:22:14.505855 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfad2664-5a86-48b8-b3a5-cad4eec57408" path="/var/lib/kubelet/pods/bfad2664-5a86-48b8-b3a5-cad4eec57408/volumes" Dec 01 16:22:20 crc kubenswrapper[4810]: E1201 16:22:20.397066 4810 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.182:43288->38.102.83.182:41089: write tcp 38.102.83.182:43288->38.102.83.182:41089: write: broken pipe Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.809570 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wqtwz"] Dec 01 16:22:29 crc kubenswrapper[4810]: E1201 16:22:29.810920 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58744d74-c602-4e76-9f82-509d711591dc" containerName="heat-engine" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.810936 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="58744d74-c602-4e76-9f82-509d711591dc" containerName="heat-engine" Dec 01 16:22:29 crc kubenswrapper[4810]: E1201 16:22:29.810955 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d8e7e3c-e906-4197-9631-ae27a3ed10b1" containerName="heat-api" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.810963 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d8e7e3c-e906-4197-9631-ae27a3ed10b1" containerName="heat-api" Dec 01 16:22:29 crc kubenswrapper[4810]: E1201 16:22:29.810976 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc84fc7d-8fdc-405a-a45f-d80723049d83" containerName="heat-cfnapi" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.810984 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc84fc7d-8fdc-405a-a45f-d80723049d83" containerName="heat-cfnapi" Dec 01 16:22:29 crc kubenswrapper[4810]: E1201 16:22:29.811005 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfad2664-5a86-48b8-b3a5-cad4eec57408" containerName="horizon" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.811013 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfad2664-5a86-48b8-b3a5-cad4eec57408" containerName="horizon" Dec 01 16:22:29 crc kubenswrapper[4810]: E1201 16:22:29.811028 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d61bd8a-151d-474a-ac48-f332b48e519f" containerName="heat-cfnapi" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.811036 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d61bd8a-151d-474a-ac48-f332b48e519f" containerName="heat-cfnapi" Dec 01 16:22:29 crc kubenswrapper[4810]: E1201 16:22:29.811053 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a" containerName="heat-api" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.811061 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a" containerName="heat-api" Dec 01 16:22:29 crc kubenswrapper[4810]: E1201 16:22:29.811074 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc84fc7d-8fdc-405a-a45f-d80723049d83" containerName="heat-cfnapi" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.811081 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc84fc7d-8fdc-405a-a45f-d80723049d83" containerName="heat-cfnapi" Dec 01 16:22:29 crc kubenswrapper[4810]: E1201 16:22:29.811098 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfad2664-5a86-48b8-b3a5-cad4eec57408" containerName="horizon-log" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.811110 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfad2664-5a86-48b8-b3a5-cad4eec57408" containerName="horizon-log" Dec 01 16:22:29 crc kubenswrapper[4810]: E1201 16:22:29.811122 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a" containerName="heat-api" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.811129 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a" containerName="heat-api" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.811373 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfad2664-5a86-48b8-b3a5-cad4eec57408" containerName="horizon" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.811394 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfad2664-5a86-48b8-b3a5-cad4eec57408" containerName="horizon-log" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.811407 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc84fc7d-8fdc-405a-a45f-d80723049d83" containerName="heat-cfnapi" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.811419 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d8e7e3c-e906-4197-9631-ae27a3ed10b1" containerName="heat-api" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.811435 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a" containerName="heat-api" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.811449 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d61bd8a-151d-474a-ac48-f332b48e519f" containerName="heat-cfnapi" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.811457 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="58744d74-c602-4e76-9f82-509d711591dc" containerName="heat-engine" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.811933 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc84fc7d-8fdc-405a-a45f-d80723049d83" containerName="heat-cfnapi" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.811962 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c60ab7f-e4c6-4ab5-ae10-03fdbeacda7a" containerName="heat-api" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.813363 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wqtwz" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.841829 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wqtwz"] Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.943579 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51e9683a-6e27-449b-a237-4013a84619d2-catalog-content\") pod \"redhat-marketplace-wqtwz\" (UID: \"51e9683a-6e27-449b-a237-4013a84619d2\") " pod="openshift-marketplace/redhat-marketplace-wqtwz" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.943692 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51e9683a-6e27-449b-a237-4013a84619d2-utilities\") pod \"redhat-marketplace-wqtwz\" (UID: \"51e9683a-6e27-449b-a237-4013a84619d2\") " pod="openshift-marketplace/redhat-marketplace-wqtwz" Dec 01 16:22:29 crc kubenswrapper[4810]: I1201 16:22:29.943772 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph487\" (UniqueName: \"kubernetes.io/projected/51e9683a-6e27-449b-a237-4013a84619d2-kube-api-access-ph487\") pod \"redhat-marketplace-wqtwz\" (UID: \"51e9683a-6e27-449b-a237-4013a84619d2\") " pod="openshift-marketplace/redhat-marketplace-wqtwz" Dec 01 16:22:30 crc kubenswrapper[4810]: I1201 16:22:30.045362 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51e9683a-6e27-449b-a237-4013a84619d2-catalog-content\") pod \"redhat-marketplace-wqtwz\" (UID: \"51e9683a-6e27-449b-a237-4013a84619d2\") " pod="openshift-marketplace/redhat-marketplace-wqtwz" Dec 01 16:22:30 crc kubenswrapper[4810]: I1201 16:22:30.045502 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51e9683a-6e27-449b-a237-4013a84619d2-utilities\") pod \"redhat-marketplace-wqtwz\" (UID: \"51e9683a-6e27-449b-a237-4013a84619d2\") " pod="openshift-marketplace/redhat-marketplace-wqtwz" Dec 01 16:22:30 crc kubenswrapper[4810]: I1201 16:22:30.045551 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph487\" (UniqueName: \"kubernetes.io/projected/51e9683a-6e27-449b-a237-4013a84619d2-kube-api-access-ph487\") pod \"redhat-marketplace-wqtwz\" (UID: \"51e9683a-6e27-449b-a237-4013a84619d2\") " pod="openshift-marketplace/redhat-marketplace-wqtwz" Dec 01 16:22:30 crc kubenswrapper[4810]: I1201 16:22:30.045921 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51e9683a-6e27-449b-a237-4013a84619d2-catalog-content\") pod \"redhat-marketplace-wqtwz\" (UID: \"51e9683a-6e27-449b-a237-4013a84619d2\") " pod="openshift-marketplace/redhat-marketplace-wqtwz" Dec 01 16:22:30 crc kubenswrapper[4810]: I1201 16:22:30.045976 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51e9683a-6e27-449b-a237-4013a84619d2-utilities\") pod \"redhat-marketplace-wqtwz\" (UID: \"51e9683a-6e27-449b-a237-4013a84619d2\") " pod="openshift-marketplace/redhat-marketplace-wqtwz" Dec 01 16:22:30 crc kubenswrapper[4810]: I1201 16:22:30.068285 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph487\" (UniqueName: \"kubernetes.io/projected/51e9683a-6e27-449b-a237-4013a84619d2-kube-api-access-ph487\") pod \"redhat-marketplace-wqtwz\" (UID: \"51e9683a-6e27-449b-a237-4013a84619d2\") " pod="openshift-marketplace/redhat-marketplace-wqtwz" Dec 01 16:22:30 crc kubenswrapper[4810]: I1201 16:22:30.141942 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wqtwz" Dec 01 16:22:30 crc kubenswrapper[4810]: I1201 16:22:30.675148 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wqtwz"] Dec 01 16:22:30 crc kubenswrapper[4810]: I1201 16:22:30.834426 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x"] Dec 01 16:22:30 crc kubenswrapper[4810]: I1201 16:22:30.836445 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x" Dec 01 16:22:30 crc kubenswrapper[4810]: I1201 16:22:30.842149 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 16:22:30 crc kubenswrapper[4810]: I1201 16:22:30.845288 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x"] Dec 01 16:22:30 crc kubenswrapper[4810]: I1201 16:22:30.963509 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/135d31a5-5184-4906-8c27-9f8d645ea84c-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x\" (UID: \"135d31a5-5184-4906-8c27-9f8d645ea84c\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x" Dec 01 16:22:30 crc kubenswrapper[4810]: I1201 16:22:30.963598 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dcx9\" (UniqueName: \"kubernetes.io/projected/135d31a5-5184-4906-8c27-9f8d645ea84c-kube-api-access-9dcx9\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x\" (UID: \"135d31a5-5184-4906-8c27-9f8d645ea84c\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x" Dec 01 16:22:30 crc kubenswrapper[4810]: I1201 16:22:30.963691 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/135d31a5-5184-4906-8c27-9f8d645ea84c-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x\" (UID: \"135d31a5-5184-4906-8c27-9f8d645ea84c\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x" Dec 01 16:22:31 crc kubenswrapper[4810]: I1201 16:22:31.065635 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/135d31a5-5184-4906-8c27-9f8d645ea84c-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x\" (UID: \"135d31a5-5184-4906-8c27-9f8d645ea84c\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x" Dec 01 16:22:31 crc kubenswrapper[4810]: I1201 16:22:31.065726 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dcx9\" (UniqueName: \"kubernetes.io/projected/135d31a5-5184-4906-8c27-9f8d645ea84c-kube-api-access-9dcx9\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x\" (UID: \"135d31a5-5184-4906-8c27-9f8d645ea84c\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x" Dec 01 16:22:31 crc kubenswrapper[4810]: I1201 16:22:31.065797 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/135d31a5-5184-4906-8c27-9f8d645ea84c-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x\" (UID: \"135d31a5-5184-4906-8c27-9f8d645ea84c\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x" Dec 01 16:22:31 crc kubenswrapper[4810]: I1201 16:22:31.066257 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/135d31a5-5184-4906-8c27-9f8d645ea84c-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x\" (UID: \"135d31a5-5184-4906-8c27-9f8d645ea84c\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x" Dec 01 16:22:31 crc kubenswrapper[4810]: I1201 16:22:31.066286 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/135d31a5-5184-4906-8c27-9f8d645ea84c-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x\" (UID: \"135d31a5-5184-4906-8c27-9f8d645ea84c\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x" Dec 01 16:22:31 crc kubenswrapper[4810]: I1201 16:22:31.083392 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dcx9\" (UniqueName: \"kubernetes.io/projected/135d31a5-5184-4906-8c27-9f8d645ea84c-kube-api-access-9dcx9\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x\" (UID: \"135d31a5-5184-4906-8c27-9f8d645ea84c\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x" Dec 01 16:22:31 crc kubenswrapper[4810]: I1201 16:22:31.165657 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x" Dec 01 16:22:31 crc kubenswrapper[4810]: I1201 16:22:31.413200 4810 generic.go:334] "Generic (PLEG): container finished" podID="51e9683a-6e27-449b-a237-4013a84619d2" containerID="3a89404c12be55963fdff956e3e961b4604968645a0e0181052773f5324baf54" exitCode=0 Dec 01 16:22:31 crc kubenswrapper[4810]: I1201 16:22:31.413346 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqtwz" event={"ID":"51e9683a-6e27-449b-a237-4013a84619d2","Type":"ContainerDied","Data":"3a89404c12be55963fdff956e3e961b4604968645a0e0181052773f5324baf54"} Dec 01 16:22:31 crc kubenswrapper[4810]: I1201 16:22:31.413456 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqtwz" event={"ID":"51e9683a-6e27-449b-a237-4013a84619d2","Type":"ContainerStarted","Data":"4d9ef6401a992dd8db3d378edb84f471354f30cdfaa8737c578dfcc7a8495e14"} Dec 01 16:22:31 crc kubenswrapper[4810]: I1201 16:22:31.415899 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:22:31 crc kubenswrapper[4810]: I1201 16:22:31.635928 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x"] Dec 01 16:22:31 crc kubenswrapper[4810]: W1201 16:22:31.642681 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod135d31a5_5184_4906_8c27_9f8d645ea84c.slice/crio-e8688189d2d58b521bb1247c480e354ea249a296a86faa7b4583e1ca10e89c3e WatchSource:0}: Error finding container e8688189d2d58b521bb1247c480e354ea249a296a86faa7b4583e1ca10e89c3e: Status 404 returned error can't find the container with id e8688189d2d58b521bb1247c480e354ea249a296a86faa7b4583e1ca10e89c3e Dec 01 16:22:32 crc kubenswrapper[4810]: I1201 16:22:32.425590 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqtwz" event={"ID":"51e9683a-6e27-449b-a237-4013a84619d2","Type":"ContainerStarted","Data":"b6eedae0063b6b83ebfea64a42b821eb294f765d0564c7dbc497b12bf7c70088"} Dec 01 16:22:32 crc kubenswrapper[4810]: I1201 16:22:32.427457 4810 generic.go:334] "Generic (PLEG): container finished" podID="135d31a5-5184-4906-8c27-9f8d645ea84c" containerID="8b7637307cc7fd8bcd9aec7822289d411599b9f1360a80e1c601909486f21b3a" exitCode=0 Dec 01 16:22:32 crc kubenswrapper[4810]: I1201 16:22:32.427529 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x" event={"ID":"135d31a5-5184-4906-8c27-9f8d645ea84c","Type":"ContainerDied","Data":"8b7637307cc7fd8bcd9aec7822289d411599b9f1360a80e1c601909486f21b3a"} Dec 01 16:22:32 crc kubenswrapper[4810]: I1201 16:22:32.427559 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x" event={"ID":"135d31a5-5184-4906-8c27-9f8d645ea84c","Type":"ContainerStarted","Data":"e8688189d2d58b521bb1247c480e354ea249a296a86faa7b4583e1ca10e89c3e"} Dec 01 16:22:32 crc kubenswrapper[4810]: I1201 16:22:32.972120 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:22:32 crc kubenswrapper[4810]: I1201 16:22:32.972466 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:22:32 crc kubenswrapper[4810]: I1201 16:22:32.972518 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 16:22:32 crc kubenswrapper[4810]: I1201 16:22:32.973277 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"13b8d9a5912bf37c10017c37926041fb03894967f1288822928bc2736df5d8df"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:22:32 crc kubenswrapper[4810]: I1201 16:22:32.973336 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://13b8d9a5912bf37c10017c37926041fb03894967f1288822928bc2736df5d8df" gracePeriod=600 Dec 01 16:22:33 crc kubenswrapper[4810]: I1201 16:22:33.440516 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="13b8d9a5912bf37c10017c37926041fb03894967f1288822928bc2736df5d8df" exitCode=0 Dec 01 16:22:33 crc kubenswrapper[4810]: I1201 16:22:33.440577 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"13b8d9a5912bf37c10017c37926041fb03894967f1288822928bc2736df5d8df"} Dec 01 16:22:33 crc kubenswrapper[4810]: I1201 16:22:33.440992 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b"} Dec 01 16:22:33 crc kubenswrapper[4810]: I1201 16:22:33.441013 4810 scope.go:117] "RemoveContainer" containerID="1055b545264eb5e8398f9e24154c38e7bca34c337fd9468f5ac780a23023ce2d" Dec 01 16:22:33 crc kubenswrapper[4810]: I1201 16:22:33.444454 4810 generic.go:334] "Generic (PLEG): container finished" podID="51e9683a-6e27-449b-a237-4013a84619d2" containerID="b6eedae0063b6b83ebfea64a42b821eb294f765d0564c7dbc497b12bf7c70088" exitCode=0 Dec 01 16:22:33 crc kubenswrapper[4810]: I1201 16:22:33.444516 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqtwz" event={"ID":"51e9683a-6e27-449b-a237-4013a84619d2","Type":"ContainerDied","Data":"b6eedae0063b6b83ebfea64a42b821eb294f765d0564c7dbc497b12bf7c70088"} Dec 01 16:22:34 crc kubenswrapper[4810]: I1201 16:22:34.462136 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqtwz" event={"ID":"51e9683a-6e27-449b-a237-4013a84619d2","Type":"ContainerStarted","Data":"ae362e75ae4fa4b37c0ddfabb15f6eec852797b0d866ea4dbd323a259f49dcd6"} Dec 01 16:22:34 crc kubenswrapper[4810]: I1201 16:22:34.491502 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wqtwz" podStartSLOduration=2.706176916 podStartE2EDuration="5.491465138s" podCreationTimestamp="2025-12-01 16:22:29 +0000 UTC" firstStartedPulling="2025-12-01 16:22:31.415445644 +0000 UTC m=+6517.178955247" lastFinishedPulling="2025-12-01 16:22:34.200733866 +0000 UTC m=+6519.964243469" observedRunningTime="2025-12-01 16:22:34.489653229 +0000 UTC m=+6520.253162862" watchObservedRunningTime="2025-12-01 16:22:34.491465138 +0000 UTC m=+6520.254974741" Dec 01 16:22:35 crc kubenswrapper[4810]: I1201 16:22:35.471606 4810 generic.go:334] "Generic (PLEG): container finished" podID="135d31a5-5184-4906-8c27-9f8d645ea84c" containerID="588ce864cb8f5a892fe6ae253a0c94de859fb0ed1d97b087ca53a99d8bea0cdf" exitCode=0 Dec 01 16:22:35 crc kubenswrapper[4810]: I1201 16:22:35.471672 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x" event={"ID":"135d31a5-5184-4906-8c27-9f8d645ea84c","Type":"ContainerDied","Data":"588ce864cb8f5a892fe6ae253a0c94de859fb0ed1d97b087ca53a99d8bea0cdf"} Dec 01 16:22:36 crc kubenswrapper[4810]: I1201 16:22:36.486417 4810 generic.go:334] "Generic (PLEG): container finished" podID="135d31a5-5184-4906-8c27-9f8d645ea84c" containerID="f287a7fea863cb14d3b9f4956d502d6828e5f376b6beaf97f3399e59e13f6a4a" exitCode=0 Dec 01 16:22:36 crc kubenswrapper[4810]: I1201 16:22:36.486595 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x" event={"ID":"135d31a5-5184-4906-8c27-9f8d645ea84c","Type":"ContainerDied","Data":"f287a7fea863cb14d3b9f4956d502d6828e5f376b6beaf97f3399e59e13f6a4a"} Dec 01 16:22:37 crc kubenswrapper[4810]: I1201 16:22:37.818970 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x" Dec 01 16:22:37 crc kubenswrapper[4810]: I1201 16:22:37.903130 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/135d31a5-5184-4906-8c27-9f8d645ea84c-bundle\") pod \"135d31a5-5184-4906-8c27-9f8d645ea84c\" (UID: \"135d31a5-5184-4906-8c27-9f8d645ea84c\") " Dec 01 16:22:37 crc kubenswrapper[4810]: I1201 16:22:37.903191 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dcx9\" (UniqueName: \"kubernetes.io/projected/135d31a5-5184-4906-8c27-9f8d645ea84c-kube-api-access-9dcx9\") pod \"135d31a5-5184-4906-8c27-9f8d645ea84c\" (UID: \"135d31a5-5184-4906-8c27-9f8d645ea84c\") " Dec 01 16:22:37 crc kubenswrapper[4810]: I1201 16:22:37.903314 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/135d31a5-5184-4906-8c27-9f8d645ea84c-util\") pod \"135d31a5-5184-4906-8c27-9f8d645ea84c\" (UID: \"135d31a5-5184-4906-8c27-9f8d645ea84c\") " Dec 01 16:22:37 crc kubenswrapper[4810]: I1201 16:22:37.908286 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/135d31a5-5184-4906-8c27-9f8d645ea84c-bundle" (OuterVolumeSpecName: "bundle") pod "135d31a5-5184-4906-8c27-9f8d645ea84c" (UID: "135d31a5-5184-4906-8c27-9f8d645ea84c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:22:37 crc kubenswrapper[4810]: I1201 16:22:37.910758 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/135d31a5-5184-4906-8c27-9f8d645ea84c-util" (OuterVolumeSpecName: "util") pod "135d31a5-5184-4906-8c27-9f8d645ea84c" (UID: "135d31a5-5184-4906-8c27-9f8d645ea84c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:22:37 crc kubenswrapper[4810]: I1201 16:22:37.942842 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/135d31a5-5184-4906-8c27-9f8d645ea84c-kube-api-access-9dcx9" (OuterVolumeSpecName: "kube-api-access-9dcx9") pod "135d31a5-5184-4906-8c27-9f8d645ea84c" (UID: "135d31a5-5184-4906-8c27-9f8d645ea84c"). InnerVolumeSpecName "kube-api-access-9dcx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:22:38 crc kubenswrapper[4810]: I1201 16:22:38.014740 4810 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/135d31a5-5184-4906-8c27-9f8d645ea84c-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:38 crc kubenswrapper[4810]: I1201 16:22:38.014782 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dcx9\" (UniqueName: \"kubernetes.io/projected/135d31a5-5184-4906-8c27-9f8d645ea84c-kube-api-access-9dcx9\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:38 crc kubenswrapper[4810]: I1201 16:22:38.014794 4810 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/135d31a5-5184-4906-8c27-9f8d645ea84c-util\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:38 crc kubenswrapper[4810]: I1201 16:22:38.509927 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x" event={"ID":"135d31a5-5184-4906-8c27-9f8d645ea84c","Type":"ContainerDied","Data":"e8688189d2d58b521bb1247c480e354ea249a296a86faa7b4583e1ca10e89c3e"} Dec 01 16:22:38 crc kubenswrapper[4810]: I1201 16:22:38.509971 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8688189d2d58b521bb1247c480e354ea249a296a86faa7b4583e1ca10e89c3e" Dec 01 16:22:38 crc kubenswrapper[4810]: I1201 16:22:38.510018 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x" Dec 01 16:22:40 crc kubenswrapper[4810]: I1201 16:22:40.143090 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wqtwz" Dec 01 16:22:40 crc kubenswrapper[4810]: I1201 16:22:40.143579 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wqtwz" Dec 01 16:22:40 crc kubenswrapper[4810]: I1201 16:22:40.188199 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wqtwz" Dec 01 16:22:40 crc kubenswrapper[4810]: I1201 16:22:40.579803 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wqtwz" Dec 01 16:22:41 crc kubenswrapper[4810]: I1201 16:22:41.990182 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wqtwz"] Dec 01 16:22:42 crc kubenswrapper[4810]: I1201 16:22:42.546197 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wqtwz" podUID="51e9683a-6e27-449b-a237-4013a84619d2" containerName="registry-server" containerID="cri-o://ae362e75ae4fa4b37c0ddfabb15f6eec852797b0d866ea4dbd323a259f49dcd6" gracePeriod=2 Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.011887 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wqtwz" Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.119457 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51e9683a-6e27-449b-a237-4013a84619d2-catalog-content\") pod \"51e9683a-6e27-449b-a237-4013a84619d2\" (UID: \"51e9683a-6e27-449b-a237-4013a84619d2\") " Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.119538 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph487\" (UniqueName: \"kubernetes.io/projected/51e9683a-6e27-449b-a237-4013a84619d2-kube-api-access-ph487\") pod \"51e9683a-6e27-449b-a237-4013a84619d2\" (UID: \"51e9683a-6e27-449b-a237-4013a84619d2\") " Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.119654 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51e9683a-6e27-449b-a237-4013a84619d2-utilities\") pod \"51e9683a-6e27-449b-a237-4013a84619d2\" (UID: \"51e9683a-6e27-449b-a237-4013a84619d2\") " Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.120351 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51e9683a-6e27-449b-a237-4013a84619d2-utilities" (OuterVolumeSpecName: "utilities") pod "51e9683a-6e27-449b-a237-4013a84619d2" (UID: "51e9683a-6e27-449b-a237-4013a84619d2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.127255 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51e9683a-6e27-449b-a237-4013a84619d2-kube-api-access-ph487" (OuterVolumeSpecName: "kube-api-access-ph487") pod "51e9683a-6e27-449b-a237-4013a84619d2" (UID: "51e9683a-6e27-449b-a237-4013a84619d2"). InnerVolumeSpecName "kube-api-access-ph487". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.139609 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51e9683a-6e27-449b-a237-4013a84619d2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "51e9683a-6e27-449b-a237-4013a84619d2" (UID: "51e9683a-6e27-449b-a237-4013a84619d2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.221646 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51e9683a-6e27-449b-a237-4013a84619d2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.221688 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph487\" (UniqueName: \"kubernetes.io/projected/51e9683a-6e27-449b-a237-4013a84619d2-kube-api-access-ph487\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.221705 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51e9683a-6e27-449b-a237-4013a84619d2-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.557935 4810 generic.go:334] "Generic (PLEG): container finished" podID="51e9683a-6e27-449b-a237-4013a84619d2" containerID="ae362e75ae4fa4b37c0ddfabb15f6eec852797b0d866ea4dbd323a259f49dcd6" exitCode=0 Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.558003 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wqtwz" Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.558022 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqtwz" event={"ID":"51e9683a-6e27-449b-a237-4013a84619d2","Type":"ContainerDied","Data":"ae362e75ae4fa4b37c0ddfabb15f6eec852797b0d866ea4dbd323a259f49dcd6"} Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.558393 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqtwz" event={"ID":"51e9683a-6e27-449b-a237-4013a84619d2","Type":"ContainerDied","Data":"4d9ef6401a992dd8db3d378edb84f471354f30cdfaa8737c578dfcc7a8495e14"} Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.558420 4810 scope.go:117] "RemoveContainer" containerID="ae362e75ae4fa4b37c0ddfabb15f6eec852797b0d866ea4dbd323a259f49dcd6" Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.600634 4810 scope.go:117] "RemoveContainer" containerID="b6eedae0063b6b83ebfea64a42b821eb294f765d0564c7dbc497b12bf7c70088" Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.619539 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wqtwz"] Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.627498 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wqtwz"] Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.676179 4810 scope.go:117] "RemoveContainer" containerID="3a89404c12be55963fdff956e3e961b4604968645a0e0181052773f5324baf54" Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.717039 4810 scope.go:117] "RemoveContainer" containerID="ae362e75ae4fa4b37c0ddfabb15f6eec852797b0d866ea4dbd323a259f49dcd6" Dec 01 16:22:43 crc kubenswrapper[4810]: E1201 16:22:43.717523 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae362e75ae4fa4b37c0ddfabb15f6eec852797b0d866ea4dbd323a259f49dcd6\": container with ID starting with ae362e75ae4fa4b37c0ddfabb15f6eec852797b0d866ea4dbd323a259f49dcd6 not found: ID does not exist" containerID="ae362e75ae4fa4b37c0ddfabb15f6eec852797b0d866ea4dbd323a259f49dcd6" Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.717560 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae362e75ae4fa4b37c0ddfabb15f6eec852797b0d866ea4dbd323a259f49dcd6"} err="failed to get container status \"ae362e75ae4fa4b37c0ddfabb15f6eec852797b0d866ea4dbd323a259f49dcd6\": rpc error: code = NotFound desc = could not find container \"ae362e75ae4fa4b37c0ddfabb15f6eec852797b0d866ea4dbd323a259f49dcd6\": container with ID starting with ae362e75ae4fa4b37c0ddfabb15f6eec852797b0d866ea4dbd323a259f49dcd6 not found: ID does not exist" Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.717586 4810 scope.go:117] "RemoveContainer" containerID="b6eedae0063b6b83ebfea64a42b821eb294f765d0564c7dbc497b12bf7c70088" Dec 01 16:22:43 crc kubenswrapper[4810]: E1201 16:22:43.717787 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6eedae0063b6b83ebfea64a42b821eb294f765d0564c7dbc497b12bf7c70088\": container with ID starting with b6eedae0063b6b83ebfea64a42b821eb294f765d0564c7dbc497b12bf7c70088 not found: ID does not exist" containerID="b6eedae0063b6b83ebfea64a42b821eb294f765d0564c7dbc497b12bf7c70088" Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.717823 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6eedae0063b6b83ebfea64a42b821eb294f765d0564c7dbc497b12bf7c70088"} err="failed to get container status \"b6eedae0063b6b83ebfea64a42b821eb294f765d0564c7dbc497b12bf7c70088\": rpc error: code = NotFound desc = could not find container \"b6eedae0063b6b83ebfea64a42b821eb294f765d0564c7dbc497b12bf7c70088\": container with ID starting with b6eedae0063b6b83ebfea64a42b821eb294f765d0564c7dbc497b12bf7c70088 not found: ID does not exist" Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.717846 4810 scope.go:117] "RemoveContainer" containerID="3a89404c12be55963fdff956e3e961b4604968645a0e0181052773f5324baf54" Dec 01 16:22:43 crc kubenswrapper[4810]: E1201 16:22:43.718053 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a89404c12be55963fdff956e3e961b4604968645a0e0181052773f5324baf54\": container with ID starting with 3a89404c12be55963fdff956e3e961b4604968645a0e0181052773f5324baf54 not found: ID does not exist" containerID="3a89404c12be55963fdff956e3e961b4604968645a0e0181052773f5324baf54" Dec 01 16:22:43 crc kubenswrapper[4810]: I1201 16:22:43.718078 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a89404c12be55963fdff956e3e961b4604968645a0e0181052773f5324baf54"} err="failed to get container status \"3a89404c12be55963fdff956e3e961b4604968645a0e0181052773f5324baf54\": rpc error: code = NotFound desc = could not find container \"3a89404c12be55963fdff956e3e961b4604968645a0e0181052773f5324baf54\": container with ID starting with 3a89404c12be55963fdff956e3e961b4604968645a0e0181052773f5324baf54 not found: ID does not exist" Dec 01 16:22:44 crc kubenswrapper[4810]: I1201 16:22:44.508039 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51e9683a-6e27-449b-a237-4013a84619d2" path="/var/lib/kubelet/pods/51e9683a-6e27-449b-a237-4013a84619d2/volumes" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.480296 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-kx7vb"] Dec 01 16:22:47 crc kubenswrapper[4810]: E1201 16:22:47.481242 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="135d31a5-5184-4906-8c27-9f8d645ea84c" containerName="extract" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.481255 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="135d31a5-5184-4906-8c27-9f8d645ea84c" containerName="extract" Dec 01 16:22:47 crc kubenswrapper[4810]: E1201 16:22:47.481278 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51e9683a-6e27-449b-a237-4013a84619d2" containerName="registry-server" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.481284 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="51e9683a-6e27-449b-a237-4013a84619d2" containerName="registry-server" Dec 01 16:22:47 crc kubenswrapper[4810]: E1201 16:22:47.481299 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51e9683a-6e27-449b-a237-4013a84619d2" containerName="extract-utilities" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.481305 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="51e9683a-6e27-449b-a237-4013a84619d2" containerName="extract-utilities" Dec 01 16:22:47 crc kubenswrapper[4810]: E1201 16:22:47.481316 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="135d31a5-5184-4906-8c27-9f8d645ea84c" containerName="pull" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.481321 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="135d31a5-5184-4906-8c27-9f8d645ea84c" containerName="pull" Dec 01 16:22:47 crc kubenswrapper[4810]: E1201 16:22:47.481332 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="135d31a5-5184-4906-8c27-9f8d645ea84c" containerName="util" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.481338 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="135d31a5-5184-4906-8c27-9f8d645ea84c" containerName="util" Dec 01 16:22:47 crc kubenswrapper[4810]: E1201 16:22:47.481351 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51e9683a-6e27-449b-a237-4013a84619d2" containerName="extract-content" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.481356 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="51e9683a-6e27-449b-a237-4013a84619d2" containerName="extract-content" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.481557 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="135d31a5-5184-4906-8c27-9f8d645ea84c" containerName="extract" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.481585 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="51e9683a-6e27-449b-a237-4013a84619d2" containerName="registry-server" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.482246 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kx7vb" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.483764 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.484007 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.484103 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-l68gd" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.512385 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-kx7vb"] Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.599820 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm"] Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.601104 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.615317 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-7slxk" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.615597 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.616171 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdgps\" (UniqueName: \"kubernetes.io/projected/56425059-45e3-4f39-8e68-92d8ffd61831-kube-api-access-xdgps\") pod \"obo-prometheus-operator-668cf9dfbb-kx7vb\" (UID: \"56425059-45e3-4f39-8e68-92d8ffd61831\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kx7vb" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.618950 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm"] Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.664524 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-g626h"] Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.665827 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-g626h" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.674210 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-g626h"] Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.725557 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a3054e82-c7b2-4419-86cf-7089e43c4612-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7769c7475b-g626h\" (UID: \"a3054e82-c7b2-4419-86cf-7089e43c4612\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-g626h" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.725895 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a3054e82-c7b2-4419-86cf-7089e43c4612-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7769c7475b-g626h\" (UID: \"a3054e82-c7b2-4419-86cf-7089e43c4612\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-g626h" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.725934 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdgps\" (UniqueName: \"kubernetes.io/projected/56425059-45e3-4f39-8e68-92d8ffd61831-kube-api-access-xdgps\") pod \"obo-prometheus-operator-668cf9dfbb-kx7vb\" (UID: \"56425059-45e3-4f39-8e68-92d8ffd61831\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kx7vb" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.725960 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5c81ea53-c375-492e-8426-88b80c2ccb25-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm\" (UID: \"5c81ea53-c375-492e-8426-88b80c2ccb25\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.726054 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5c81ea53-c375-492e-8426-88b80c2ccb25-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm\" (UID: \"5c81ea53-c375-492e-8426-88b80c2ccb25\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.751952 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdgps\" (UniqueName: \"kubernetes.io/projected/56425059-45e3-4f39-8e68-92d8ffd61831-kube-api-access-xdgps\") pod \"obo-prometheus-operator-668cf9dfbb-kx7vb\" (UID: \"56425059-45e3-4f39-8e68-92d8ffd61831\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kx7vb" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.805293 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kx7vb" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.814256 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-7f426"] Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.815846 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-7f426" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.821538 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-p5lrs" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.821686 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.831696 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a3054e82-c7b2-4419-86cf-7089e43c4612-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7769c7475b-g626h\" (UID: \"a3054e82-c7b2-4419-86cf-7089e43c4612\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-g626h" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.831770 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a3054e82-c7b2-4419-86cf-7089e43c4612-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7769c7475b-g626h\" (UID: \"a3054e82-c7b2-4419-86cf-7089e43c4612\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-g626h" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.831806 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5c81ea53-c375-492e-8426-88b80c2ccb25-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm\" (UID: \"5c81ea53-c375-492e-8426-88b80c2ccb25\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.831881 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5c81ea53-c375-492e-8426-88b80c2ccb25-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm\" (UID: \"5c81ea53-c375-492e-8426-88b80c2ccb25\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.841972 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a3054e82-c7b2-4419-86cf-7089e43c4612-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7769c7475b-g626h\" (UID: \"a3054e82-c7b2-4419-86cf-7089e43c4612\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-g626h" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.842354 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5c81ea53-c375-492e-8426-88b80c2ccb25-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm\" (UID: \"5c81ea53-c375-492e-8426-88b80c2ccb25\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.847012 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a3054e82-c7b2-4419-86cf-7089e43c4612-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7769c7475b-g626h\" (UID: \"a3054e82-c7b2-4419-86cf-7089e43c4612\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-g626h" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.853224 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5c81ea53-c375-492e-8426-88b80c2ccb25-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm\" (UID: \"5c81ea53-c375-492e-8426-88b80c2ccb25\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.929912 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-7f426"] Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.937334 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggvgl\" (UniqueName: \"kubernetes.io/projected/9f0473a9-3b74-4538-909f-59f84c97d416-kube-api-access-ggvgl\") pod \"observability-operator-d8bb48f5d-7f426\" (UID: \"9f0473a9-3b74-4538-909f-59f84c97d416\") " pod="openshift-operators/observability-operator-d8bb48f5d-7f426" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.937535 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f0473a9-3b74-4538-909f-59f84c97d416-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-7f426\" (UID: \"9f0473a9-3b74-4538-909f-59f84c97d416\") " pod="openshift-operators/observability-operator-d8bb48f5d-7f426" Dec 01 16:22:47 crc kubenswrapper[4810]: I1201 16:22:47.937874 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm" Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.018921 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-g626h" Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.026750 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-vcr8r"] Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.028381 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-vcr8r" Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.031870 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-tv9s5" Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.036723 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-vcr8r"] Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.041764 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggvgl\" (UniqueName: \"kubernetes.io/projected/9f0473a9-3b74-4538-909f-59f84c97d416-kube-api-access-ggvgl\") pod \"observability-operator-d8bb48f5d-7f426\" (UID: \"9f0473a9-3b74-4538-909f-59f84c97d416\") " pod="openshift-operators/observability-operator-d8bb48f5d-7f426" Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.041900 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f0473a9-3b74-4538-909f-59f84c97d416-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-7f426\" (UID: \"9f0473a9-3b74-4538-909f-59f84c97d416\") " pod="openshift-operators/observability-operator-d8bb48f5d-7f426" Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.057424 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f0473a9-3b74-4538-909f-59f84c97d416-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-7f426\" (UID: \"9f0473a9-3b74-4538-909f-59f84c97d416\") " pod="openshift-operators/observability-operator-d8bb48f5d-7f426" Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.087279 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggvgl\" (UniqueName: \"kubernetes.io/projected/9f0473a9-3b74-4538-909f-59f84c97d416-kube-api-access-ggvgl\") pod \"observability-operator-d8bb48f5d-7f426\" (UID: \"9f0473a9-3b74-4538-909f-59f84c97d416\") " pod="openshift-operators/observability-operator-d8bb48f5d-7f426" Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.143875 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/b78207c2-8eae-48a2-8eaf-aa967d534298-openshift-service-ca\") pod \"perses-operator-5446b9c989-vcr8r\" (UID: \"b78207c2-8eae-48a2-8eaf-aa967d534298\") " pod="openshift-operators/perses-operator-5446b9c989-vcr8r" Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.144040 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhwzg\" (UniqueName: \"kubernetes.io/projected/b78207c2-8eae-48a2-8eaf-aa967d534298-kube-api-access-qhwzg\") pod \"perses-operator-5446b9c989-vcr8r\" (UID: \"b78207c2-8eae-48a2-8eaf-aa967d534298\") " pod="openshift-operators/perses-operator-5446b9c989-vcr8r" Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.245601 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-7f426" Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.245965 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/b78207c2-8eae-48a2-8eaf-aa967d534298-openshift-service-ca\") pod \"perses-operator-5446b9c989-vcr8r\" (UID: \"b78207c2-8eae-48a2-8eaf-aa967d534298\") " pod="openshift-operators/perses-operator-5446b9c989-vcr8r" Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.246100 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhwzg\" (UniqueName: \"kubernetes.io/projected/b78207c2-8eae-48a2-8eaf-aa967d534298-kube-api-access-qhwzg\") pod \"perses-operator-5446b9c989-vcr8r\" (UID: \"b78207c2-8eae-48a2-8eaf-aa967d534298\") " pod="openshift-operators/perses-operator-5446b9c989-vcr8r" Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.247005 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/b78207c2-8eae-48a2-8eaf-aa967d534298-openshift-service-ca\") pod \"perses-operator-5446b9c989-vcr8r\" (UID: \"b78207c2-8eae-48a2-8eaf-aa967d534298\") " pod="openshift-operators/perses-operator-5446b9c989-vcr8r" Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.291213 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhwzg\" (UniqueName: \"kubernetes.io/projected/b78207c2-8eae-48a2-8eaf-aa967d534298-kube-api-access-qhwzg\") pod \"perses-operator-5446b9c989-vcr8r\" (UID: \"b78207c2-8eae-48a2-8eaf-aa967d534298\") " pod="openshift-operators/perses-operator-5446b9c989-vcr8r" Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.414646 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-vcr8r" Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.423970 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-kx7vb"] Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.636764 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kx7vb" event={"ID":"56425059-45e3-4f39-8e68-92d8ffd61831","Type":"ContainerStarted","Data":"37c05cbd0875833e9acc12f51afa945cb47a53221ec51c401e8823a0ba440a91"} Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.656282 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm"] Dec 01 16:22:48 crc kubenswrapper[4810]: I1201 16:22:48.851514 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-g626h"] Dec 01 16:22:49 crc kubenswrapper[4810]: I1201 16:22:49.003486 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-7f426"] Dec 01 16:22:49 crc kubenswrapper[4810]: W1201 16:22:49.007135 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f0473a9_3b74_4538_909f_59f84c97d416.slice/crio-4ccf2e90e97229939e7acf4bd3c3aec39443deac128ac3fda63488fa2568883e WatchSource:0}: Error finding container 4ccf2e90e97229939e7acf4bd3c3aec39443deac128ac3fda63488fa2568883e: Status 404 returned error can't find the container with id 4ccf2e90e97229939e7acf4bd3c3aec39443deac128ac3fda63488fa2568883e Dec 01 16:22:49 crc kubenswrapper[4810]: I1201 16:22:49.123218 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-vcr8r"] Dec 01 16:22:49 crc kubenswrapper[4810]: W1201 16:22:49.125003 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb78207c2_8eae_48a2_8eaf_aa967d534298.slice/crio-bd294d2ae5e0726e9a76751cfba346dbec68868f4031c64366421f6f47f0df7f WatchSource:0}: Error finding container bd294d2ae5e0726e9a76751cfba346dbec68868f4031c64366421f6f47f0df7f: Status 404 returned error can't find the container with id bd294d2ae5e0726e9a76751cfba346dbec68868f4031c64366421f6f47f0df7f Dec 01 16:22:49 crc kubenswrapper[4810]: I1201 16:22:49.656659 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-7f426" event={"ID":"9f0473a9-3b74-4538-909f-59f84c97d416","Type":"ContainerStarted","Data":"4ccf2e90e97229939e7acf4bd3c3aec39443deac128ac3fda63488fa2568883e"} Dec 01 16:22:49 crc kubenswrapper[4810]: I1201 16:22:49.658271 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-g626h" event={"ID":"a3054e82-c7b2-4419-86cf-7089e43c4612","Type":"ContainerStarted","Data":"1a2b974295e5509020761df84899efaa7f1ec50cad9edfa2a9219d324460e455"} Dec 01 16:22:49 crc kubenswrapper[4810]: I1201 16:22:49.660449 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-vcr8r" event={"ID":"b78207c2-8eae-48a2-8eaf-aa967d534298","Type":"ContainerStarted","Data":"bd294d2ae5e0726e9a76751cfba346dbec68868f4031c64366421f6f47f0df7f"} Dec 01 16:22:49 crc kubenswrapper[4810]: I1201 16:22:49.661728 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm" event={"ID":"5c81ea53-c375-492e-8426-88b80c2ccb25","Type":"ContainerStarted","Data":"25ea6e8242536a8e2f6b7440eeddb6a4126819d8d1194de1ce3b4a672e669299"} Dec 01 16:22:56 crc kubenswrapper[4810]: I1201 16:22:56.797736 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-vcr8r" event={"ID":"b78207c2-8eae-48a2-8eaf-aa967d534298","Type":"ContainerStarted","Data":"e4f20c5208099a08ab70b5bc8b60776347ca97137c7bfc3702d8dd89a3df63a3"} Dec 01 16:22:56 crc kubenswrapper[4810]: I1201 16:22:56.798415 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-vcr8r" Dec 01 16:22:56 crc kubenswrapper[4810]: I1201 16:22:56.799647 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-7f426" event={"ID":"9f0473a9-3b74-4538-909f-59f84c97d416","Type":"ContainerStarted","Data":"11051651c2215925efd60a0a52b6d797cd755537906e359486d760e841f3e256"} Dec 01 16:22:56 crc kubenswrapper[4810]: I1201 16:22:56.799911 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-7f426" Dec 01 16:22:56 crc kubenswrapper[4810]: I1201 16:22:56.801322 4810 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-7f426 container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.1.138:8081/healthz\": dial tcp 10.217.1.138:8081: connect: connection refused" start-of-body= Dec 01 16:22:56 crc kubenswrapper[4810]: I1201 16:22:56.801367 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-7f426" podUID="9f0473a9-3b74-4538-909f-59f84c97d416" containerName="operator" probeResult="failure" output="Get \"http://10.217.1.138:8081/healthz\": dial tcp 10.217.1.138:8081: connect: connection refused" Dec 01 16:22:56 crc kubenswrapper[4810]: I1201 16:22:56.827266 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-vcr8r" podStartSLOduration=2.645621638 podStartE2EDuration="9.827232955s" podCreationTimestamp="2025-12-01 16:22:47 +0000 UTC" firstStartedPulling="2025-12-01 16:22:49.128212961 +0000 UTC m=+6534.891722564" lastFinishedPulling="2025-12-01 16:22:56.309824278 +0000 UTC m=+6542.073333881" observedRunningTime="2025-12-01 16:22:56.821429037 +0000 UTC m=+6542.584938640" watchObservedRunningTime="2025-12-01 16:22:56.827232955 +0000 UTC m=+6542.590742558" Dec 01 16:22:56 crc kubenswrapper[4810]: I1201 16:22:56.858881 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-7f426" podStartSLOduration=2.4812992879999998 podStartE2EDuration="9.858858924s" podCreationTimestamp="2025-12-01 16:22:47 +0000 UTC" firstStartedPulling="2025-12-01 16:22:49.010323071 +0000 UTC m=+6534.773832674" lastFinishedPulling="2025-12-01 16:22:56.387882707 +0000 UTC m=+6542.151392310" observedRunningTime="2025-12-01 16:22:56.845583373 +0000 UTC m=+6542.609092976" watchObservedRunningTime="2025-12-01 16:22:56.858858924 +0000 UTC m=+6542.622368537" Dec 01 16:22:57 crc kubenswrapper[4810]: I1201 16:22:57.813105 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-g626h" event={"ID":"a3054e82-c7b2-4419-86cf-7089e43c4612","Type":"ContainerStarted","Data":"cbe986ffeb91f30b666b3219a21d901cf04cd79141950d23a1f5583c198134b0"} Dec 01 16:22:57 crc kubenswrapper[4810]: I1201 16:22:57.816140 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kx7vb" event={"ID":"56425059-45e3-4f39-8e68-92d8ffd61831","Type":"ContainerStarted","Data":"85bc7542726a273f78d6e377741fae0062ddaf8c6e980e21214d0e8e0107a904"} Dec 01 16:22:57 crc kubenswrapper[4810]: I1201 16:22:57.817840 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm" event={"ID":"5c81ea53-c375-492e-8426-88b80c2ccb25","Type":"ContainerStarted","Data":"695f6ab4246d0258b19d4d86b183a971118e9828727d93ceee4986ea1508f38c"} Dec 01 16:22:57 crc kubenswrapper[4810]: I1201 16:22:57.836162 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-g626h" podStartSLOduration=3.442058787 podStartE2EDuration="10.836132103s" podCreationTimestamp="2025-12-01 16:22:47 +0000 UTC" firstStartedPulling="2025-12-01 16:22:48.859360892 +0000 UTC m=+6534.622870495" lastFinishedPulling="2025-12-01 16:22:56.253434188 +0000 UTC m=+6542.016943811" observedRunningTime="2025-12-01 16:22:57.833769519 +0000 UTC m=+6543.597279132" watchObservedRunningTime="2025-12-01 16:22:57.836132103 +0000 UTC m=+6543.599641726" Dec 01 16:22:57 crc kubenswrapper[4810]: I1201 16:22:57.852372 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-7f426" Dec 01 16:22:57 crc kubenswrapper[4810]: I1201 16:22:57.857635 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm" podStartSLOduration=3.281612303 podStartE2EDuration="10.857616357s" podCreationTimestamp="2025-12-01 16:22:47 +0000 UTC" firstStartedPulling="2025-12-01 16:22:48.678816101 +0000 UTC m=+6534.442325704" lastFinishedPulling="2025-12-01 16:22:56.254820155 +0000 UTC m=+6542.018329758" observedRunningTime="2025-12-01 16:22:57.849715302 +0000 UTC m=+6543.613224905" watchObservedRunningTime="2025-12-01 16:22:57.857616357 +0000 UTC m=+6543.621125980" Dec 01 16:22:57 crc kubenswrapper[4810]: I1201 16:22:57.943901 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-kx7vb" podStartSLOduration=3.105286656 podStartE2EDuration="10.943878688s" podCreationTimestamp="2025-12-01 16:22:47 +0000 UTC" firstStartedPulling="2025-12-01 16:22:48.470822455 +0000 UTC m=+6534.234332058" lastFinishedPulling="2025-12-01 16:22:56.309414487 +0000 UTC m=+6542.072924090" observedRunningTime="2025-12-01 16:22:57.887041445 +0000 UTC m=+6543.650551078" watchObservedRunningTime="2025-12-01 16:22:57.943878688 +0000 UTC m=+6543.707388291" Dec 01 16:23:08 crc kubenswrapper[4810]: I1201 16:23:08.419180 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-vcr8r" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.589741 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.590473 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="e94a01a9-14d5-4dd9-b98a-5fcd66f2239b" containerName="openstackclient" containerID="cri-o://73589ed04e6cb6c6075477260b43267dffa628c16a6dc292c1ae630bdd4bd7b0" gracePeriod=2 Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.597956 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.719725 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 16:23:11 crc kubenswrapper[4810]: E1201 16:23:11.720551 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e94a01a9-14d5-4dd9-b98a-5fcd66f2239b" containerName="openstackclient" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.720569 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e94a01a9-14d5-4dd9-b98a-5fcd66f2239b" containerName="openstackclient" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.731745 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e94a01a9-14d5-4dd9-b98a-5fcd66f2239b" containerName="openstackclient" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.733003 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.743519 4810 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="e94a01a9-14d5-4dd9-b98a-5fcd66f2239b" podUID="bec7da3e-6c13-4883-9c56-4e98c688c4c5" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.754980 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.766528 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fglw2\" (UniqueName: \"kubernetes.io/projected/bec7da3e-6c13-4883-9c56-4e98c688c4c5-kube-api-access-fglw2\") pod \"openstackclient\" (UID: \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\") " pod="openstack/openstackclient" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.766640 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bec7da3e-6c13-4883-9c56-4e98c688c4c5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\") " pod="openstack/openstackclient" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.766700 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bec7da3e-6c13-4883-9c56-4e98c688c4c5-openstack-config\") pod \"openstackclient\" (UID: \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\") " pod="openstack/openstackclient" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.766750 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bec7da3e-6c13-4883-9c56-4e98c688c4c5-openstack-config-secret\") pod \"openstackclient\" (UID: \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\") " pod="openstack/openstackclient" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.816657 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.818103 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.821456 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-4tmsl" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.836944 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.871768 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fglw2\" (UniqueName: \"kubernetes.io/projected/bec7da3e-6c13-4883-9c56-4e98c688c4c5-kube-api-access-fglw2\") pod \"openstackclient\" (UID: \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\") " pod="openstack/openstackclient" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.871878 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bec7da3e-6c13-4883-9c56-4e98c688c4c5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\") " pod="openstack/openstackclient" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.871935 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bec7da3e-6c13-4883-9c56-4e98c688c4c5-openstack-config\") pod \"openstackclient\" (UID: \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\") " pod="openstack/openstackclient" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.871981 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bec7da3e-6c13-4883-9c56-4e98c688c4c5-openstack-config-secret\") pod \"openstackclient\" (UID: \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\") " pod="openstack/openstackclient" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.874331 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bec7da3e-6c13-4883-9c56-4e98c688c4c5-openstack-config\") pod \"openstackclient\" (UID: \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\") " pod="openstack/openstackclient" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.878191 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bec7da3e-6c13-4883-9c56-4e98c688c4c5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\") " pod="openstack/openstackclient" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.884381 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bec7da3e-6c13-4883-9c56-4e98c688c4c5-openstack-config-secret\") pod \"openstackclient\" (UID: \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\") " pod="openstack/openstackclient" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.928832 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fglw2\" (UniqueName: \"kubernetes.io/projected/bec7da3e-6c13-4883-9c56-4e98c688c4c5-kube-api-access-fglw2\") pod \"openstackclient\" (UID: \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\") " pod="openstack/openstackclient" Dec 01 16:23:11 crc kubenswrapper[4810]: I1201 16:23:11.985419 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r77sj\" (UniqueName: \"kubernetes.io/projected/67e37611-812c-4270-9225-2ed16b14c0a2-kube-api-access-r77sj\") pod \"kube-state-metrics-0\" (UID: \"67e37611-812c-4270-9225-2ed16b14c0a2\") " pod="openstack/kube-state-metrics-0" Dec 01 16:23:12 crc kubenswrapper[4810]: I1201 16:23:12.008283 4810 scope.go:117] "RemoveContainer" containerID="f98456462717d9ca893cd8ec8685e17043a3d1a8de5ca47ee03ad3253cdab906" Dec 01 16:23:12 crc kubenswrapper[4810]: I1201 16:23:12.070973 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 16:23:12 crc kubenswrapper[4810]: I1201 16:23:12.086274 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r77sj\" (UniqueName: \"kubernetes.io/projected/67e37611-812c-4270-9225-2ed16b14c0a2-kube-api-access-r77sj\") pod \"kube-state-metrics-0\" (UID: \"67e37611-812c-4270-9225-2ed16b14c0a2\") " pod="openstack/kube-state-metrics-0" Dec 01 16:23:12 crc kubenswrapper[4810]: I1201 16:23:12.121223 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r77sj\" (UniqueName: \"kubernetes.io/projected/67e37611-812c-4270-9225-2ed16b14c0a2-kube-api-access-r77sj\") pod \"kube-state-metrics-0\" (UID: \"67e37611-812c-4270-9225-2ed16b14c0a2\") " pod="openstack/kube-state-metrics-0" Dec 01 16:23:12 crc kubenswrapper[4810]: I1201 16:23:12.156408 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 16:23:12 crc kubenswrapper[4810]: W1201 16:23:12.913357 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbec7da3e_6c13_4883_9c56_4e98c688c4c5.slice/crio-9655656e07fcdffcf0ded7ec69a7f41d70db952e2e9d4064811d0ef1ab9ea4ce WatchSource:0}: Error finding container 9655656e07fcdffcf0ded7ec69a7f41d70db952e2e9d4064811d0ef1ab9ea4ce: Status 404 returned error can't find the container with id 9655656e07fcdffcf0ded7ec69a7f41d70db952e2e9d4064811d0ef1ab9ea4ce Dec 01 16:23:12 crc kubenswrapper[4810]: I1201 16:23:12.913646 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 16:23:12 crc kubenswrapper[4810]: I1201 16:23:12.971325 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"bec7da3e-6c13-4883-9c56-4e98c688c4c5","Type":"ContainerStarted","Data":"9655656e07fcdffcf0ded7ec69a7f41d70db952e2e9d4064811d0ef1ab9ea4ce"} Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.027432 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.133241 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.138229 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.143288 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-gqh87" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.143562 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.153616 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.155002 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.155055 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.155087 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.155108 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.221655 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6b2c284-91c8-4419-b078-66f0dfcf46de-config\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.221722 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f6b2c284-91c8-4419-b078-66f0dfcf46de-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.221748 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f6b2c284-91c8-4419-b078-66f0dfcf46de-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.221776 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-be429350-f439-458e-a105-c9855c2edb42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be429350-f439-458e-a105-c9855c2edb42\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.221827 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f6b2c284-91c8-4419-b078-66f0dfcf46de-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.221846 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7xzf\" (UniqueName: \"kubernetes.io/projected/f6b2c284-91c8-4419-b078-66f0dfcf46de-kube-api-access-g7xzf\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.221902 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f6b2c284-91c8-4419-b078-66f0dfcf46de-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.221942 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f6b2c284-91c8-4419-b078-66f0dfcf46de-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.323843 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f6b2c284-91c8-4419-b078-66f0dfcf46de-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.323896 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f6b2c284-91c8-4419-b078-66f0dfcf46de-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.323927 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-be429350-f439-458e-a105-c9855c2edb42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be429350-f439-458e-a105-c9855c2edb42\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.323980 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f6b2c284-91c8-4419-b078-66f0dfcf46de-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.324002 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7xzf\" (UniqueName: \"kubernetes.io/projected/f6b2c284-91c8-4419-b078-66f0dfcf46de-kube-api-access-g7xzf\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.324057 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f6b2c284-91c8-4419-b078-66f0dfcf46de-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.324097 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f6b2c284-91c8-4419-b078-66f0dfcf46de-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.324149 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6b2c284-91c8-4419-b078-66f0dfcf46de-config\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.325098 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f6b2c284-91c8-4419-b078-66f0dfcf46de-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.331438 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f6b2c284-91c8-4419-b078-66f0dfcf46de-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.331554 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f6b2c284-91c8-4419-b078-66f0dfcf46de-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.333239 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6b2c284-91c8-4419-b078-66f0dfcf46de-config\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.341980 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f6b2c284-91c8-4419-b078-66f0dfcf46de-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.343877 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.343899 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-be429350-f439-458e-a105-c9855c2edb42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be429350-f439-458e-a105-c9855c2edb42\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5ec5cb75b4a42b22102ef267695ee69cf2fa8f177a6aa2457575c731cb182193/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.346436 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f6b2c284-91c8-4419-b078-66f0dfcf46de-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.361198 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7xzf\" (UniqueName: \"kubernetes.io/projected/f6b2c284-91c8-4419-b078-66f0dfcf46de-kube-api-access-g7xzf\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.538218 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-be429350-f439-458e-a105-c9855c2edb42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be429350-f439-458e-a105-c9855c2edb42\") pod \"prometheus-metric-storage-0\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:13 crc kubenswrapper[4810]: I1201 16:23:13.760001 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.006507 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"67e37611-812c-4270-9225-2ed16b14c0a2","Type":"ContainerStarted","Data":"2474303523ecd36c3c4c15fd281847fc3259faaaf1bcd7a17867ffa92b23bb70"} Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.006775 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"67e37611-812c-4270-9225-2ed16b14c0a2","Type":"ContainerStarted","Data":"c6402d52428d471ca1fbf8a60ae556386a2e8fc2e4895ed63ee4321b7875b86d"} Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.007585 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.043032 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.660533617 podStartE2EDuration="3.04301395s" podCreationTimestamp="2025-12-01 16:23:11 +0000 UTC" firstStartedPulling="2025-12-01 16:23:13.034153992 +0000 UTC m=+6558.797663585" lastFinishedPulling="2025-12-01 16:23:13.416634315 +0000 UTC m=+6559.180143918" observedRunningTime="2025-12-01 16:23:14.039556196 +0000 UTC m=+6559.803065809" watchObservedRunningTime="2025-12-01 16:23:14.04301395 +0000 UTC m=+6559.806523553" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.051192 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"bec7da3e-6c13-4883-9c56-4e98c688c4c5","Type":"ContainerStarted","Data":"1c128fe541adca617bb9b988c393195875144309c38e34224d3ffe479f2a327f"} Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.086061 4810 generic.go:334] "Generic (PLEG): container finished" podID="e94a01a9-14d5-4dd9-b98a-5fcd66f2239b" containerID="73589ed04e6cb6c6075477260b43267dffa628c16a6dc292c1ae630bdd4bd7b0" exitCode=137 Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.092886 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.092866533 podStartE2EDuration="3.092866533s" podCreationTimestamp="2025-12-01 16:23:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:23:14.080195879 +0000 UTC m=+6559.843705482" watchObservedRunningTime="2025-12-01 16:23:14.092866533 +0000 UTC m=+6559.856376136" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.336895 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.339712 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.348595 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.349022 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-5k2j5" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.349269 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.352771 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.352977 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.375799 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.473102 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.473268 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n62qk\" (UniqueName: \"kubernetes.io/projected/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-kube-api-access-n62qk\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.473310 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.473397 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.473592 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.473696 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.473940 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.579378 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.579560 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.579691 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.579736 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n62qk\" (UniqueName: \"kubernetes.io/projected/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-kube-api-access-n62qk\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.579759 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.579795 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.579817 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.590570 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.598043 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.602031 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.602334 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.605566 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.608129 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.610261 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.671392 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n62qk\" (UniqueName: \"kubernetes.io/projected/8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7-kube-api-access-n62qk\") pod \"alertmanager-metric-storage-0\" (UID: \"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.677319 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.973591 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 16:23:14 crc kubenswrapper[4810]: I1201 16:23:14.981189 4810 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="e94a01a9-14d5-4dd9-b98a-5fcd66f2239b" podUID="bec7da3e-6c13-4883-9c56-4e98c688c4c5" Dec 01 16:23:15 crc kubenswrapper[4810]: I1201 16:23:15.104472 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-openstack-config-secret\") pod \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\" (UID: \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\") " Dec 01 16:23:15 crc kubenswrapper[4810]: I1201 16:23:15.104661 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vn6f\" (UniqueName: \"kubernetes.io/projected/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-kube-api-access-6vn6f\") pod \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\" (UID: \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\") " Dec 01 16:23:15 crc kubenswrapper[4810]: I1201 16:23:15.104679 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6b2c284-91c8-4419-b078-66f0dfcf46de","Type":"ContainerStarted","Data":"c5023a0636cb8ef089d46f71d741f6f1b2b0ecbcb27cf968f599a0cf2cd0439c"} Dec 01 16:23:15 crc kubenswrapper[4810]: I1201 16:23:15.104777 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-combined-ca-bundle\") pod \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\" (UID: \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\") " Dec 01 16:23:15 crc kubenswrapper[4810]: I1201 16:23:15.104870 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-openstack-config\") pod \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\" (UID: \"e94a01a9-14d5-4dd9-b98a-5fcd66f2239b\") " Dec 01 16:23:15 crc kubenswrapper[4810]: I1201 16:23:15.110866 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 16:23:15 crc kubenswrapper[4810]: I1201 16:23:15.110728 4810 scope.go:117] "RemoveContainer" containerID="73589ed04e6cb6c6075477260b43267dffa628c16a6dc292c1ae630bdd4bd7b0" Dec 01 16:23:15 crc kubenswrapper[4810]: I1201 16:23:15.112144 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-kube-api-access-6vn6f" (OuterVolumeSpecName: "kube-api-access-6vn6f") pod "e94a01a9-14d5-4dd9-b98a-5fcd66f2239b" (UID: "e94a01a9-14d5-4dd9-b98a-5fcd66f2239b"). InnerVolumeSpecName "kube-api-access-6vn6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:23:15 crc kubenswrapper[4810]: I1201 16:23:15.115197 4810 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="e94a01a9-14d5-4dd9-b98a-5fcd66f2239b" podUID="bec7da3e-6c13-4883-9c56-4e98c688c4c5" Dec 01 16:23:15 crc kubenswrapper[4810]: I1201 16:23:15.171653 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "e94a01a9-14d5-4dd9-b98a-5fcd66f2239b" (UID: "e94a01a9-14d5-4dd9-b98a-5fcd66f2239b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:23:15 crc kubenswrapper[4810]: I1201 16:23:15.186719 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e94a01a9-14d5-4dd9-b98a-5fcd66f2239b" (UID: "e94a01a9-14d5-4dd9-b98a-5fcd66f2239b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:23:15 crc kubenswrapper[4810]: I1201 16:23:15.190316 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "e94a01a9-14d5-4dd9-b98a-5fcd66f2239b" (UID: "e94a01a9-14d5-4dd9-b98a-5fcd66f2239b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:23:15 crc kubenswrapper[4810]: I1201 16:23:15.216153 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:23:15 crc kubenswrapper[4810]: I1201 16:23:15.216457 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 01 16:23:15 crc kubenswrapper[4810]: I1201 16:23:15.216491 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vn6f\" (UniqueName: \"kubernetes.io/projected/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-kube-api-access-6vn6f\") on node \"crc\" DevicePath \"\"" Dec 01 16:23:15 crc kubenswrapper[4810]: I1201 16:23:15.216502 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:23:15 crc kubenswrapper[4810]: I1201 16:23:15.430312 4810 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="e94a01a9-14d5-4dd9-b98a-5fcd66f2239b" podUID="bec7da3e-6c13-4883-9c56-4e98c688c4c5" Dec 01 16:23:15 crc kubenswrapper[4810]: I1201 16:23:15.456010 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 01 16:23:16 crc kubenswrapper[4810]: I1201 16:23:16.122265 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7","Type":"ContainerStarted","Data":"752784e2e6c205b18b17552b2dbdca6fd29aaa39f4e942b5d0e1844e6d21cf9b"} Dec 01 16:23:16 crc kubenswrapper[4810]: I1201 16:23:16.506270 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e94a01a9-14d5-4dd9-b98a-5fcd66f2239b" path="/var/lib/kubelet/pods/e94a01a9-14d5-4dd9-b98a-5fcd66f2239b/volumes" Dec 01 16:23:18 crc kubenswrapper[4810]: I1201 16:23:18.036184 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-481d-account-create-update-2rdl9"] Dec 01 16:23:18 crc kubenswrapper[4810]: I1201 16:23:18.045160 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-9lwc4"] Dec 01 16:23:18 crc kubenswrapper[4810]: I1201 16:23:18.056003 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-9lwc4"] Dec 01 16:23:18 crc kubenswrapper[4810]: I1201 16:23:18.065198 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-481d-account-create-update-2rdl9"] Dec 01 16:23:18 crc kubenswrapper[4810]: I1201 16:23:18.511285 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25c3e988-ae27-4369-aa98-4bd7a3c5848a" path="/var/lib/kubelet/pods/25c3e988-ae27-4369-aa98-4bd7a3c5848a/volumes" Dec 01 16:23:18 crc kubenswrapper[4810]: I1201 16:23:18.513312 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="607341d7-10b4-4d40-a2a2-7647ceff0c14" path="/var/lib/kubelet/pods/607341d7-10b4-4d40-a2a2-7647ceff0c14/volumes" Dec 01 16:23:21 crc kubenswrapper[4810]: I1201 16:23:21.170825 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6b2c284-91c8-4419-b078-66f0dfcf46de","Type":"ContainerStarted","Data":"8671c51fb48482df2b04b4150d35fd4083bc424f37bf501f80446e3585f33786"} Dec 01 16:23:21 crc kubenswrapper[4810]: I1201 16:23:21.173029 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7","Type":"ContainerStarted","Data":"aee0b73da16d56496e596789192c1c199baef368090e72f65d9e704a624a6a91"} Dec 01 16:23:22 crc kubenswrapper[4810]: I1201 16:23:22.161084 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 01 16:23:25 crc kubenswrapper[4810]: I1201 16:23:25.070514 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-4q58g"] Dec 01 16:23:25 crc kubenswrapper[4810]: I1201 16:23:25.087770 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-4q58g"] Dec 01 16:23:26 crc kubenswrapper[4810]: I1201 16:23:26.503753 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="908c7ff6-8b25-4011-b1da-159a16755ac5" path="/var/lib/kubelet/pods/908c7ff6-8b25-4011-b1da-159a16755ac5/volumes" Dec 01 16:23:28 crc kubenswrapper[4810]: I1201 16:23:28.259041 4810 generic.go:334] "Generic (PLEG): container finished" podID="8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7" containerID="aee0b73da16d56496e596789192c1c199baef368090e72f65d9e704a624a6a91" exitCode=0 Dec 01 16:23:28 crc kubenswrapper[4810]: I1201 16:23:28.259198 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7","Type":"ContainerDied","Data":"aee0b73da16d56496e596789192c1c199baef368090e72f65d9e704a624a6a91"} Dec 01 16:23:28 crc kubenswrapper[4810]: I1201 16:23:28.261878 4810 generic.go:334] "Generic (PLEG): container finished" podID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerID="8671c51fb48482df2b04b4150d35fd4083bc424f37bf501f80446e3585f33786" exitCode=0 Dec 01 16:23:28 crc kubenswrapper[4810]: I1201 16:23:28.261939 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6b2c284-91c8-4419-b078-66f0dfcf46de","Type":"ContainerDied","Data":"8671c51fb48482df2b04b4150d35fd4083bc424f37bf501f80446e3585f33786"} Dec 01 16:23:33 crc kubenswrapper[4810]: I1201 16:23:33.326197 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7","Type":"ContainerStarted","Data":"61d2ed24b5fd7f3a9779ff079cb093ccd280891a1a49207007048fb3ec611264"} Dec 01 16:23:36 crc kubenswrapper[4810]: I1201 16:23:36.369588 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6b2c284-91c8-4419-b078-66f0dfcf46de","Type":"ContainerStarted","Data":"15ecf553d2f6db20425e4b6d1a9fd862b0af424fc27c0b249931118f089be67a"} Dec 01 16:23:37 crc kubenswrapper[4810]: I1201 16:23:37.387500 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7","Type":"ContainerStarted","Data":"8d81a93c05efd7ddf0b0b3a470f259ecb6487059d56122b047feace6b6cb6317"} Dec 01 16:23:37 crc kubenswrapper[4810]: I1201 16:23:37.387878 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:37 crc kubenswrapper[4810]: I1201 16:23:37.391898 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Dec 01 16:23:37 crc kubenswrapper[4810]: I1201 16:23:37.413838 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=6.095155025 podStartE2EDuration="23.413818822s" podCreationTimestamp="2025-12-01 16:23:14 +0000 UTC" firstStartedPulling="2025-12-01 16:23:15.471930291 +0000 UTC m=+6561.235439894" lastFinishedPulling="2025-12-01 16:23:32.790594088 +0000 UTC m=+6578.554103691" observedRunningTime="2025-12-01 16:23:37.410333398 +0000 UTC m=+6583.173843021" watchObservedRunningTime="2025-12-01 16:23:37.413818822 +0000 UTC m=+6583.177328425" Dec 01 16:23:40 crc kubenswrapper[4810]: I1201 16:23:40.419863 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6b2c284-91c8-4419-b078-66f0dfcf46de","Type":"ContainerStarted","Data":"ac41df1b33599bfe7b04da23ca73bb5f7d795c5312a516b31eb8fa53a417e62f"} Dec 01 16:23:43 crc kubenswrapper[4810]: I1201 16:23:43.456233 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6b2c284-91c8-4419-b078-66f0dfcf46de","Type":"ContainerStarted","Data":"e0308580cdf2047b1427f6efb397b8d831a9568fd6e530c3da34b986302f9b79"} Dec 01 16:23:43 crc kubenswrapper[4810]: I1201 16:23:43.480671 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.332259329 podStartE2EDuration="31.480653048s" podCreationTimestamp="2025-12-01 16:23:12 +0000 UTC" firstStartedPulling="2025-12-01 16:23:14.593237657 +0000 UTC m=+6560.356747260" lastFinishedPulling="2025-12-01 16:23:42.741631376 +0000 UTC m=+6588.505140979" observedRunningTime="2025-12-01 16:23:43.479186678 +0000 UTC m=+6589.242696281" watchObservedRunningTime="2025-12-01 16:23:43.480653048 +0000 UTC m=+6589.244162651" Dec 01 16:23:43 crc kubenswrapper[4810]: I1201 16:23:43.761642 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:43 crc kubenswrapper[4810]: I1201 16:23:43.761692 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:43 crc kubenswrapper[4810]: I1201 16:23:43.763811 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:44 crc kubenswrapper[4810]: I1201 16:23:44.466220 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:45 crc kubenswrapper[4810]: I1201 16:23:45.962722 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 01 16:23:45 crc kubenswrapper[4810]: I1201 16:23:45.963232 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="bec7da3e-6c13-4883-9c56-4e98c688c4c5" containerName="openstackclient" containerID="cri-o://1c128fe541adca617bb9b988c393195875144309c38e34224d3ffe479f2a327f" gracePeriod=2 Dec 01 16:23:45 crc kubenswrapper[4810]: I1201 16:23:45.978951 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.005391 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 16:23:46 crc kubenswrapper[4810]: E1201 16:23:46.005983 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bec7da3e-6c13-4883-9c56-4e98c688c4c5" containerName="openstackclient" Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.006002 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bec7da3e-6c13-4883-9c56-4e98c688c4c5" containerName="openstackclient" Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.006288 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bec7da3e-6c13-4883-9c56-4e98c688c4c5" containerName="openstackclient" Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.007230 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.014252 4810 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bec7da3e-6c13-4883-9c56-4e98c688c4c5" podUID="20828f9e-5524-4c73-8754-2e1d60bb775a" Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.016933 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.145425 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/20828f9e-5524-4c73-8754-2e1d60bb775a-openstack-config\") pod \"openstackclient\" (UID: \"20828f9e-5524-4c73-8754-2e1d60bb775a\") " pod="openstack/openstackclient" Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.145818 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20828f9e-5524-4c73-8754-2e1d60bb775a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"20828f9e-5524-4c73-8754-2e1d60bb775a\") " pod="openstack/openstackclient" Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.145917 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/20828f9e-5524-4c73-8754-2e1d60bb775a-openstack-config-secret\") pod \"openstackclient\" (UID: \"20828f9e-5524-4c73-8754-2e1d60bb775a\") " pod="openstack/openstackclient" Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.146116 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkxn7\" (UniqueName: \"kubernetes.io/projected/20828f9e-5524-4c73-8754-2e1d60bb775a-kube-api-access-jkxn7\") pod \"openstackclient\" (UID: \"20828f9e-5524-4c73-8754-2e1d60bb775a\") " pod="openstack/openstackclient" Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.248010 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20828f9e-5524-4c73-8754-2e1d60bb775a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"20828f9e-5524-4c73-8754-2e1d60bb775a\") " pod="openstack/openstackclient" Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.248085 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/20828f9e-5524-4c73-8754-2e1d60bb775a-openstack-config-secret\") pod \"openstackclient\" (UID: \"20828f9e-5524-4c73-8754-2e1d60bb775a\") " pod="openstack/openstackclient" Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.248189 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkxn7\" (UniqueName: \"kubernetes.io/projected/20828f9e-5524-4c73-8754-2e1d60bb775a-kube-api-access-jkxn7\") pod \"openstackclient\" (UID: \"20828f9e-5524-4c73-8754-2e1d60bb775a\") " pod="openstack/openstackclient" Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.248308 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/20828f9e-5524-4c73-8754-2e1d60bb775a-openstack-config\") pod \"openstackclient\" (UID: \"20828f9e-5524-4c73-8754-2e1d60bb775a\") " pod="openstack/openstackclient" Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.249402 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/20828f9e-5524-4c73-8754-2e1d60bb775a-openstack-config\") pod \"openstackclient\" (UID: \"20828f9e-5524-4c73-8754-2e1d60bb775a\") " pod="openstack/openstackclient" Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.255071 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/20828f9e-5524-4c73-8754-2e1d60bb775a-openstack-config-secret\") pod \"openstackclient\" (UID: \"20828f9e-5524-4c73-8754-2e1d60bb775a\") " pod="openstack/openstackclient" Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.262840 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20828f9e-5524-4c73-8754-2e1d60bb775a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"20828f9e-5524-4c73-8754-2e1d60bb775a\") " pod="openstack/openstackclient" Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.269392 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkxn7\" (UniqueName: \"kubernetes.io/projected/20828f9e-5524-4c73-8754-2e1d60bb775a-kube-api-access-jkxn7\") pod \"openstackclient\" (UID: \"20828f9e-5524-4c73-8754-2e1d60bb775a\") " pod="openstack/openstackclient" Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.336594 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 16:23:46 crc kubenswrapper[4810]: I1201 16:23:46.886008 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 16:23:46 crc kubenswrapper[4810]: W1201 16:23:46.892866 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20828f9e_5524_4c73_8754_2e1d60bb775a.slice/crio-ad6f617639e7dad660b95cc87f8d0d32ea4230a875c4b62d068cf6bae6467f29 WatchSource:0}: Error finding container ad6f617639e7dad660b95cc87f8d0d32ea4230a875c4b62d068cf6bae6467f29: Status 404 returned error can't find the container with id ad6f617639e7dad660b95cc87f8d0d32ea4230a875c4b62d068cf6bae6467f29 Dec 01 16:23:47 crc kubenswrapper[4810]: I1201 16:23:47.298690 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 16:23:47 crc kubenswrapper[4810]: I1201 16:23:47.499970 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"20828f9e-5524-4c73-8754-2e1d60bb775a","Type":"ContainerStarted","Data":"c70447372e3313b2fda13d3349eeaa484accbe076295b69e3cb2d543243e795d"} Dec 01 16:23:47 crc kubenswrapper[4810]: I1201 16:23:47.500379 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"20828f9e-5524-4c73-8754-2e1d60bb775a","Type":"ContainerStarted","Data":"ad6f617639e7dad660b95cc87f8d0d32ea4230a875c4b62d068cf6bae6467f29"} Dec 01 16:23:47 crc kubenswrapper[4810]: I1201 16:23:47.500172 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerName="thanos-sidecar" containerID="cri-o://e0308580cdf2047b1427f6efb397b8d831a9568fd6e530c3da34b986302f9b79" gracePeriod=600 Dec 01 16:23:47 crc kubenswrapper[4810]: I1201 16:23:47.500110 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerName="prometheus" containerID="cri-o://15ecf553d2f6db20425e4b6d1a9fd862b0af424fc27c0b249931118f089be67a" gracePeriod=600 Dec 01 16:23:47 crc kubenswrapper[4810]: I1201 16:23:47.500232 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerName="config-reloader" containerID="cri-o://ac41df1b33599bfe7b04da23ca73bb5f7d795c5312a516b31eb8fa53a417e62f" gracePeriod=600 Dec 01 16:23:47 crc kubenswrapper[4810]: I1201 16:23:47.529698 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.529675096 podStartE2EDuration="2.529675096s" podCreationTimestamp="2025-12-01 16:23:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:23:47.516752646 +0000 UTC m=+6593.280262249" watchObservedRunningTime="2025-12-01 16:23:47.529675096 +0000 UTC m=+6593.293184699" Dec 01 16:23:48 crc kubenswrapper[4810]: I1201 16:23:48.512776 4810 generic.go:334] "Generic (PLEG): container finished" podID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerID="e0308580cdf2047b1427f6efb397b8d831a9568fd6e530c3da34b986302f9b79" exitCode=0 Dec 01 16:23:48 crc kubenswrapper[4810]: I1201 16:23:48.513129 4810 generic.go:334] "Generic (PLEG): container finished" podID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerID="ac41df1b33599bfe7b04da23ca73bb5f7d795c5312a516b31eb8fa53a417e62f" exitCode=0 Dec 01 16:23:48 crc kubenswrapper[4810]: I1201 16:23:48.513140 4810 generic.go:334] "Generic (PLEG): container finished" podID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerID="15ecf553d2f6db20425e4b6d1a9fd862b0af424fc27c0b249931118f089be67a" exitCode=0 Dec 01 16:23:48 crc kubenswrapper[4810]: I1201 16:23:48.512806 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6b2c284-91c8-4419-b078-66f0dfcf46de","Type":"ContainerDied","Data":"e0308580cdf2047b1427f6efb397b8d831a9568fd6e530c3da34b986302f9b79"} Dec 01 16:23:48 crc kubenswrapper[4810]: I1201 16:23:48.513218 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6b2c284-91c8-4419-b078-66f0dfcf46de","Type":"ContainerDied","Data":"ac41df1b33599bfe7b04da23ca73bb5f7d795c5312a516b31eb8fa53a417e62f"} Dec 01 16:23:48 crc kubenswrapper[4810]: I1201 16:23:48.513235 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6b2c284-91c8-4419-b078-66f0dfcf46de","Type":"ContainerDied","Data":"15ecf553d2f6db20425e4b6d1a9fd862b0af424fc27c0b249931118f089be67a"} Dec 01 16:23:48 crc kubenswrapper[4810]: I1201 16:23:48.514674 4810 generic.go:334] "Generic (PLEG): container finished" podID="bec7da3e-6c13-4883-9c56-4e98c688c4c5" containerID="1c128fe541adca617bb9b988c393195875144309c38e34224d3ffe479f2a327f" exitCode=137 Dec 01 16:23:48 crc kubenswrapper[4810]: I1201 16:23:48.760967 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.1.142:9090/-/ready\": dial tcp 10.217.1.142:9090: connect: connection refused" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.077103 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.220175 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bec7da3e-6c13-4883-9c56-4e98c688c4c5-openstack-config-secret\") pod \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\" (UID: \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\") " Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.220232 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fglw2\" (UniqueName: \"kubernetes.io/projected/bec7da3e-6c13-4883-9c56-4e98c688c4c5-kube-api-access-fglw2\") pod \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\" (UID: \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\") " Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.220414 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bec7da3e-6c13-4883-9c56-4e98c688c4c5-openstack-config\") pod \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\" (UID: \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\") " Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.220522 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bec7da3e-6c13-4883-9c56-4e98c688c4c5-combined-ca-bundle\") pod \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\" (UID: \"bec7da3e-6c13-4883-9c56-4e98c688c4c5\") " Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.229061 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bec7da3e-6c13-4883-9c56-4e98c688c4c5-kube-api-access-fglw2" (OuterVolumeSpecName: "kube-api-access-fglw2") pod "bec7da3e-6c13-4883-9c56-4e98c688c4c5" (UID: "bec7da3e-6c13-4883-9c56-4e98c688c4c5"). InnerVolumeSpecName "kube-api-access-fglw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.264790 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bec7da3e-6c13-4883-9c56-4e98c688c4c5-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "bec7da3e-6c13-4883-9c56-4e98c688c4c5" (UID: "bec7da3e-6c13-4883-9c56-4e98c688c4c5"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.266848 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bec7da3e-6c13-4883-9c56-4e98c688c4c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bec7da3e-6c13-4883-9c56-4e98c688c4c5" (UID: "bec7da3e-6c13-4883-9c56-4e98c688c4c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.355036 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fglw2\" (UniqueName: \"kubernetes.io/projected/bec7da3e-6c13-4883-9c56-4e98c688c4c5-kube-api-access-fglw2\") on node \"crc\" DevicePath \"\"" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.355082 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bec7da3e-6c13-4883-9c56-4e98c688c4c5-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.355097 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bec7da3e-6c13-4883-9c56-4e98c688c4c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.356316 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bec7da3e-6c13-4883-9c56-4e98c688c4c5-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "bec7da3e-6c13-4883-9c56-4e98c688c4c5" (UID: "bec7da3e-6c13-4883-9c56-4e98c688c4c5"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.426561 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.458408 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bec7da3e-6c13-4883-9c56-4e98c688c4c5-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.527935 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f6b2c284-91c8-4419-b078-66f0dfcf46de","Type":"ContainerDied","Data":"c5023a0636cb8ef089d46f71d741f6f1b2b0ecbcb27cf968f599a0cf2cd0439c"} Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.528003 4810 scope.go:117] "RemoveContainer" containerID="e0308580cdf2047b1427f6efb397b8d831a9568fd6e530c3da34b986302f9b79" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.528214 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.531610 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.556301 4810 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bec7da3e-6c13-4883-9c56-4e98c688c4c5" podUID="20828f9e-5524-4c73-8754-2e1d60bb775a" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.560499 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f6b2c284-91c8-4419-b078-66f0dfcf46de-web-config\") pod \"f6b2c284-91c8-4419-b078-66f0dfcf46de\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.560805 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f6b2c284-91c8-4419-b078-66f0dfcf46de-tls-assets\") pod \"f6b2c284-91c8-4419-b078-66f0dfcf46de\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.561034 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7xzf\" (UniqueName: \"kubernetes.io/projected/f6b2c284-91c8-4419-b078-66f0dfcf46de-kube-api-access-g7xzf\") pod \"f6b2c284-91c8-4419-b078-66f0dfcf46de\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.561165 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6b2c284-91c8-4419-b078-66f0dfcf46de-config\") pod \"f6b2c284-91c8-4419-b078-66f0dfcf46de\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.561271 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f6b2c284-91c8-4419-b078-66f0dfcf46de-config-out\") pod \"f6b2c284-91c8-4419-b078-66f0dfcf46de\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.561410 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f6b2c284-91c8-4419-b078-66f0dfcf46de-prometheus-metric-storage-rulefiles-0\") pod \"f6b2c284-91c8-4419-b078-66f0dfcf46de\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.561696 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f6b2c284-91c8-4419-b078-66f0dfcf46de-thanos-prometheus-http-client-file\") pod \"f6b2c284-91c8-4419-b078-66f0dfcf46de\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.561991 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be429350-f439-458e-a105-c9855c2edb42\") pod \"f6b2c284-91c8-4419-b078-66f0dfcf46de\" (UID: \"f6b2c284-91c8-4419-b078-66f0dfcf46de\") " Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.566188 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6b2c284-91c8-4419-b078-66f0dfcf46de-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "f6b2c284-91c8-4419-b078-66f0dfcf46de" (UID: "f6b2c284-91c8-4419-b078-66f0dfcf46de"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.570311 4810 scope.go:117] "RemoveContainer" containerID="ac41df1b33599bfe7b04da23ca73bb5f7d795c5312a516b31eb8fa53a417e62f" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.571174 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b2c284-91c8-4419-b078-66f0dfcf46de-config" (OuterVolumeSpecName: "config") pod "f6b2c284-91c8-4419-b078-66f0dfcf46de" (UID: "f6b2c284-91c8-4419-b078-66f0dfcf46de"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.572033 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6b2c284-91c8-4419-b078-66f0dfcf46de-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "f6b2c284-91c8-4419-b078-66f0dfcf46de" (UID: "f6b2c284-91c8-4419-b078-66f0dfcf46de"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.573332 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b2c284-91c8-4419-b078-66f0dfcf46de-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "f6b2c284-91c8-4419-b078-66f0dfcf46de" (UID: "f6b2c284-91c8-4419-b078-66f0dfcf46de"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.574091 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6b2c284-91c8-4419-b078-66f0dfcf46de-config-out" (OuterVolumeSpecName: "config-out") pod "f6b2c284-91c8-4419-b078-66f0dfcf46de" (UID: "f6b2c284-91c8-4419-b078-66f0dfcf46de"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.575653 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6b2c284-91c8-4419-b078-66f0dfcf46de-kube-api-access-g7xzf" (OuterVolumeSpecName: "kube-api-access-g7xzf") pod "f6b2c284-91c8-4419-b078-66f0dfcf46de" (UID: "f6b2c284-91c8-4419-b078-66f0dfcf46de"). InnerVolumeSpecName "kube-api-access-g7xzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.594690 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be429350-f439-458e-a105-c9855c2edb42" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "f6b2c284-91c8-4419-b078-66f0dfcf46de" (UID: "f6b2c284-91c8-4419-b078-66f0dfcf46de"). InnerVolumeSpecName "pvc-be429350-f439-458e-a105-c9855c2edb42". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.598162 4810 scope.go:117] "RemoveContainer" containerID="15ecf553d2f6db20425e4b6d1a9fd862b0af424fc27c0b249931118f089be67a" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.609667 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b2c284-91c8-4419-b078-66f0dfcf46de-web-config" (OuterVolumeSpecName: "web-config") pod "f6b2c284-91c8-4419-b078-66f0dfcf46de" (UID: "f6b2c284-91c8-4419-b078-66f0dfcf46de"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.624916 4810 scope.go:117] "RemoveContainer" containerID="8671c51fb48482df2b04b4150d35fd4083bc424f37bf501f80446e3585f33786" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.646438 4810 scope.go:117] "RemoveContainer" containerID="1c128fe541adca617bb9b988c393195875144309c38e34224d3ffe479f2a327f" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.665870 4810 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f6b2c284-91c8-4419-b078-66f0dfcf46de-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.667175 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-be429350-f439-458e-a105-c9855c2edb42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be429350-f439-458e-a105-c9855c2edb42\") on node \"crc\" " Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.667276 4810 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f6b2c284-91c8-4419-b078-66f0dfcf46de-web-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.667296 4810 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f6b2c284-91c8-4419-b078-66f0dfcf46de-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.667310 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7xzf\" (UniqueName: \"kubernetes.io/projected/f6b2c284-91c8-4419-b078-66f0dfcf46de-kube-api-access-g7xzf\") on node \"crc\" DevicePath \"\"" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.667325 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/f6b2c284-91c8-4419-b078-66f0dfcf46de-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.667369 4810 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f6b2c284-91c8-4419-b078-66f0dfcf46de-config-out\") on node \"crc\" DevicePath \"\"" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.667399 4810 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f6b2c284-91c8-4419-b078-66f0dfcf46de-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.700745 4810 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.700962 4810 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-be429350-f439-458e-a105-c9855c2edb42" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be429350-f439-458e-a105-c9855c2edb42") on node "crc" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.768966 4810 reconciler_common.go:293] "Volume detached for volume \"pvc-be429350-f439-458e-a105-c9855c2edb42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be429350-f439-458e-a105-c9855c2edb42\") on node \"crc\" DevicePath \"\"" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.880176 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.892765 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.907106 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 16:23:49 crc kubenswrapper[4810]: E1201 16:23:49.907597 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerName="prometheus" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.907624 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerName="prometheus" Dec 01 16:23:49 crc kubenswrapper[4810]: E1201 16:23:49.907653 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerName="init-config-reloader" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.907663 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerName="init-config-reloader" Dec 01 16:23:49 crc kubenswrapper[4810]: E1201 16:23:49.907678 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerName="config-reloader" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.907687 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerName="config-reloader" Dec 01 16:23:49 crc kubenswrapper[4810]: E1201 16:23:49.907702 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerName="thanos-sidecar" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.907709 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerName="thanos-sidecar" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.907933 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerName="prometheus" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.907962 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerName="config-reloader" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.907977 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6b2c284-91c8-4419-b078-66f0dfcf46de" containerName="thanos-sidecar" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.909999 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.912825 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.913147 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.914224 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-gqh87" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.914402 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.914739 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.914906 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.936547 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 01 16:23:49 crc kubenswrapper[4810]: I1201 16:23:49.970811 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.082208 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e98b441b-4ef4-45d4-8358-ab8603841e8a-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.082525 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e98b441b-4ef4-45d4-8358-ab8603841e8a-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.082674 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e98b441b-4ef4-45d4-8358-ab8603841e8a-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.082811 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e98b441b-4ef4-45d4-8358-ab8603841e8a-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.083027 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e98b441b-4ef4-45d4-8358-ab8603841e8a-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.083218 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e98b441b-4ef4-45d4-8358-ab8603841e8a-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.083396 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-be429350-f439-458e-a105-c9855c2edb42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be429350-f439-458e-a105-c9855c2edb42\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.083863 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e98b441b-4ef4-45d4-8358-ab8603841e8a-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.084037 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e98b441b-4ef4-45d4-8358-ab8603841e8a-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.084195 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r87cg\" (UniqueName: \"kubernetes.io/projected/e98b441b-4ef4-45d4-8358-ab8603841e8a-kube-api-access-r87cg\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.084337 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e98b441b-4ef4-45d4-8358-ab8603841e8a-config\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.186214 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e98b441b-4ef4-45d4-8358-ab8603841e8a-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.186323 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e98b441b-4ef4-45d4-8358-ab8603841e8a-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.186352 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e98b441b-4ef4-45d4-8358-ab8603841e8a-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.186383 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e98b441b-4ef4-45d4-8358-ab8603841e8a-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.186426 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e98b441b-4ef4-45d4-8358-ab8603841e8a-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.186515 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e98b441b-4ef4-45d4-8358-ab8603841e8a-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.186547 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-be429350-f439-458e-a105-c9855c2edb42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be429350-f439-458e-a105-c9855c2edb42\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.186573 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e98b441b-4ef4-45d4-8358-ab8603841e8a-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.186601 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e98b441b-4ef4-45d4-8358-ab8603841e8a-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.186631 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r87cg\" (UniqueName: \"kubernetes.io/projected/e98b441b-4ef4-45d4-8358-ab8603841e8a-kube-api-access-r87cg\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.186649 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e98b441b-4ef4-45d4-8358-ab8603841e8a-config\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.188192 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e98b441b-4ef4-45d4-8358-ab8603841e8a-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.190447 4810 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.190501 4810 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-be429350-f439-458e-a105-c9855c2edb42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be429350-f439-458e-a105-c9855c2edb42\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5ec5cb75b4a42b22102ef267695ee69cf2fa8f177a6aa2457575c731cb182193/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.190596 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e98b441b-4ef4-45d4-8358-ab8603841e8a-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.192244 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e98b441b-4ef4-45d4-8358-ab8603841e8a-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.192659 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e98b441b-4ef4-45d4-8358-ab8603841e8a-config\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.194776 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e98b441b-4ef4-45d4-8358-ab8603841e8a-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.195315 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e98b441b-4ef4-45d4-8358-ab8603841e8a-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.195784 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e98b441b-4ef4-45d4-8358-ab8603841e8a-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.198782 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e98b441b-4ef4-45d4-8358-ab8603841e8a-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.206664 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r87cg\" (UniqueName: \"kubernetes.io/projected/e98b441b-4ef4-45d4-8358-ab8603841e8a-kube-api-access-r87cg\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.220199 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e98b441b-4ef4-45d4-8358-ab8603841e8a-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.232693 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-be429350-f439-458e-a105-c9855c2edb42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be429350-f439-458e-a105-c9855c2edb42\") pod \"prometheus-metric-storage-0\" (UID: \"e98b441b-4ef4-45d4-8358-ab8603841e8a\") " pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.323969 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.512281 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bec7da3e-6c13-4883-9c56-4e98c688c4c5" path="/var/lib/kubelet/pods/bec7da3e-6c13-4883-9c56-4e98c688c4c5/volumes" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.513586 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6b2c284-91c8-4419-b078-66f0dfcf46de" path="/var/lib/kubelet/pods/f6b2c284-91c8-4419-b078-66f0dfcf46de/volumes" Dec 01 16:23:50 crc kubenswrapper[4810]: I1201 16:23:50.855204 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 16:23:50 crc kubenswrapper[4810]: W1201 16:23:50.864435 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode98b441b_4ef4_45d4_8358_ab8603841e8a.slice/crio-6aac4067ca89112e9ff22538fc842b3100eee2fb4757173f9d9608de9725b13c WatchSource:0}: Error finding container 6aac4067ca89112e9ff22538fc842b3100eee2fb4757173f9d9608de9725b13c: Status 404 returned error can't find the container with id 6aac4067ca89112e9ff22538fc842b3100eee2fb4757173f9d9608de9725b13c Dec 01 16:23:51 crc kubenswrapper[4810]: I1201 16:23:51.562999 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e98b441b-4ef4-45d4-8358-ab8603841e8a","Type":"ContainerStarted","Data":"6aac4067ca89112e9ff22538fc842b3100eee2fb4757173f9d9608de9725b13c"} Dec 01 16:23:52 crc kubenswrapper[4810]: I1201 16:23:52.879740 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 16:23:52 crc kubenswrapper[4810]: I1201 16:23:52.884365 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 16:23:52 crc kubenswrapper[4810]: I1201 16:23:52.886401 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 16:23:52 crc kubenswrapper[4810]: I1201 16:23:52.895620 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 16:23:52 crc kubenswrapper[4810]: I1201 16:23:52.896570 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.046353 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-2ca7-account-create-update-fkfmv"] Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.053059 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db82ec37-860a-4baa-9dcd-e967dd687658-run-httpd\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.053265 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.053582 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qvf9\" (UniqueName: \"kubernetes.io/projected/db82ec37-860a-4baa-9dcd-e967dd687658-kube-api-access-9qvf9\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.053694 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.053730 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db82ec37-860a-4baa-9dcd-e967dd687658-log-httpd\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.053856 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-config-data\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.053898 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-scripts\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.055680 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-2zlzh"] Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.063411 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-2ca7-account-create-update-fkfmv"] Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.070844 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-2zlzh"] Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.155570 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-config-data\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.156654 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-scripts\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.156706 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db82ec37-860a-4baa-9dcd-e967dd687658-run-httpd\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.156756 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.156842 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qvf9\" (UniqueName: \"kubernetes.io/projected/db82ec37-860a-4baa-9dcd-e967dd687658-kube-api-access-9qvf9\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.156880 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.156902 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db82ec37-860a-4baa-9dcd-e967dd687658-log-httpd\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.157253 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db82ec37-860a-4baa-9dcd-e967dd687658-log-httpd\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.157245 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db82ec37-860a-4baa-9dcd-e967dd687658-run-httpd\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.163268 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.163869 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.164620 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-scripts\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.177391 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qvf9\" (UniqueName: \"kubernetes.io/projected/db82ec37-860a-4baa-9dcd-e967dd687658-kube-api-access-9qvf9\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.251765 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-config-data\") pod \"ceilometer-0\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.269234 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 16:23:53 crc kubenswrapper[4810]: W1201 16:23:53.788893 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb82ec37_860a_4baa_9dcd_e967dd687658.slice/crio-545e315c4b9d9c8780713e2d214bede795b918aa9ccdcd82467cd835790947b9 WatchSource:0}: Error finding container 545e315c4b9d9c8780713e2d214bede795b918aa9ccdcd82467cd835790947b9: Status 404 returned error can't find the container with id 545e315c4b9d9c8780713e2d214bede795b918aa9ccdcd82467cd835790947b9 Dec 01 16:23:53 crc kubenswrapper[4810]: I1201 16:23:53.790958 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 16:23:54 crc kubenswrapper[4810]: I1201 16:23:54.509356 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c1aa2aa-21be-4db1-9b97-552b303cfa57" path="/var/lib/kubelet/pods/1c1aa2aa-21be-4db1-9b97-552b303cfa57/volumes" Dec 01 16:23:54 crc kubenswrapper[4810]: I1201 16:23:54.511019 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0f1746a-f775-4c67-9122-903cd261dc87" path="/var/lib/kubelet/pods/b0f1746a-f775-4c67-9122-903cd261dc87/volumes" Dec 01 16:23:54 crc kubenswrapper[4810]: I1201 16:23:54.597573 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e98b441b-4ef4-45d4-8358-ab8603841e8a","Type":"ContainerStarted","Data":"faf77038808bf4aa9b1b1921c43b6133dd877d710cc368b40b114abb5bfa2323"} Dec 01 16:23:54 crc kubenswrapper[4810]: I1201 16:23:54.598829 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db82ec37-860a-4baa-9dcd-e967dd687658","Type":"ContainerStarted","Data":"545e315c4b9d9c8780713e2d214bede795b918aa9ccdcd82467cd835790947b9"} Dec 01 16:23:55 crc kubenswrapper[4810]: I1201 16:23:55.609823 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db82ec37-860a-4baa-9dcd-e967dd687658","Type":"ContainerStarted","Data":"b5af3a32078ae23ef4575623ef13f026da5ca28468f6e916be9439972df2e4f3"} Dec 01 16:23:55 crc kubenswrapper[4810]: I1201 16:23:55.610144 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db82ec37-860a-4baa-9dcd-e967dd687658","Type":"ContainerStarted","Data":"806cd5b011ecf8247f10aa1d1790e394521c847694025998395ed54776b43d6b"} Dec 01 16:23:57 crc kubenswrapper[4810]: I1201 16:23:57.628573 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db82ec37-860a-4baa-9dcd-e967dd687658","Type":"ContainerStarted","Data":"429c0b9d412531309efc726969a27406fe1de8537dc4abb6524ab729d17c8546"} Dec 01 16:23:58 crc kubenswrapper[4810]: I1201 16:23:58.669555 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db82ec37-860a-4baa-9dcd-e967dd687658","Type":"ContainerStarted","Data":"21231bb387042290f49bf476bb2ca9e18537490fe26b9a023815f2b32c96378d"} Dec 01 16:23:58 crc kubenswrapper[4810]: I1201 16:23:58.671102 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 16:23:58 crc kubenswrapper[4810]: I1201 16:23:58.703319 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.127891379 podStartE2EDuration="6.703295687s" podCreationTimestamp="2025-12-01 16:23:52 +0000 UTC" firstStartedPulling="2025-12-01 16:23:53.793829729 +0000 UTC m=+6599.557339332" lastFinishedPulling="2025-12-01 16:23:58.369234037 +0000 UTC m=+6604.132743640" observedRunningTime="2025-12-01 16:23:58.688313959 +0000 UTC m=+6604.451823562" watchObservedRunningTime="2025-12-01 16:23:58.703295687 +0000 UTC m=+6604.466805290" Dec 01 16:23:59 crc kubenswrapper[4810]: I1201 16:23:59.041179 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-dwrp5"] Dec 01 16:23:59 crc kubenswrapper[4810]: I1201 16:23:59.052901 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-dwrp5"] Dec 01 16:24:00 crc kubenswrapper[4810]: I1201 16:24:00.527052 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84039e04-f053-496f-921a-18e12ff432e1" path="/var/lib/kubelet/pods/84039e04-f053-496f-921a-18e12ff432e1/volumes" Dec 01 16:24:00 crc kubenswrapper[4810]: I1201 16:24:00.687255 4810 generic.go:334] "Generic (PLEG): container finished" podID="e98b441b-4ef4-45d4-8358-ab8603841e8a" containerID="faf77038808bf4aa9b1b1921c43b6133dd877d710cc368b40b114abb5bfa2323" exitCode=0 Dec 01 16:24:00 crc kubenswrapper[4810]: I1201 16:24:00.687301 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e98b441b-4ef4-45d4-8358-ab8603841e8a","Type":"ContainerDied","Data":"faf77038808bf4aa9b1b1921c43b6133dd877d710cc368b40b114abb5bfa2323"} Dec 01 16:24:01 crc kubenswrapper[4810]: I1201 16:24:01.710034 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e98b441b-4ef4-45d4-8358-ab8603841e8a","Type":"ContainerStarted","Data":"256ef834930e41865da8dccf0fd82716f68803aaa4c2aac60b8b0f3dd22ad1c0"} Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.302818 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-vgp88"] Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.304600 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-vgp88" Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.315517 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-vgp88"] Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.392675 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r8c9\" (UniqueName: \"kubernetes.io/projected/bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e-kube-api-access-8r8c9\") pod \"aodh-db-create-vgp88\" (UID: \"bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e\") " pod="openstack/aodh-db-create-vgp88" Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.392854 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e-operator-scripts\") pod \"aodh-db-create-vgp88\" (UID: \"bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e\") " pod="openstack/aodh-db-create-vgp88" Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.409416 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-2833-account-create-update-x4f4n"] Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.410966 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-2833-account-create-update-x4f4n" Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.414059 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.420968 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-2833-account-create-update-x4f4n"] Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.502945 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e-operator-scripts\") pod \"aodh-db-create-vgp88\" (UID: \"bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e\") " pod="openstack/aodh-db-create-vgp88" Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.503186 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0b4462a-f09b-4d97-9069-c8b248855311-operator-scripts\") pod \"aodh-2833-account-create-update-x4f4n\" (UID: \"b0b4462a-f09b-4d97-9069-c8b248855311\") " pod="openstack/aodh-2833-account-create-update-x4f4n" Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.503232 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvn8b\" (UniqueName: \"kubernetes.io/projected/b0b4462a-f09b-4d97-9069-c8b248855311-kube-api-access-wvn8b\") pod \"aodh-2833-account-create-update-x4f4n\" (UID: \"b0b4462a-f09b-4d97-9069-c8b248855311\") " pod="openstack/aodh-2833-account-create-update-x4f4n" Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.503282 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r8c9\" (UniqueName: \"kubernetes.io/projected/bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e-kube-api-access-8r8c9\") pod \"aodh-db-create-vgp88\" (UID: \"bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e\") " pod="openstack/aodh-db-create-vgp88" Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.505056 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e-operator-scripts\") pod \"aodh-db-create-vgp88\" (UID: \"bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e\") " pod="openstack/aodh-db-create-vgp88" Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.528927 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r8c9\" (UniqueName: \"kubernetes.io/projected/bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e-kube-api-access-8r8c9\") pod \"aodh-db-create-vgp88\" (UID: \"bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e\") " pod="openstack/aodh-db-create-vgp88" Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.606832 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0b4462a-f09b-4d97-9069-c8b248855311-operator-scripts\") pod \"aodh-2833-account-create-update-x4f4n\" (UID: \"b0b4462a-f09b-4d97-9069-c8b248855311\") " pod="openstack/aodh-2833-account-create-update-x4f4n" Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.606890 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvn8b\" (UniqueName: \"kubernetes.io/projected/b0b4462a-f09b-4d97-9069-c8b248855311-kube-api-access-wvn8b\") pod \"aodh-2833-account-create-update-x4f4n\" (UID: \"b0b4462a-f09b-4d97-9069-c8b248855311\") " pod="openstack/aodh-2833-account-create-update-x4f4n" Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.608214 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0b4462a-f09b-4d97-9069-c8b248855311-operator-scripts\") pod \"aodh-2833-account-create-update-x4f4n\" (UID: \"b0b4462a-f09b-4d97-9069-c8b248855311\") " pod="openstack/aodh-2833-account-create-update-x4f4n" Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.627997 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-vgp88" Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.650532 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvn8b\" (UniqueName: \"kubernetes.io/projected/b0b4462a-f09b-4d97-9069-c8b248855311-kube-api-access-wvn8b\") pod \"aodh-2833-account-create-update-x4f4n\" (UID: \"b0b4462a-f09b-4d97-9069-c8b248855311\") " pod="openstack/aodh-2833-account-create-update-x4f4n" Dec 01 16:24:03 crc kubenswrapper[4810]: I1201 16:24:03.750852 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-2833-account-create-update-x4f4n" Dec 01 16:24:04 crc kubenswrapper[4810]: I1201 16:24:04.113048 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-vgp88"] Dec 01 16:24:04 crc kubenswrapper[4810]: I1201 16:24:04.288771 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-2833-account-create-update-x4f4n"] Dec 01 16:24:04 crc kubenswrapper[4810]: W1201 16:24:04.554817 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbcd2b18_92ab_425c_bcc9_00c8eacd7c6e.slice/crio-497acf4f3126d9d2395c8638f172fc687941d89ea9e89cc76fdcbc2ff637c7bd WatchSource:0}: Error finding container 497acf4f3126d9d2395c8638f172fc687941d89ea9e89cc76fdcbc2ff637c7bd: Status 404 returned error can't find the container with id 497acf4f3126d9d2395c8638f172fc687941d89ea9e89cc76fdcbc2ff637c7bd Dec 01 16:24:04 crc kubenswrapper[4810]: I1201 16:24:04.738137 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-vgp88" event={"ID":"bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e","Type":"ContainerStarted","Data":"497acf4f3126d9d2395c8638f172fc687941d89ea9e89cc76fdcbc2ff637c7bd"} Dec 01 16:24:04 crc kubenswrapper[4810]: I1201 16:24:04.739163 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-2833-account-create-update-x4f4n" event={"ID":"b0b4462a-f09b-4d97-9069-c8b248855311","Type":"ContainerStarted","Data":"1dce37f1319d1941934416acfb1846b4fa6678a7f7a403b8e402d981a0a701a6"} Dec 01 16:24:05 crc kubenswrapper[4810]: I1201 16:24:05.750321 4810 generic.go:334] "Generic (PLEG): container finished" podID="b0b4462a-f09b-4d97-9069-c8b248855311" containerID="e53c4f4ffada9a1df56e799ff33e717732604ff8999d584f19a7aeff4729156e" exitCode=0 Dec 01 16:24:05 crc kubenswrapper[4810]: I1201 16:24:05.750503 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-2833-account-create-update-x4f4n" event={"ID":"b0b4462a-f09b-4d97-9069-c8b248855311","Type":"ContainerDied","Data":"e53c4f4ffada9a1df56e799ff33e717732604ff8999d584f19a7aeff4729156e"} Dec 01 16:24:05 crc kubenswrapper[4810]: I1201 16:24:05.752403 4810 generic.go:334] "Generic (PLEG): container finished" podID="bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e" containerID="09a9240e4da76772e26368cb6bc54742aa9b3279565331075191dc6d9408af3b" exitCode=0 Dec 01 16:24:05 crc kubenswrapper[4810]: I1201 16:24:05.752443 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-vgp88" event={"ID":"bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e","Type":"ContainerDied","Data":"09a9240e4da76772e26368cb6bc54742aa9b3279565331075191dc6d9408af3b"} Dec 01 16:24:06 crc kubenswrapper[4810]: I1201 16:24:06.769398 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e98b441b-4ef4-45d4-8358-ab8603841e8a","Type":"ContainerStarted","Data":"b37d35a900a90aaeb113b1e5b373c0ebcc86f0f5b251035472b0529a7d96df00"} Dec 01 16:24:06 crc kubenswrapper[4810]: I1201 16:24:06.769952 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e98b441b-4ef4-45d4-8358-ab8603841e8a","Type":"ContainerStarted","Data":"876d3b1e80741f86e2c9eaf5fdb2d6c272addff854d399165780637372135173"} Dec 01 16:24:06 crc kubenswrapper[4810]: I1201 16:24:06.823656 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=17.823621308 podStartE2EDuration="17.823621308s" podCreationTimestamp="2025-12-01 16:23:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:24:06.80343892 +0000 UTC m=+6612.566948563" watchObservedRunningTime="2025-12-01 16:24:06.823621308 +0000 UTC m=+6612.587130951" Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.293894 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-vgp88" Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.304287 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-2833-account-create-update-x4f4n" Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.393885 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0b4462a-f09b-4d97-9069-c8b248855311-operator-scripts\") pod \"b0b4462a-f09b-4d97-9069-c8b248855311\" (UID: \"b0b4462a-f09b-4d97-9069-c8b248855311\") " Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.394162 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e-operator-scripts\") pod \"bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e\" (UID: \"bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e\") " Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.394425 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvn8b\" (UniqueName: \"kubernetes.io/projected/b0b4462a-f09b-4d97-9069-c8b248855311-kube-api-access-wvn8b\") pod \"b0b4462a-f09b-4d97-9069-c8b248855311\" (UID: \"b0b4462a-f09b-4d97-9069-c8b248855311\") " Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.394716 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r8c9\" (UniqueName: \"kubernetes.io/projected/bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e-kube-api-access-8r8c9\") pod \"bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e\" (UID: \"bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e\") " Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.394447 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0b4462a-f09b-4d97-9069-c8b248855311-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b0b4462a-f09b-4d97-9069-c8b248855311" (UID: "b0b4462a-f09b-4d97-9069-c8b248855311"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.394761 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e" (UID: "bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.395592 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0b4462a-f09b-4d97-9069-c8b248855311-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.395650 4810 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.400157 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0b4462a-f09b-4d97-9069-c8b248855311-kube-api-access-wvn8b" (OuterVolumeSpecName: "kube-api-access-wvn8b") pod "b0b4462a-f09b-4d97-9069-c8b248855311" (UID: "b0b4462a-f09b-4d97-9069-c8b248855311"). InnerVolumeSpecName "kube-api-access-wvn8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.400670 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e-kube-api-access-8r8c9" (OuterVolumeSpecName: "kube-api-access-8r8c9") pod "bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e" (UID: "bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e"). InnerVolumeSpecName "kube-api-access-8r8c9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.497981 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvn8b\" (UniqueName: \"kubernetes.io/projected/b0b4462a-f09b-4d97-9069-c8b248855311-kube-api-access-wvn8b\") on node \"crc\" DevicePath \"\"" Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.498023 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8r8c9\" (UniqueName: \"kubernetes.io/projected/bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e-kube-api-access-8r8c9\") on node \"crc\" DevicePath \"\"" Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.780918 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-vgp88" Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.780929 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-vgp88" event={"ID":"bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e","Type":"ContainerDied","Data":"497acf4f3126d9d2395c8638f172fc687941d89ea9e89cc76fdcbc2ff637c7bd"} Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.781301 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="497acf4f3126d9d2395c8638f172fc687941d89ea9e89cc76fdcbc2ff637c7bd" Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.783324 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-2833-account-create-update-x4f4n" event={"ID":"b0b4462a-f09b-4d97-9069-c8b248855311","Type":"ContainerDied","Data":"1dce37f1319d1941934416acfb1846b4fa6678a7f7a403b8e402d981a0a701a6"} Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.783350 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-2833-account-create-update-x4f4n" Dec 01 16:24:07 crc kubenswrapper[4810]: I1201 16:24:07.783389 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dce37f1319d1941934416acfb1846b4fa6678a7f7a403b8e402d981a0a701a6" Dec 01 16:24:08 crc kubenswrapper[4810]: I1201 16:24:08.895978 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-t8xft"] Dec 01 16:24:08 crc kubenswrapper[4810]: E1201 16:24:08.896412 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e" containerName="mariadb-database-create" Dec 01 16:24:08 crc kubenswrapper[4810]: I1201 16:24:08.896423 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e" containerName="mariadb-database-create" Dec 01 16:24:08 crc kubenswrapper[4810]: E1201 16:24:08.896446 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0b4462a-f09b-4d97-9069-c8b248855311" containerName="mariadb-account-create-update" Dec 01 16:24:08 crc kubenswrapper[4810]: I1201 16:24:08.896452 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0b4462a-f09b-4d97-9069-c8b248855311" containerName="mariadb-account-create-update" Dec 01 16:24:08 crc kubenswrapper[4810]: I1201 16:24:08.896663 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e" containerName="mariadb-database-create" Dec 01 16:24:08 crc kubenswrapper[4810]: I1201 16:24:08.896688 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0b4462a-f09b-4d97-9069-c8b248855311" containerName="mariadb-account-create-update" Dec 01 16:24:08 crc kubenswrapper[4810]: I1201 16:24:08.897337 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-t8xft" Dec 01 16:24:08 crc kubenswrapper[4810]: I1201 16:24:08.899238 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 01 16:24:08 crc kubenswrapper[4810]: I1201 16:24:08.899909 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-js2tk" Dec 01 16:24:08 crc kubenswrapper[4810]: I1201 16:24:08.900185 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 16:24:08 crc kubenswrapper[4810]: I1201 16:24:08.900555 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 01 16:24:08 crc kubenswrapper[4810]: I1201 16:24:08.908524 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-t8xft"] Dec 01 16:24:09 crc kubenswrapper[4810]: I1201 16:24:09.030681 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tl46\" (UniqueName: \"kubernetes.io/projected/85bae6bb-65c7-4004-99a3-d41d528f1be3-kube-api-access-2tl46\") pod \"aodh-db-sync-t8xft\" (UID: \"85bae6bb-65c7-4004-99a3-d41d528f1be3\") " pod="openstack/aodh-db-sync-t8xft" Dec 01 16:24:09 crc kubenswrapper[4810]: I1201 16:24:09.030742 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85bae6bb-65c7-4004-99a3-d41d528f1be3-config-data\") pod \"aodh-db-sync-t8xft\" (UID: \"85bae6bb-65c7-4004-99a3-d41d528f1be3\") " pod="openstack/aodh-db-sync-t8xft" Dec 01 16:24:09 crc kubenswrapper[4810]: I1201 16:24:09.030777 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85bae6bb-65c7-4004-99a3-d41d528f1be3-combined-ca-bundle\") pod \"aodh-db-sync-t8xft\" (UID: \"85bae6bb-65c7-4004-99a3-d41d528f1be3\") " pod="openstack/aodh-db-sync-t8xft" Dec 01 16:24:09 crc kubenswrapper[4810]: I1201 16:24:09.030902 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85bae6bb-65c7-4004-99a3-d41d528f1be3-scripts\") pod \"aodh-db-sync-t8xft\" (UID: \"85bae6bb-65c7-4004-99a3-d41d528f1be3\") " pod="openstack/aodh-db-sync-t8xft" Dec 01 16:24:09 crc kubenswrapper[4810]: I1201 16:24:09.133185 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tl46\" (UniqueName: \"kubernetes.io/projected/85bae6bb-65c7-4004-99a3-d41d528f1be3-kube-api-access-2tl46\") pod \"aodh-db-sync-t8xft\" (UID: \"85bae6bb-65c7-4004-99a3-d41d528f1be3\") " pod="openstack/aodh-db-sync-t8xft" Dec 01 16:24:09 crc kubenswrapper[4810]: I1201 16:24:09.133246 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85bae6bb-65c7-4004-99a3-d41d528f1be3-config-data\") pod \"aodh-db-sync-t8xft\" (UID: \"85bae6bb-65c7-4004-99a3-d41d528f1be3\") " pod="openstack/aodh-db-sync-t8xft" Dec 01 16:24:09 crc kubenswrapper[4810]: I1201 16:24:09.133284 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85bae6bb-65c7-4004-99a3-d41d528f1be3-combined-ca-bundle\") pod \"aodh-db-sync-t8xft\" (UID: \"85bae6bb-65c7-4004-99a3-d41d528f1be3\") " pod="openstack/aodh-db-sync-t8xft" Dec 01 16:24:09 crc kubenswrapper[4810]: I1201 16:24:09.133422 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85bae6bb-65c7-4004-99a3-d41d528f1be3-scripts\") pod \"aodh-db-sync-t8xft\" (UID: \"85bae6bb-65c7-4004-99a3-d41d528f1be3\") " pod="openstack/aodh-db-sync-t8xft" Dec 01 16:24:09 crc kubenswrapper[4810]: I1201 16:24:09.141772 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85bae6bb-65c7-4004-99a3-d41d528f1be3-config-data\") pod \"aodh-db-sync-t8xft\" (UID: \"85bae6bb-65c7-4004-99a3-d41d528f1be3\") " pod="openstack/aodh-db-sync-t8xft" Dec 01 16:24:09 crc kubenswrapper[4810]: I1201 16:24:09.142073 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85bae6bb-65c7-4004-99a3-d41d528f1be3-scripts\") pod \"aodh-db-sync-t8xft\" (UID: \"85bae6bb-65c7-4004-99a3-d41d528f1be3\") " pod="openstack/aodh-db-sync-t8xft" Dec 01 16:24:09 crc kubenswrapper[4810]: I1201 16:24:09.142417 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85bae6bb-65c7-4004-99a3-d41d528f1be3-combined-ca-bundle\") pod \"aodh-db-sync-t8xft\" (UID: \"85bae6bb-65c7-4004-99a3-d41d528f1be3\") " pod="openstack/aodh-db-sync-t8xft" Dec 01 16:24:09 crc kubenswrapper[4810]: I1201 16:24:09.162572 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tl46\" (UniqueName: \"kubernetes.io/projected/85bae6bb-65c7-4004-99a3-d41d528f1be3-kube-api-access-2tl46\") pod \"aodh-db-sync-t8xft\" (UID: \"85bae6bb-65c7-4004-99a3-d41d528f1be3\") " pod="openstack/aodh-db-sync-t8xft" Dec 01 16:24:09 crc kubenswrapper[4810]: I1201 16:24:09.219714 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-t8xft" Dec 01 16:24:09 crc kubenswrapper[4810]: I1201 16:24:09.803910 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-t8xft"] Dec 01 16:24:10 crc kubenswrapper[4810]: I1201 16:24:10.324634 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 01 16:24:10 crc kubenswrapper[4810]: I1201 16:24:10.828502 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-t8xft" event={"ID":"85bae6bb-65c7-4004-99a3-d41d528f1be3","Type":"ContainerStarted","Data":"249388c29780cd265d256bf7b150e87352aeeebe5acda9f34fba5295ab4b4393"} Dec 01 16:24:12 crc kubenswrapper[4810]: I1201 16:24:12.281982 4810 scope.go:117] "RemoveContainer" containerID="ac95e8544d2d9d177ad553c30d64e480d85ac3dfaf5c112025a6760b5aab16a2" Dec 01 16:24:14 crc kubenswrapper[4810]: I1201 16:24:14.178536 4810 scope.go:117] "RemoveContainer" containerID="2ffb281642b4705e534267c46825b7a1113616d4e1a85b7062d9692bd49640b1" Dec 01 16:24:14 crc kubenswrapper[4810]: I1201 16:24:14.227133 4810 scope.go:117] "RemoveContainer" containerID="bb9403ef5d0fd0c4f80ae6cc912979f62b9a82165899bfe05216834631ccc2a8" Dec 01 16:24:14 crc kubenswrapper[4810]: I1201 16:24:14.373596 4810 scope.go:117] "RemoveContainer" containerID="21be20ccde4a34f9f5035ce0cc253e3abf1c6cea80658e626418f9db0ee7a24c" Dec 01 16:24:14 crc kubenswrapper[4810]: I1201 16:24:14.418793 4810 scope.go:117] "RemoveContainer" containerID="bc86f9e27fdb7f8ebcdb878b0e8d795189133c8c8c3c46b170bd0b3aae36e026" Dec 01 16:24:14 crc kubenswrapper[4810]: I1201 16:24:14.471244 4810 scope.go:117] "RemoveContainer" containerID="b17090bed39c4e268f652595deba6c7b1d0907729166a71e93979d05ea4332ec" Dec 01 16:24:14 crc kubenswrapper[4810]: I1201 16:24:14.496681 4810 scope.go:117] "RemoveContainer" containerID="d53fe94224d2bd2ea49669d0f69129401ffd965650a924d2f4da3e096289347d" Dec 01 16:24:14 crc kubenswrapper[4810]: I1201 16:24:14.564008 4810 scope.go:117] "RemoveContainer" containerID="c4abe3f0f775cd77dd7ee7c754d96148cc205726254d214cbbb8fbbccc4f53be" Dec 01 16:24:14 crc kubenswrapper[4810]: I1201 16:24:14.877462 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-t8xft" event={"ID":"85bae6bb-65c7-4004-99a3-d41d528f1be3","Type":"ContainerStarted","Data":"04a2372a0b8b8c0e8f34fafcd609ae79e2b5c96f79520cbcc7412d88f535df6f"} Dec 01 16:24:14 crc kubenswrapper[4810]: I1201 16:24:14.897160 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-t8xft" podStartSLOduration=2.494061638 podStartE2EDuration="6.897140118s" podCreationTimestamp="2025-12-01 16:24:08 +0000 UTC" firstStartedPulling="2025-12-01 16:24:09.825792277 +0000 UTC m=+6615.589301880" lastFinishedPulling="2025-12-01 16:24:14.228870757 +0000 UTC m=+6619.992380360" observedRunningTime="2025-12-01 16:24:14.89019678 +0000 UTC m=+6620.653706383" watchObservedRunningTime="2025-12-01 16:24:14.897140118 +0000 UTC m=+6620.660649721" Dec 01 16:24:17 crc kubenswrapper[4810]: I1201 16:24:17.912126 4810 generic.go:334] "Generic (PLEG): container finished" podID="85bae6bb-65c7-4004-99a3-d41d528f1be3" containerID="04a2372a0b8b8c0e8f34fafcd609ae79e2b5c96f79520cbcc7412d88f535df6f" exitCode=0 Dec 01 16:24:17 crc kubenswrapper[4810]: I1201 16:24:17.912241 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-t8xft" event={"ID":"85bae6bb-65c7-4004-99a3-d41d528f1be3","Type":"ContainerDied","Data":"04a2372a0b8b8c0e8f34fafcd609ae79e2b5c96f79520cbcc7412d88f535df6f"} Dec 01 16:24:19 crc kubenswrapper[4810]: I1201 16:24:19.303570 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-t8xft" Dec 01 16:24:19 crc kubenswrapper[4810]: I1201 16:24:19.470008 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85bae6bb-65c7-4004-99a3-d41d528f1be3-config-data\") pod \"85bae6bb-65c7-4004-99a3-d41d528f1be3\" (UID: \"85bae6bb-65c7-4004-99a3-d41d528f1be3\") " Dec 01 16:24:19 crc kubenswrapper[4810]: I1201 16:24:19.470118 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tl46\" (UniqueName: \"kubernetes.io/projected/85bae6bb-65c7-4004-99a3-d41d528f1be3-kube-api-access-2tl46\") pod \"85bae6bb-65c7-4004-99a3-d41d528f1be3\" (UID: \"85bae6bb-65c7-4004-99a3-d41d528f1be3\") " Dec 01 16:24:19 crc kubenswrapper[4810]: I1201 16:24:19.470137 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85bae6bb-65c7-4004-99a3-d41d528f1be3-scripts\") pod \"85bae6bb-65c7-4004-99a3-d41d528f1be3\" (UID: \"85bae6bb-65c7-4004-99a3-d41d528f1be3\") " Dec 01 16:24:19 crc kubenswrapper[4810]: I1201 16:24:19.470214 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85bae6bb-65c7-4004-99a3-d41d528f1be3-combined-ca-bundle\") pod \"85bae6bb-65c7-4004-99a3-d41d528f1be3\" (UID: \"85bae6bb-65c7-4004-99a3-d41d528f1be3\") " Dec 01 16:24:19 crc kubenswrapper[4810]: I1201 16:24:19.475886 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85bae6bb-65c7-4004-99a3-d41d528f1be3-kube-api-access-2tl46" (OuterVolumeSpecName: "kube-api-access-2tl46") pod "85bae6bb-65c7-4004-99a3-d41d528f1be3" (UID: "85bae6bb-65c7-4004-99a3-d41d528f1be3"). InnerVolumeSpecName "kube-api-access-2tl46". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:24:19 crc kubenswrapper[4810]: I1201 16:24:19.483067 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85bae6bb-65c7-4004-99a3-d41d528f1be3-scripts" (OuterVolumeSpecName: "scripts") pod "85bae6bb-65c7-4004-99a3-d41d528f1be3" (UID: "85bae6bb-65c7-4004-99a3-d41d528f1be3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:24:19 crc kubenswrapper[4810]: I1201 16:24:19.516434 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85bae6bb-65c7-4004-99a3-d41d528f1be3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85bae6bb-65c7-4004-99a3-d41d528f1be3" (UID: "85bae6bb-65c7-4004-99a3-d41d528f1be3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:24:19 crc kubenswrapper[4810]: I1201 16:24:19.516485 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85bae6bb-65c7-4004-99a3-d41d528f1be3-config-data" (OuterVolumeSpecName: "config-data") pod "85bae6bb-65c7-4004-99a3-d41d528f1be3" (UID: "85bae6bb-65c7-4004-99a3-d41d528f1be3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:24:19 crc kubenswrapper[4810]: I1201 16:24:19.573229 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85bae6bb-65c7-4004-99a3-d41d528f1be3-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:24:19 crc kubenswrapper[4810]: I1201 16:24:19.573271 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tl46\" (UniqueName: \"kubernetes.io/projected/85bae6bb-65c7-4004-99a3-d41d528f1be3-kube-api-access-2tl46\") on node \"crc\" DevicePath \"\"" Dec 01 16:24:19 crc kubenswrapper[4810]: I1201 16:24:19.573283 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85bae6bb-65c7-4004-99a3-d41d528f1be3-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:24:19 crc kubenswrapper[4810]: I1201 16:24:19.573291 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85bae6bb-65c7-4004-99a3-d41d528f1be3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:24:19 crc kubenswrapper[4810]: I1201 16:24:19.938198 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-t8xft" event={"ID":"85bae6bb-65c7-4004-99a3-d41d528f1be3","Type":"ContainerDied","Data":"249388c29780cd265d256bf7b150e87352aeeebe5acda9f34fba5295ab4b4393"} Dec 01 16:24:19 crc kubenswrapper[4810]: I1201 16:24:19.938249 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-t8xft" Dec 01 16:24:19 crc kubenswrapper[4810]: I1201 16:24:19.938257 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="249388c29780cd265d256bf7b150e87352aeeebe5acda9f34fba5295ab4b4393" Dec 01 16:24:20 crc kubenswrapper[4810]: I1201 16:24:20.324489 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 01 16:24:20 crc kubenswrapper[4810]: I1201 16:24:20.334999 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 01 16:24:20 crc kubenswrapper[4810]: I1201 16:24:20.952904 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 01 16:24:23 crc kubenswrapper[4810]: I1201 16:24:23.275325 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.030067 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 01 16:24:24 crc kubenswrapper[4810]: E1201 16:24:24.031083 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85bae6bb-65c7-4004-99a3-d41d528f1be3" containerName="aodh-db-sync" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.031111 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="85bae6bb-65c7-4004-99a3-d41d528f1be3" containerName="aodh-db-sync" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.031393 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="85bae6bb-65c7-4004-99a3-d41d528f1be3" containerName="aodh-db-sync" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.033801 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.038798 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-js2tk" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.039448 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.039603 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.058976 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.113807 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v694p\" (UniqueName: \"kubernetes.io/projected/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-kube-api-access-v694p\") pod \"aodh-0\" (UID: \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\") " pod="openstack/aodh-0" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.113902 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\") " pod="openstack/aodh-0" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.113945 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-config-data\") pod \"aodh-0\" (UID: \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\") " pod="openstack/aodh-0" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.113981 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-scripts\") pod \"aodh-0\" (UID: \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\") " pod="openstack/aodh-0" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.216039 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\") " pod="openstack/aodh-0" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.216119 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-config-data\") pod \"aodh-0\" (UID: \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\") " pod="openstack/aodh-0" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.216271 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-scripts\") pod \"aodh-0\" (UID: \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\") " pod="openstack/aodh-0" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.216380 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v694p\" (UniqueName: \"kubernetes.io/projected/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-kube-api-access-v694p\") pod \"aodh-0\" (UID: \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\") " pod="openstack/aodh-0" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.227745 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-config-data\") pod \"aodh-0\" (UID: \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\") " pod="openstack/aodh-0" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.232610 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-scripts\") pod \"aodh-0\" (UID: \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\") " pod="openstack/aodh-0" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.235081 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\") " pod="openstack/aodh-0" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.238190 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v694p\" (UniqueName: \"kubernetes.io/projected/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-kube-api-access-v694p\") pod \"aodh-0\" (UID: \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\") " pod="openstack/aodh-0" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.370699 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 16:24:24 crc kubenswrapper[4810]: I1201 16:24:24.921570 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 16:24:25 crc kubenswrapper[4810]: I1201 16:24:25.011293 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1","Type":"ContainerStarted","Data":"89efac2d97004dc6229b3922cedb754555ccb061bd49cc316c6838cd861d5a83"} Dec 01 16:24:26 crc kubenswrapper[4810]: I1201 16:24:26.022489 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1","Type":"ContainerStarted","Data":"5449f9949a4c937ef3e6c920d57228d915efeb6e7f2b4523f10ca5b42af150cd"} Dec 01 16:24:27 crc kubenswrapper[4810]: I1201 16:24:27.981627 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 16:24:27 crc kubenswrapper[4810]: I1201 16:24:27.982689 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db82ec37-860a-4baa-9dcd-e967dd687658" containerName="ceilometer-central-agent" containerID="cri-o://806cd5b011ecf8247f10aa1d1790e394521c847694025998395ed54776b43d6b" gracePeriod=30 Dec 01 16:24:27 crc kubenswrapper[4810]: I1201 16:24:27.982742 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db82ec37-860a-4baa-9dcd-e967dd687658" containerName="proxy-httpd" containerID="cri-o://21231bb387042290f49bf476bb2ca9e18537490fe26b9a023815f2b32c96378d" gracePeriod=30 Dec 01 16:24:27 crc kubenswrapper[4810]: I1201 16:24:27.982770 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db82ec37-860a-4baa-9dcd-e967dd687658" containerName="sg-core" containerID="cri-o://429c0b9d412531309efc726969a27406fe1de8537dc4abb6524ab729d17c8546" gracePeriod=30 Dec 01 16:24:27 crc kubenswrapper[4810]: I1201 16:24:27.982802 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="db82ec37-860a-4baa-9dcd-e967dd687658" containerName="ceilometer-notification-agent" containerID="cri-o://b5af3a32078ae23ef4575623ef13f026da5ca28468f6e916be9439972df2e4f3" gracePeriod=30 Dec 01 16:24:28 crc kubenswrapper[4810]: I1201 16:24:28.061821 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1","Type":"ContainerStarted","Data":"7f0109a8ed436e14a762a62a1a5f553d064ed6038171fde7792832f931dd3fea"} Dec 01 16:24:29 crc kubenswrapper[4810]: I1201 16:24:29.079973 4810 generic.go:334] "Generic (PLEG): container finished" podID="db82ec37-860a-4baa-9dcd-e967dd687658" containerID="21231bb387042290f49bf476bb2ca9e18537490fe26b9a023815f2b32c96378d" exitCode=0 Dec 01 16:24:29 crc kubenswrapper[4810]: I1201 16:24:29.080315 4810 generic.go:334] "Generic (PLEG): container finished" podID="db82ec37-860a-4baa-9dcd-e967dd687658" containerID="429c0b9d412531309efc726969a27406fe1de8537dc4abb6524ab729d17c8546" exitCode=2 Dec 01 16:24:29 crc kubenswrapper[4810]: I1201 16:24:29.080327 4810 generic.go:334] "Generic (PLEG): container finished" podID="db82ec37-860a-4baa-9dcd-e967dd687658" containerID="806cd5b011ecf8247f10aa1d1790e394521c847694025998395ed54776b43d6b" exitCode=0 Dec 01 16:24:29 crc kubenswrapper[4810]: I1201 16:24:29.080162 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db82ec37-860a-4baa-9dcd-e967dd687658","Type":"ContainerDied","Data":"21231bb387042290f49bf476bb2ca9e18537490fe26b9a023815f2b32c96378d"} Dec 01 16:24:29 crc kubenswrapper[4810]: I1201 16:24:29.080367 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db82ec37-860a-4baa-9dcd-e967dd687658","Type":"ContainerDied","Data":"429c0b9d412531309efc726969a27406fe1de8537dc4abb6524ab729d17c8546"} Dec 01 16:24:29 crc kubenswrapper[4810]: I1201 16:24:29.080385 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db82ec37-860a-4baa-9dcd-e967dd687658","Type":"ContainerDied","Data":"806cd5b011ecf8247f10aa1d1790e394521c847694025998395ed54776b43d6b"} Dec 01 16:24:30 crc kubenswrapper[4810]: I1201 16:24:30.013142 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 01 16:24:30 crc kubenswrapper[4810]: I1201 16:24:30.091292 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1","Type":"ContainerStarted","Data":"3f6a81dd21ec6774340043e8b2875b5ad3b64d6dc01c1f25c335c8415b24d0f5"} Dec 01 16:24:30 crc kubenswrapper[4810]: I1201 16:24:30.593196 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 16:24:30 crc kubenswrapper[4810]: I1201 16:24:30.593823 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="67e37611-812c-4270-9225-2ed16b14c0a2" containerName="kube-state-metrics" containerID="cri-o://2474303523ecd36c3c4c15fd281847fc3259faaaf1bcd7a17867ffa92b23bb70" gracePeriod=30 Dec 01 16:24:31 crc kubenswrapper[4810]: I1201 16:24:31.108594 4810 generic.go:334] "Generic (PLEG): container finished" podID="67e37611-812c-4270-9225-2ed16b14c0a2" containerID="2474303523ecd36c3c4c15fd281847fc3259faaaf1bcd7a17867ffa92b23bb70" exitCode=2 Dec 01 16:24:31 crc kubenswrapper[4810]: I1201 16:24:31.109147 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"67e37611-812c-4270-9225-2ed16b14c0a2","Type":"ContainerDied","Data":"2474303523ecd36c3c4c15fd281847fc3259faaaf1bcd7a17867ffa92b23bb70"} Dec 01 16:24:31 crc kubenswrapper[4810]: I1201 16:24:31.109184 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"67e37611-812c-4270-9225-2ed16b14c0a2","Type":"ContainerDied","Data":"c6402d52428d471ca1fbf8a60ae556386a2e8fc2e4895ed63ee4321b7875b86d"} Dec 01 16:24:31 crc kubenswrapper[4810]: I1201 16:24:31.109199 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6402d52428d471ca1fbf8a60ae556386a2e8fc2e4895ed63ee4321b7875b86d" Dec 01 16:24:31 crc kubenswrapper[4810]: I1201 16:24:31.211668 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 16:24:31 crc kubenswrapper[4810]: I1201 16:24:31.280484 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r77sj\" (UniqueName: \"kubernetes.io/projected/67e37611-812c-4270-9225-2ed16b14c0a2-kube-api-access-r77sj\") pod \"67e37611-812c-4270-9225-2ed16b14c0a2\" (UID: \"67e37611-812c-4270-9225-2ed16b14c0a2\") " Dec 01 16:24:31 crc kubenswrapper[4810]: I1201 16:24:31.305921 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67e37611-812c-4270-9225-2ed16b14c0a2-kube-api-access-r77sj" (OuterVolumeSpecName: "kube-api-access-r77sj") pod "67e37611-812c-4270-9225-2ed16b14c0a2" (UID: "67e37611-812c-4270-9225-2ed16b14c0a2"). InnerVolumeSpecName "kube-api-access-r77sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:24:31 crc kubenswrapper[4810]: I1201 16:24:31.382998 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r77sj\" (UniqueName: \"kubernetes.io/projected/67e37611-812c-4270-9225-2ed16b14c0a2-kube-api-access-r77sj\") on node \"crc\" DevicePath \"\"" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.145969 4810 generic.go:334] "Generic (PLEG): container finished" podID="db82ec37-860a-4baa-9dcd-e967dd687658" containerID="b5af3a32078ae23ef4575623ef13f026da5ca28468f6e916be9439972df2e4f3" exitCode=0 Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.146402 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db82ec37-860a-4baa-9dcd-e967dd687658","Type":"ContainerDied","Data":"b5af3a32078ae23ef4575623ef13f026da5ca28468f6e916be9439972df2e4f3"} Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.146534 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.306915 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.332214 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.350917 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 16:24:32 crc kubenswrapper[4810]: E1201 16:24:32.351445 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67e37611-812c-4270-9225-2ed16b14c0a2" containerName="kube-state-metrics" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.351459 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="67e37611-812c-4270-9225-2ed16b14c0a2" containerName="kube-state-metrics" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.351753 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="67e37611-812c-4270-9225-2ed16b14c0a2" containerName="kube-state-metrics" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.352932 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.356314 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.356457 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.374995 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.407875 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b853756-ebf7-4dac-9e50-67524f3b8abe-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"4b853756-ebf7-4dac-9e50-67524f3b8abe\") " pod="openstack/kube-state-metrics-0" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.407979 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhlqq\" (UniqueName: \"kubernetes.io/projected/4b853756-ebf7-4dac-9e50-67524f3b8abe-kube-api-access-nhlqq\") pod \"kube-state-metrics-0\" (UID: \"4b853756-ebf7-4dac-9e50-67524f3b8abe\") " pod="openstack/kube-state-metrics-0" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.408101 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/4b853756-ebf7-4dac-9e50-67524f3b8abe-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"4b853756-ebf7-4dac-9e50-67524f3b8abe\") " pod="openstack/kube-state-metrics-0" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.408198 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b853756-ebf7-4dac-9e50-67524f3b8abe-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"4b853756-ebf7-4dac-9e50-67524f3b8abe\") " pod="openstack/kube-state-metrics-0" Dec 01 16:24:32 crc kubenswrapper[4810]: E1201 16:24:32.498822 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb82ec37_860a_4baa_9dcd_e967dd687658.slice/crio-b5af3a32078ae23ef4575623ef13f026da5ca28468f6e916be9439972df2e4f3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67e37611_812c_4270_9225_2ed16b14c0a2.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb82ec37_860a_4baa_9dcd_e967dd687658.slice/crio-conmon-b5af3a32078ae23ef4575623ef13f026da5ca28468f6e916be9439972df2e4f3.scope\": RecentStats: unable to find data in memory cache]" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.509794 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b853756-ebf7-4dac-9e50-67524f3b8abe-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"4b853756-ebf7-4dac-9e50-67524f3b8abe\") " pod="openstack/kube-state-metrics-0" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.509911 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b853756-ebf7-4dac-9e50-67524f3b8abe-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"4b853756-ebf7-4dac-9e50-67524f3b8abe\") " pod="openstack/kube-state-metrics-0" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.509965 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhlqq\" (UniqueName: \"kubernetes.io/projected/4b853756-ebf7-4dac-9e50-67524f3b8abe-kube-api-access-nhlqq\") pod \"kube-state-metrics-0\" (UID: \"4b853756-ebf7-4dac-9e50-67524f3b8abe\") " pod="openstack/kube-state-metrics-0" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.510035 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/4b853756-ebf7-4dac-9e50-67524f3b8abe-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"4b853756-ebf7-4dac-9e50-67524f3b8abe\") " pod="openstack/kube-state-metrics-0" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.514942 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/4b853756-ebf7-4dac-9e50-67524f3b8abe-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"4b853756-ebf7-4dac-9e50-67524f3b8abe\") " pod="openstack/kube-state-metrics-0" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.517600 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67e37611-812c-4270-9225-2ed16b14c0a2" path="/var/lib/kubelet/pods/67e37611-812c-4270-9225-2ed16b14c0a2/volumes" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.520940 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b853756-ebf7-4dac-9e50-67524f3b8abe-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"4b853756-ebf7-4dac-9e50-67524f3b8abe\") " pod="openstack/kube-state-metrics-0" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.521948 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b853756-ebf7-4dac-9e50-67524f3b8abe-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"4b853756-ebf7-4dac-9e50-67524f3b8abe\") " pod="openstack/kube-state-metrics-0" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.530666 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhlqq\" (UniqueName: \"kubernetes.io/projected/4b853756-ebf7-4dac-9e50-67524f3b8abe-kube-api-access-nhlqq\") pod \"kube-state-metrics-0\" (UID: \"4b853756-ebf7-4dac-9e50-67524f3b8abe\") " pod="openstack/kube-state-metrics-0" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.582786 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.685396 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.715333 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qvf9\" (UniqueName: \"kubernetes.io/projected/db82ec37-860a-4baa-9dcd-e967dd687658-kube-api-access-9qvf9\") pod \"db82ec37-860a-4baa-9dcd-e967dd687658\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.715396 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-scripts\") pod \"db82ec37-860a-4baa-9dcd-e967dd687658\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.715526 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-sg-core-conf-yaml\") pod \"db82ec37-860a-4baa-9dcd-e967dd687658\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.715636 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db82ec37-860a-4baa-9dcd-e967dd687658-log-httpd\") pod \"db82ec37-860a-4baa-9dcd-e967dd687658\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.715740 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-combined-ca-bundle\") pod \"db82ec37-860a-4baa-9dcd-e967dd687658\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.715885 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-config-data\") pod \"db82ec37-860a-4baa-9dcd-e967dd687658\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.715912 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db82ec37-860a-4baa-9dcd-e967dd687658-run-httpd\") pod \"db82ec37-860a-4baa-9dcd-e967dd687658\" (UID: \"db82ec37-860a-4baa-9dcd-e967dd687658\") " Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.717176 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db82ec37-860a-4baa-9dcd-e967dd687658-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "db82ec37-860a-4baa-9dcd-e967dd687658" (UID: "db82ec37-860a-4baa-9dcd-e967dd687658"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.717592 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db82ec37-860a-4baa-9dcd-e967dd687658-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "db82ec37-860a-4baa-9dcd-e967dd687658" (UID: "db82ec37-860a-4baa-9dcd-e967dd687658"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.720573 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db82ec37-860a-4baa-9dcd-e967dd687658-kube-api-access-9qvf9" (OuterVolumeSpecName: "kube-api-access-9qvf9") pod "db82ec37-860a-4baa-9dcd-e967dd687658" (UID: "db82ec37-860a-4baa-9dcd-e967dd687658"). InnerVolumeSpecName "kube-api-access-9qvf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.721494 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-scripts" (OuterVolumeSpecName: "scripts") pod "db82ec37-860a-4baa-9dcd-e967dd687658" (UID: "db82ec37-860a-4baa-9dcd-e967dd687658"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.753443 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "db82ec37-860a-4baa-9dcd-e967dd687658" (UID: "db82ec37-860a-4baa-9dcd-e967dd687658"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.809825 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db82ec37-860a-4baa-9dcd-e967dd687658" (UID: "db82ec37-860a-4baa-9dcd-e967dd687658"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.821279 4810 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db82ec37-860a-4baa-9dcd-e967dd687658-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.821327 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qvf9\" (UniqueName: \"kubernetes.io/projected/db82ec37-860a-4baa-9dcd-e967dd687658-kube-api-access-9qvf9\") on node \"crc\" DevicePath \"\"" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.821340 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.821348 4810 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.821359 4810 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/db82ec37-860a-4baa-9dcd-e967dd687658-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.821385 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.852380 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-config-data" (OuterVolumeSpecName: "config-data") pod "db82ec37-860a-4baa-9dcd-e967dd687658" (UID: "db82ec37-860a-4baa-9dcd-e967dd687658"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:24:32 crc kubenswrapper[4810]: I1201 16:24:32.923220 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db82ec37-860a-4baa-9dcd-e967dd687658-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:24:33 crc kubenswrapper[4810]: W1201 16:24:33.153030 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b853756_ebf7_4dac_9e50_67524f3b8abe.slice/crio-95123467f92e667e38679f04e7c0607d3757411f99c9bb82355ea42508dd9e56 WatchSource:0}: Error finding container 95123467f92e667e38679f04e7c0607d3757411f99c9bb82355ea42508dd9e56: Status 404 returned error can't find the container with id 95123467f92e667e38679f04e7c0607d3757411f99c9bb82355ea42508dd9e56 Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.158371 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.163688 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1","Type":"ContainerStarted","Data":"e32b40afabc8b48a4fbd851c616d46f5b4edc522f56e548ca30c74dd1bb7ec49"} Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.163861 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerName="aodh-api" containerID="cri-o://5449f9949a4c937ef3e6c920d57228d915efeb6e7f2b4523f10ca5b42af150cd" gracePeriod=30 Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.163984 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerName="aodh-notifier" containerID="cri-o://3f6a81dd21ec6774340043e8b2875b5ad3b64d6dc01c1f25c335c8415b24d0f5" gracePeriod=30 Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.164223 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerName="aodh-evaluator" containerID="cri-o://7f0109a8ed436e14a762a62a1a5f553d064ed6038171fde7792832f931dd3fea" gracePeriod=30 Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.164223 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerName="aodh-listener" containerID="cri-o://e32b40afabc8b48a4fbd851c616d46f5b4edc522f56e548ca30c74dd1bb7ec49" gracePeriod=30 Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.174576 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"db82ec37-860a-4baa-9dcd-e967dd687658","Type":"ContainerDied","Data":"545e315c4b9d9c8780713e2d214bede795b918aa9ccdcd82467cd835790947b9"} Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.174644 4810 scope.go:117] "RemoveContainer" containerID="21231bb387042290f49bf476bb2ca9e18537490fe26b9a023815f2b32c96378d" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.174863 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.212411 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=3.130829169 podStartE2EDuration="10.212382641s" podCreationTimestamp="2025-12-01 16:24:23 +0000 UTC" firstStartedPulling="2025-12-01 16:24:24.930244577 +0000 UTC m=+6630.693754180" lastFinishedPulling="2025-12-01 16:24:32.011798049 +0000 UTC m=+6637.775307652" observedRunningTime="2025-12-01 16:24:33.199997836 +0000 UTC m=+6638.963507459" watchObservedRunningTime="2025-12-01 16:24:33.212382641 +0000 UTC m=+6638.975892234" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.269551 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.273169 4810 scope.go:117] "RemoveContainer" containerID="429c0b9d412531309efc726969a27406fe1de8537dc4abb6524ab729d17c8546" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.301586 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.311064 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.318682 4810 scope.go:117] "RemoveContainer" containerID="b5af3a32078ae23ef4575623ef13f026da5ca28468f6e916be9439972df2e4f3" Dec 01 16:24:33 crc kubenswrapper[4810]: E1201 16:24:33.320744 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db82ec37-860a-4baa-9dcd-e967dd687658" containerName="proxy-httpd" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.320781 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="db82ec37-860a-4baa-9dcd-e967dd687658" containerName="proxy-httpd" Dec 01 16:24:33 crc kubenswrapper[4810]: E1201 16:24:33.320819 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db82ec37-860a-4baa-9dcd-e967dd687658" containerName="sg-core" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.320826 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="db82ec37-860a-4baa-9dcd-e967dd687658" containerName="sg-core" Dec 01 16:24:33 crc kubenswrapper[4810]: E1201 16:24:33.320849 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db82ec37-860a-4baa-9dcd-e967dd687658" containerName="ceilometer-notification-agent" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.320855 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="db82ec37-860a-4baa-9dcd-e967dd687658" containerName="ceilometer-notification-agent" Dec 01 16:24:33 crc kubenswrapper[4810]: E1201 16:24:33.320864 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db82ec37-860a-4baa-9dcd-e967dd687658" containerName="ceilometer-central-agent" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.320870 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="db82ec37-860a-4baa-9dcd-e967dd687658" containerName="ceilometer-central-agent" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.321072 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="db82ec37-860a-4baa-9dcd-e967dd687658" containerName="sg-core" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.321095 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="db82ec37-860a-4baa-9dcd-e967dd687658" containerName="ceilometer-notification-agent" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.321105 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="db82ec37-860a-4baa-9dcd-e967dd687658" containerName="proxy-httpd" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.321123 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="db82ec37-860a-4baa-9dcd-e967dd687658" containerName="ceilometer-central-agent" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.323963 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.327650 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.329407 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.329647 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.356319 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.360828 4810 scope.go:117] "RemoveContainer" containerID="806cd5b011ecf8247f10aa1d1790e394521c847694025998395ed54776b43d6b" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.458861 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qql8b\" (UniqueName: \"kubernetes.io/projected/40aec98f-dc77-482d-b562-c5e5c9eed98d-kube-api-access-qql8b\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.459712 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40aec98f-dc77-482d-b562-c5e5c9eed98d-log-httpd\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.459896 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40aec98f-dc77-482d-b562-c5e5c9eed98d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.460140 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40aec98f-dc77-482d-b562-c5e5c9eed98d-config-data\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.460233 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/40aec98f-dc77-482d-b562-c5e5c9eed98d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.460917 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40aec98f-dc77-482d-b562-c5e5c9eed98d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.461012 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40aec98f-dc77-482d-b562-c5e5c9eed98d-run-httpd\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.461097 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40aec98f-dc77-482d-b562-c5e5c9eed98d-scripts\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.563056 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40aec98f-dc77-482d-b562-c5e5c9eed98d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.563332 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40aec98f-dc77-482d-b562-c5e5c9eed98d-run-httpd\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.563412 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40aec98f-dc77-482d-b562-c5e5c9eed98d-scripts\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.563543 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qql8b\" (UniqueName: \"kubernetes.io/projected/40aec98f-dc77-482d-b562-c5e5c9eed98d-kube-api-access-qql8b\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.563639 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40aec98f-dc77-482d-b562-c5e5c9eed98d-log-httpd\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.563725 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40aec98f-dc77-482d-b562-c5e5c9eed98d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.563820 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40aec98f-dc77-482d-b562-c5e5c9eed98d-run-httpd\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.564112 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40aec98f-dc77-482d-b562-c5e5c9eed98d-config-data\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.564232 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/40aec98f-dc77-482d-b562-c5e5c9eed98d-log-httpd\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.564235 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/40aec98f-dc77-482d-b562-c5e5c9eed98d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.572437 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/40aec98f-dc77-482d-b562-c5e5c9eed98d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.572584 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40aec98f-dc77-482d-b562-c5e5c9eed98d-config-data\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.572903 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40aec98f-dc77-482d-b562-c5e5c9eed98d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.573915 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/40aec98f-dc77-482d-b562-c5e5c9eed98d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.582851 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40aec98f-dc77-482d-b562-c5e5c9eed98d-scripts\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.583481 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qql8b\" (UniqueName: \"kubernetes.io/projected/40aec98f-dc77-482d-b562-c5e5c9eed98d-kube-api-access-qql8b\") pod \"ceilometer-0\" (UID: \"40aec98f-dc77-482d-b562-c5e5c9eed98d\") " pod="openstack/ceilometer-0" Dec 01 16:24:33 crc kubenswrapper[4810]: I1201 16:24:33.657105 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 16:24:34 crc kubenswrapper[4810]: I1201 16:24:34.186982 4810 generic.go:334] "Generic (PLEG): container finished" podID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerID="7f0109a8ed436e14a762a62a1a5f553d064ed6038171fde7792832f931dd3fea" exitCode=0 Dec 01 16:24:34 crc kubenswrapper[4810]: I1201 16:24:34.188244 4810 generic.go:334] "Generic (PLEG): container finished" podID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerID="5449f9949a4c937ef3e6c920d57228d915efeb6e7f2b4523f10ca5b42af150cd" exitCode=0 Dec 01 16:24:34 crc kubenswrapper[4810]: I1201 16:24:34.187162 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1","Type":"ContainerDied","Data":"7f0109a8ed436e14a762a62a1a5f553d064ed6038171fde7792832f931dd3fea"} Dec 01 16:24:34 crc kubenswrapper[4810]: I1201 16:24:34.188323 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1","Type":"ContainerDied","Data":"5449f9949a4c937ef3e6c920d57228d915efeb6e7f2b4523f10ca5b42af150cd"} Dec 01 16:24:34 crc kubenswrapper[4810]: I1201 16:24:34.191057 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4b853756-ebf7-4dac-9e50-67524f3b8abe","Type":"ContainerStarted","Data":"95123467f92e667e38679f04e7c0607d3757411f99c9bb82355ea42508dd9e56"} Dec 01 16:24:34 crc kubenswrapper[4810]: I1201 16:24:34.351532 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 16:24:34 crc kubenswrapper[4810]: W1201 16:24:34.379397 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40aec98f_dc77_482d_b562_c5e5c9eed98d.slice/crio-84022d5820ee07eb3458020eeedc8fce335ae537ae6ea74605c3bde15ba09d78 WatchSource:0}: Error finding container 84022d5820ee07eb3458020eeedc8fce335ae537ae6ea74605c3bde15ba09d78: Status 404 returned error can't find the container with id 84022d5820ee07eb3458020eeedc8fce335ae537ae6ea74605c3bde15ba09d78 Dec 01 16:24:34 crc kubenswrapper[4810]: I1201 16:24:34.504773 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db82ec37-860a-4baa-9dcd-e967dd687658" path="/var/lib/kubelet/pods/db82ec37-860a-4baa-9dcd-e967dd687658/volumes" Dec 01 16:24:35 crc kubenswrapper[4810]: I1201 16:24:35.207374 4810 generic.go:334] "Generic (PLEG): container finished" podID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerID="3f6a81dd21ec6774340043e8b2875b5ad3b64d6dc01c1f25c335c8415b24d0f5" exitCode=0 Dec 01 16:24:35 crc kubenswrapper[4810]: I1201 16:24:35.207442 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1","Type":"ContainerDied","Data":"3f6a81dd21ec6774340043e8b2875b5ad3b64d6dc01c1f25c335c8415b24d0f5"} Dec 01 16:24:35 crc kubenswrapper[4810]: I1201 16:24:35.209818 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40aec98f-dc77-482d-b562-c5e5c9eed98d","Type":"ContainerStarted","Data":"84022d5820ee07eb3458020eeedc8fce335ae537ae6ea74605c3bde15ba09d78"} Dec 01 16:24:35 crc kubenswrapper[4810]: I1201 16:24:35.211818 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4b853756-ebf7-4dac-9e50-67524f3b8abe","Type":"ContainerStarted","Data":"fcde3d344f292bbb7a6c90e80dc2f2ef3fe56a899cccfd79f7d98641fae4bfff"} Dec 01 16:24:35 crc kubenswrapper[4810]: I1201 16:24:35.212458 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 01 16:24:35 crc kubenswrapper[4810]: I1201 16:24:35.233912 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.8212506400000001 podStartE2EDuration="3.233889599s" podCreationTimestamp="2025-12-01 16:24:32 +0000 UTC" firstStartedPulling="2025-12-01 16:24:33.159083164 +0000 UTC m=+6638.922592767" lastFinishedPulling="2025-12-01 16:24:34.571722123 +0000 UTC m=+6640.335231726" observedRunningTime="2025-12-01 16:24:35.230323893 +0000 UTC m=+6640.993833506" watchObservedRunningTime="2025-12-01 16:24:35.233889599 +0000 UTC m=+6640.997399202" Dec 01 16:24:36 crc kubenswrapper[4810]: I1201 16:24:36.222795 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40aec98f-dc77-482d-b562-c5e5c9eed98d","Type":"ContainerStarted","Data":"1e74550416afdd900784c51a103f352012df16c5b15dafa9dc097d4a19f4aa0d"} Dec 01 16:24:36 crc kubenswrapper[4810]: I1201 16:24:36.222853 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40aec98f-dc77-482d-b562-c5e5c9eed98d","Type":"ContainerStarted","Data":"27d9c34e780627448ddfb7d9cb05b9a21b6c5eddb2b0de02fd2c42e129d20787"} Dec 01 16:24:37 crc kubenswrapper[4810]: I1201 16:24:37.236124 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40aec98f-dc77-482d-b562-c5e5c9eed98d","Type":"ContainerStarted","Data":"5b13e64edda2b2b4b4f3d54fcdf22febe04f80a343d9cfcebe1d9f6d1fb0276b"} Dec 01 16:24:39 crc kubenswrapper[4810]: I1201 16:24:39.270419 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"40aec98f-dc77-482d-b562-c5e5c9eed98d","Type":"ContainerStarted","Data":"0cbbb6bca6e97a8a74425877819341a00b3f9c47cd51d32e1e79b3b79421b596"} Dec 01 16:24:39 crc kubenswrapper[4810]: I1201 16:24:39.271004 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 16:24:39 crc kubenswrapper[4810]: I1201 16:24:39.301953 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.54998924 podStartE2EDuration="6.301932914s" podCreationTimestamp="2025-12-01 16:24:33 +0000 UTC" firstStartedPulling="2025-12-01 16:24:34.3877861 +0000 UTC m=+6640.151295703" lastFinishedPulling="2025-12-01 16:24:38.139729774 +0000 UTC m=+6643.903239377" observedRunningTime="2025-12-01 16:24:39.2936514 +0000 UTC m=+6645.057161093" watchObservedRunningTime="2025-12-01 16:24:39.301932914 +0000 UTC m=+6645.065442517" Dec 01 16:24:42 crc kubenswrapper[4810]: I1201 16:24:42.696910 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 01 16:24:56 crc kubenswrapper[4810]: I1201 16:24:56.046871 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-942ws"] Dec 01 16:24:56 crc kubenswrapper[4810]: I1201 16:24:56.059862 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-nfcz6"] Dec 01 16:24:56 crc kubenswrapper[4810]: I1201 16:24:56.068385 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-a524-account-create-update-4cr2g"] Dec 01 16:24:56 crc kubenswrapper[4810]: I1201 16:24:56.078082 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-378c-account-create-update-rbsww"] Dec 01 16:24:56 crc kubenswrapper[4810]: I1201 16:24:56.086600 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-942ws"] Dec 01 16:24:56 crc kubenswrapper[4810]: I1201 16:24:56.095309 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-a524-account-create-update-4cr2g"] Dec 01 16:24:56 crc kubenswrapper[4810]: I1201 16:24:56.103919 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-nfcz6"] Dec 01 16:24:56 crc kubenswrapper[4810]: I1201 16:24:56.112675 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-378c-account-create-update-rbsww"] Dec 01 16:24:56 crc kubenswrapper[4810]: I1201 16:24:56.513877 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f61fca3-87ad-4722-9d8d-f212af501835" path="/var/lib/kubelet/pods/0f61fca3-87ad-4722-9d8d-f212af501835/volumes" Dec 01 16:24:56 crc kubenswrapper[4810]: I1201 16:24:56.515053 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25ad5508-0e9b-4c6e-86ad-472613af32be" path="/var/lib/kubelet/pods/25ad5508-0e9b-4c6e-86ad-472613af32be/volumes" Dec 01 16:24:56 crc kubenswrapper[4810]: I1201 16:24:56.516090 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4" path="/var/lib/kubelet/pods/ae98057f-ede8-47b0-b0d8-fea7d5d2e6f4/volumes" Dec 01 16:24:56 crc kubenswrapper[4810]: I1201 16:24:56.517040 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2d19437-5742-4516-8c26-52267458dcec" path="/var/lib/kubelet/pods/b2d19437-5742-4516-8c26-52267458dcec/volumes" Dec 01 16:24:57 crc kubenswrapper[4810]: I1201 16:24:57.034984 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-0fac-account-create-update-57jps"] Dec 01 16:24:57 crc kubenswrapper[4810]: I1201 16:24:57.047082 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-kdnv8"] Dec 01 16:24:57 crc kubenswrapper[4810]: I1201 16:24:57.055814 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-0fac-account-create-update-57jps"] Dec 01 16:24:57 crc kubenswrapper[4810]: I1201 16:24:57.064688 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-kdnv8"] Dec 01 16:24:58 crc kubenswrapper[4810]: I1201 16:24:58.508595 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d789e4e-7e8f-48c6-94e2-cf1d081d0018" path="/var/lib/kubelet/pods/1d789e4e-7e8f-48c6-94e2-cf1d081d0018/volumes" Dec 01 16:24:58 crc kubenswrapper[4810]: I1201 16:24:58.510191 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6e67c08-6bc5-4c10-ac10-80bb918b95d5" path="/var/lib/kubelet/pods/e6e67c08-6bc5-4c10-ac10-80bb918b95d5/volumes" Dec 01 16:25:02 crc kubenswrapper[4810]: I1201 16:25:02.972806 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:25:02 crc kubenswrapper[4810]: I1201 16:25:02.973424 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:25:03 crc kubenswrapper[4810]: E1201 16:25:03.459370 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2d472ec_cb41_4b4e_b57f_acbd8d91b3b1.slice/crio-e32b40afabc8b48a4fbd851c616d46f5b4edc522f56e548ca30c74dd1bb7ec49.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2d472ec_cb41_4b4e_b57f_acbd8d91b3b1.slice/crio-conmon-e32b40afabc8b48a4fbd851c616d46f5b4edc522f56e548ca30c74dd1bb7ec49.scope\": RecentStats: unable to find data in memory cache]" Dec 01 16:25:03 crc kubenswrapper[4810]: I1201 16:25:03.667447 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 16:25:03 crc kubenswrapper[4810]: I1201 16:25:03.801170 4810 generic.go:334] "Generic (PLEG): container finished" podID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerID="e32b40afabc8b48a4fbd851c616d46f5b4edc522f56e548ca30c74dd1bb7ec49" exitCode=137 Dec 01 16:25:03 crc kubenswrapper[4810]: I1201 16:25:03.801227 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1","Type":"ContainerDied","Data":"e32b40afabc8b48a4fbd851c616d46f5b4edc522f56e548ca30c74dd1bb7ec49"} Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.051005 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.144826 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v694p\" (UniqueName: \"kubernetes.io/projected/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-kube-api-access-v694p\") pod \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\" (UID: \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\") " Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.144900 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-config-data\") pod \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\" (UID: \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\") " Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.145012 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-combined-ca-bundle\") pod \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\" (UID: \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\") " Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.145047 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-scripts\") pod \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\" (UID: \"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1\") " Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.164000 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-kube-api-access-v694p" (OuterVolumeSpecName: "kube-api-access-v694p") pod "e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" (UID: "e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1"). InnerVolumeSpecName "kube-api-access-v694p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.164899 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-scripts" (OuterVolumeSpecName: "scripts") pod "e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" (UID: "e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.247784 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v694p\" (UniqueName: \"kubernetes.io/projected/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-kube-api-access-v694p\") on node \"crc\" DevicePath \"\"" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.247834 4810 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.278301 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" (UID: "e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.308661 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-config-data" (OuterVolumeSpecName: "config-data") pod "e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" (UID: "e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.349602 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.349652 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.815954 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1","Type":"ContainerDied","Data":"89efac2d97004dc6229b3922cedb754555ccb061bd49cc316c6838cd861d5a83"} Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.816094 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.816258 4810 scope.go:117] "RemoveContainer" containerID="e32b40afabc8b48a4fbd851c616d46f5b4edc522f56e548ca30c74dd1bb7ec49" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.848257 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.856182 4810 scope.go:117] "RemoveContainer" containerID="3f6a81dd21ec6774340043e8b2875b5ad3b64d6dc01c1f25c335c8415b24d0f5" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.860086 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.920214 4810 scope.go:117] "RemoveContainer" containerID="7f0109a8ed436e14a762a62a1a5f553d064ed6038171fde7792832f931dd3fea" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.922599 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 01 16:25:04 crc kubenswrapper[4810]: E1201 16:25:04.923019 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerName="aodh-api" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.923032 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerName="aodh-api" Dec 01 16:25:04 crc kubenswrapper[4810]: E1201 16:25:04.923054 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerName="aodh-evaluator" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.923060 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerName="aodh-evaluator" Dec 01 16:25:04 crc kubenswrapper[4810]: E1201 16:25:04.923070 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerName="aodh-listener" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.923076 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerName="aodh-listener" Dec 01 16:25:04 crc kubenswrapper[4810]: E1201 16:25:04.923100 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerName="aodh-notifier" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.923105 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerName="aodh-notifier" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.923290 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerName="aodh-evaluator" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.923343 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerName="aodh-api" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.923350 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerName="aodh-listener" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.923363 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" containerName="aodh-notifier" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.926952 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.931841 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.932314 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-js2tk" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.932486 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.932577 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.932619 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.963821 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52d88aa8-f7c3-4775-81f4-7d2ce2c4832a-scripts\") pod \"aodh-0\" (UID: \"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a\") " pod="openstack/aodh-0" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.963943 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/52d88aa8-f7c3-4775-81f4-7d2ce2c4832a-internal-tls-certs\") pod \"aodh-0\" (UID: \"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a\") " pod="openstack/aodh-0" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.964002 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52d88aa8-f7c3-4775-81f4-7d2ce2c4832a-config-data\") pod \"aodh-0\" (UID: \"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a\") " pod="openstack/aodh-0" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.964048 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpkq8\" (UniqueName: \"kubernetes.io/projected/52d88aa8-f7c3-4775-81f4-7d2ce2c4832a-kube-api-access-fpkq8\") pod \"aodh-0\" (UID: \"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a\") " pod="openstack/aodh-0" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.964094 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52d88aa8-f7c3-4775-81f4-7d2ce2c4832a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a\") " pod="openstack/aodh-0" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.964179 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52d88aa8-f7c3-4775-81f4-7d2ce2c4832a-public-tls-certs\") pod \"aodh-0\" (UID: \"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a\") " pod="openstack/aodh-0" Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.998520 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 16:25:04 crc kubenswrapper[4810]: I1201 16:25:04.999608 4810 scope.go:117] "RemoveContainer" containerID="5449f9949a4c937ef3e6c920d57228d915efeb6e7f2b4523f10ca5b42af150cd" Dec 01 16:25:05 crc kubenswrapper[4810]: I1201 16:25:05.067901 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52d88aa8-f7c3-4775-81f4-7d2ce2c4832a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a\") " pod="openstack/aodh-0" Dec 01 16:25:05 crc kubenswrapper[4810]: I1201 16:25:05.068757 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52d88aa8-f7c3-4775-81f4-7d2ce2c4832a-public-tls-certs\") pod \"aodh-0\" (UID: \"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a\") " pod="openstack/aodh-0" Dec 01 16:25:05 crc kubenswrapper[4810]: I1201 16:25:05.068948 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52d88aa8-f7c3-4775-81f4-7d2ce2c4832a-scripts\") pod \"aodh-0\" (UID: \"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a\") " pod="openstack/aodh-0" Dec 01 16:25:05 crc kubenswrapper[4810]: I1201 16:25:05.069089 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/52d88aa8-f7c3-4775-81f4-7d2ce2c4832a-internal-tls-certs\") pod \"aodh-0\" (UID: \"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a\") " pod="openstack/aodh-0" Dec 01 16:25:05 crc kubenswrapper[4810]: I1201 16:25:05.069318 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52d88aa8-f7c3-4775-81f4-7d2ce2c4832a-config-data\") pod \"aodh-0\" (UID: \"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a\") " pod="openstack/aodh-0" Dec 01 16:25:05 crc kubenswrapper[4810]: I1201 16:25:05.069498 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpkq8\" (UniqueName: \"kubernetes.io/projected/52d88aa8-f7c3-4775-81f4-7d2ce2c4832a-kube-api-access-fpkq8\") pod \"aodh-0\" (UID: \"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a\") " pod="openstack/aodh-0" Dec 01 16:25:05 crc kubenswrapper[4810]: I1201 16:25:05.079180 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52d88aa8-f7c3-4775-81f4-7d2ce2c4832a-public-tls-certs\") pod \"aodh-0\" (UID: \"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a\") " pod="openstack/aodh-0" Dec 01 16:25:05 crc kubenswrapper[4810]: I1201 16:25:05.079930 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52d88aa8-f7c3-4775-81f4-7d2ce2c4832a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a\") " pod="openstack/aodh-0" Dec 01 16:25:05 crc kubenswrapper[4810]: I1201 16:25:05.083612 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52d88aa8-f7c3-4775-81f4-7d2ce2c4832a-scripts\") pod \"aodh-0\" (UID: \"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a\") " pod="openstack/aodh-0" Dec 01 16:25:05 crc kubenswrapper[4810]: I1201 16:25:05.087725 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52d88aa8-f7c3-4775-81f4-7d2ce2c4832a-config-data\") pod \"aodh-0\" (UID: \"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a\") " pod="openstack/aodh-0" Dec 01 16:25:05 crc kubenswrapper[4810]: I1201 16:25:05.094265 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/52d88aa8-f7c3-4775-81f4-7d2ce2c4832a-internal-tls-certs\") pod \"aodh-0\" (UID: \"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a\") " pod="openstack/aodh-0" Dec 01 16:25:05 crc kubenswrapper[4810]: I1201 16:25:05.099554 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpkq8\" (UniqueName: \"kubernetes.io/projected/52d88aa8-f7c3-4775-81f4-7d2ce2c4832a-kube-api-access-fpkq8\") pod \"aodh-0\" (UID: \"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a\") " pod="openstack/aodh-0" Dec 01 16:25:05 crc kubenswrapper[4810]: I1201 16:25:05.123138 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-82k7s"] Dec 01 16:25:05 crc kubenswrapper[4810]: I1201 16:25:05.139460 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-82k7s"] Dec 01 16:25:05 crc kubenswrapper[4810]: I1201 16:25:05.287964 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 01 16:25:05 crc kubenswrapper[4810]: I1201 16:25:05.781460 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 01 16:25:05 crc kubenswrapper[4810]: W1201 16:25:05.782263 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52d88aa8_f7c3_4775_81f4_7d2ce2c4832a.slice/crio-bd6267148c91fe26ac87a5f7b4159fc58f4791b51116525c68e459a7e28d9c8c WatchSource:0}: Error finding container bd6267148c91fe26ac87a5f7b4159fc58f4791b51116525c68e459a7e28d9c8c: Status 404 returned error can't find the container with id bd6267148c91fe26ac87a5f7b4159fc58f4791b51116525c68e459a7e28d9c8c Dec 01 16:25:05 crc kubenswrapper[4810]: I1201 16:25:05.833356 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a","Type":"ContainerStarted","Data":"bd6267148c91fe26ac87a5f7b4159fc58f4791b51116525c68e459a7e28d9c8c"} Dec 01 16:25:06 crc kubenswrapper[4810]: I1201 16:25:06.505092 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1" path="/var/lib/kubelet/pods/e2d472ec-cb41-4b4e-b57f-acbd8d91b3b1/volumes" Dec 01 16:25:06 crc kubenswrapper[4810]: I1201 16:25:06.508098 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed991281-68e4-4e60-ab15-889f970b898d" path="/var/lib/kubelet/pods/ed991281-68e4-4e60-ab15-889f970b898d/volumes" Dec 01 16:25:07 crc kubenswrapper[4810]: I1201 16:25:07.868836 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a","Type":"ContainerStarted","Data":"465d8c8e0077104a094b5f448ed130681c34d1550bf4a43073454cc5a97a2bd5"} Dec 01 16:25:07 crc kubenswrapper[4810]: I1201 16:25:07.869655 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a","Type":"ContainerStarted","Data":"43fdf21ead1adf45e4f4e5ca88e73db4cea0566d8dff7d2deb340e167462416f"} Dec 01 16:25:08 crc kubenswrapper[4810]: I1201 16:25:08.880466 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a","Type":"ContainerStarted","Data":"3df8e513293dbbd20b837da7bc7f7c5129431100a00bf78e588264f7eaecbbad"} Dec 01 16:25:09 crc kubenswrapper[4810]: I1201 16:25:09.890858 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"52d88aa8-f7c3-4775-81f4-7d2ce2c4832a","Type":"ContainerStarted","Data":"6c61ad53c7c312606c8cb53e276be568d20fc1d54fa9b372b56ff3a3be9adcaa"} Dec 01 16:25:09 crc kubenswrapper[4810]: I1201 16:25:09.914127 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.264644107 podStartE2EDuration="5.914103878s" podCreationTimestamp="2025-12-01 16:25:04 +0000 UTC" firstStartedPulling="2025-12-01 16:25:05.786121867 +0000 UTC m=+6671.549631470" lastFinishedPulling="2025-12-01 16:25:09.435581638 +0000 UTC m=+6675.199091241" observedRunningTime="2025-12-01 16:25:09.90824851 +0000 UTC m=+6675.671758113" watchObservedRunningTime="2025-12-01 16:25:09.914103878 +0000 UTC m=+6675.677613471" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.389403 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bf4588c47-pv8np"] Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.395068 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.397663 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.401983 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bf4588c47-pv8np"] Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.554138 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-ovsdbserver-nb\") pod \"dnsmasq-dns-7bf4588c47-pv8np\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.554200 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-config\") pod \"dnsmasq-dns-7bf4588c47-pv8np\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.554251 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-ovsdbserver-sb\") pod \"dnsmasq-dns-7bf4588c47-pv8np\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.554316 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-openstack-cell1\") pod \"dnsmasq-dns-7bf4588c47-pv8np\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.554342 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-dns-svc\") pod \"dnsmasq-dns-7bf4588c47-pv8np\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.554446 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xnnb\" (UniqueName: \"kubernetes.io/projected/e85267ea-ec74-4851-a062-05d2aa643a30-kube-api-access-9xnnb\") pod \"dnsmasq-dns-7bf4588c47-pv8np\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.656886 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xnnb\" (UniqueName: \"kubernetes.io/projected/e85267ea-ec74-4851-a062-05d2aa643a30-kube-api-access-9xnnb\") pod \"dnsmasq-dns-7bf4588c47-pv8np\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.656991 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-ovsdbserver-nb\") pod \"dnsmasq-dns-7bf4588c47-pv8np\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.657041 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-config\") pod \"dnsmasq-dns-7bf4588c47-pv8np\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.657109 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-ovsdbserver-sb\") pod \"dnsmasq-dns-7bf4588c47-pv8np\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.657202 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-openstack-cell1\") pod \"dnsmasq-dns-7bf4588c47-pv8np\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.657236 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-dns-svc\") pod \"dnsmasq-dns-7bf4588c47-pv8np\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.658000 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-ovsdbserver-nb\") pod \"dnsmasq-dns-7bf4588c47-pv8np\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.658072 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-openstack-cell1\") pod \"dnsmasq-dns-7bf4588c47-pv8np\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.658137 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-config\") pod \"dnsmasq-dns-7bf4588c47-pv8np\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.658312 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-dns-svc\") pod \"dnsmasq-dns-7bf4588c47-pv8np\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.658355 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-ovsdbserver-sb\") pod \"dnsmasq-dns-7bf4588c47-pv8np\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.684390 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xnnb\" (UniqueName: \"kubernetes.io/projected/e85267ea-ec74-4851-a062-05d2aa643a30-kube-api-access-9xnnb\") pod \"dnsmasq-dns-7bf4588c47-pv8np\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:11 crc kubenswrapper[4810]: I1201 16:25:11.732011 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:12 crc kubenswrapper[4810]: I1201 16:25:12.219912 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bf4588c47-pv8np"] Dec 01 16:25:12 crc kubenswrapper[4810]: W1201 16:25:12.222737 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode85267ea_ec74_4851_a062_05d2aa643a30.slice/crio-e456bf1f51519c21372b60609eb62a07a88712094ff3ba65205d744fcd9d4ddb WatchSource:0}: Error finding container e456bf1f51519c21372b60609eb62a07a88712094ff3ba65205d744fcd9d4ddb: Status 404 returned error can't find the container with id e456bf1f51519c21372b60609eb62a07a88712094ff3ba65205d744fcd9d4ddb Dec 01 16:25:12 crc kubenswrapper[4810]: I1201 16:25:12.925954 4810 generic.go:334] "Generic (PLEG): container finished" podID="e85267ea-ec74-4851-a062-05d2aa643a30" containerID="7ffc1a91a2d3d962a6447e51dafd2a8f25a606374f8237ede2b1d78037e8b50d" exitCode=0 Dec 01 16:25:12 crc kubenswrapper[4810]: I1201 16:25:12.926032 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" event={"ID":"e85267ea-ec74-4851-a062-05d2aa643a30","Type":"ContainerDied","Data":"7ffc1a91a2d3d962a6447e51dafd2a8f25a606374f8237ede2b1d78037e8b50d"} Dec 01 16:25:12 crc kubenswrapper[4810]: I1201 16:25:12.926376 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" event={"ID":"e85267ea-ec74-4851-a062-05d2aa643a30","Type":"ContainerStarted","Data":"e456bf1f51519c21372b60609eb62a07a88712094ff3ba65205d744fcd9d4ddb"} Dec 01 16:25:13 crc kubenswrapper[4810]: I1201 16:25:13.937452 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" event={"ID":"e85267ea-ec74-4851-a062-05d2aa643a30","Type":"ContainerStarted","Data":"d364219299f1865f970f8e38bf5f9fc34676576785a380ab0de83c9ec63eaf4d"} Dec 01 16:25:13 crc kubenswrapper[4810]: I1201 16:25:13.937989 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:13 crc kubenswrapper[4810]: I1201 16:25:13.962064 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" podStartSLOduration=2.962045388 podStartE2EDuration="2.962045388s" podCreationTimestamp="2025-12-01 16:25:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:25:13.959220701 +0000 UTC m=+6679.722730314" watchObservedRunningTime="2025-12-01 16:25:13.962045388 +0000 UTC m=+6679.725555001" Dec 01 16:25:14 crc kubenswrapper[4810]: I1201 16:25:14.813331 4810 scope.go:117] "RemoveContainer" containerID="e485960863af96767deee39cafeca708cce075ba3a5668045121fb9388bc3c7e" Dec 01 16:25:14 crc kubenswrapper[4810]: I1201 16:25:14.840450 4810 scope.go:117] "RemoveContainer" containerID="ad04d91114a4cde163687103220e8844b8b24c733d5779dd77f33abbcedb2121" Dec 01 16:25:14 crc kubenswrapper[4810]: I1201 16:25:14.921124 4810 scope.go:117] "RemoveContainer" containerID="93ab28f60c9265602b534910e42c28f3e5aad0b872db38a20f5be6c8fbf56c26" Dec 01 16:25:14 crc kubenswrapper[4810]: I1201 16:25:14.954188 4810 scope.go:117] "RemoveContainer" containerID="2330e916d744addde52cc40fdaefdfbe1ae4811254a7cf013517f884ba2016b6" Dec 01 16:25:15 crc kubenswrapper[4810]: I1201 16:25:15.004129 4810 scope.go:117] "RemoveContainer" containerID="1e1050cbe92252ae560a925d4edb7788b2d75040a21bbffd2bd8ad3eb7fcd5e8" Dec 01 16:25:15 crc kubenswrapper[4810]: I1201 16:25:15.054520 4810 scope.go:117] "RemoveContainer" containerID="685a66adc9ea7bc3944304fbd1630b0589690e6c8f287c51a0f91b25365ce8d4" Dec 01 16:25:15 crc kubenswrapper[4810]: I1201 16:25:15.108792 4810 scope.go:117] "RemoveContainer" containerID="51816684532c34a75051e9eaf7cc7f9a5a123b3b8686b2d8945d109e810b1eab" Dec 01 16:25:21 crc kubenswrapper[4810]: I1201 16:25:21.734606 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:21 crc kubenswrapper[4810]: I1201 16:25:21.823407 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-797f6f57f5-n86v7"] Dec 01 16:25:21 crc kubenswrapper[4810]: I1201 16:25:21.824976 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" podUID="4c524b0f-d099-489b-9675-a44654fc1c53" containerName="dnsmasq-dns" containerID="cri-o://461e34c10cc81aba237ad62874026ed718e6581607abf907f1cab8e9ae099d1a" gracePeriod=10 Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.003324 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5555dfd759-gjw5c"] Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.006109 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.035450 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5555dfd759-gjw5c"] Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.056875 4810 generic.go:334] "Generic (PLEG): container finished" podID="4c524b0f-d099-489b-9675-a44654fc1c53" containerID="461e34c10cc81aba237ad62874026ed718e6581607abf907f1cab8e9ae099d1a" exitCode=0 Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.056919 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" event={"ID":"4c524b0f-d099-489b-9675-a44654fc1c53","Type":"ContainerDied","Data":"461e34c10cc81aba237ad62874026ed718e6581607abf907f1cab8e9ae099d1a"} Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.181489 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cca49c71-6d05-4439-86a0-1e920aa27fc8-dns-svc\") pod \"dnsmasq-dns-5555dfd759-gjw5c\" (UID: \"cca49c71-6d05-4439-86a0-1e920aa27fc8\") " pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.181553 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/cca49c71-6d05-4439-86a0-1e920aa27fc8-openstack-cell1\") pod \"dnsmasq-dns-5555dfd759-gjw5c\" (UID: \"cca49c71-6d05-4439-86a0-1e920aa27fc8\") " pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.187906 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cca49c71-6d05-4439-86a0-1e920aa27fc8-config\") pod \"dnsmasq-dns-5555dfd759-gjw5c\" (UID: \"cca49c71-6d05-4439-86a0-1e920aa27fc8\") " pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.187989 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cca49c71-6d05-4439-86a0-1e920aa27fc8-ovsdbserver-nb\") pod \"dnsmasq-dns-5555dfd759-gjw5c\" (UID: \"cca49c71-6d05-4439-86a0-1e920aa27fc8\") " pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.188208 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qmms\" (UniqueName: \"kubernetes.io/projected/cca49c71-6d05-4439-86a0-1e920aa27fc8-kube-api-access-4qmms\") pod \"dnsmasq-dns-5555dfd759-gjw5c\" (UID: \"cca49c71-6d05-4439-86a0-1e920aa27fc8\") " pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.188280 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cca49c71-6d05-4439-86a0-1e920aa27fc8-ovsdbserver-sb\") pod \"dnsmasq-dns-5555dfd759-gjw5c\" (UID: \"cca49c71-6d05-4439-86a0-1e920aa27fc8\") " pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.291124 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cca49c71-6d05-4439-86a0-1e920aa27fc8-dns-svc\") pod \"dnsmasq-dns-5555dfd759-gjw5c\" (UID: \"cca49c71-6d05-4439-86a0-1e920aa27fc8\") " pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.291183 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/cca49c71-6d05-4439-86a0-1e920aa27fc8-openstack-cell1\") pod \"dnsmasq-dns-5555dfd759-gjw5c\" (UID: \"cca49c71-6d05-4439-86a0-1e920aa27fc8\") " pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.292296 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/cca49c71-6d05-4439-86a0-1e920aa27fc8-openstack-cell1\") pod \"dnsmasq-dns-5555dfd759-gjw5c\" (UID: \"cca49c71-6d05-4439-86a0-1e920aa27fc8\") " pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.292363 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cca49c71-6d05-4439-86a0-1e920aa27fc8-dns-svc\") pod \"dnsmasq-dns-5555dfd759-gjw5c\" (UID: \"cca49c71-6d05-4439-86a0-1e920aa27fc8\") " pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.292375 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cca49c71-6d05-4439-86a0-1e920aa27fc8-config\") pod \"dnsmasq-dns-5555dfd759-gjw5c\" (UID: \"cca49c71-6d05-4439-86a0-1e920aa27fc8\") " pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.292445 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cca49c71-6d05-4439-86a0-1e920aa27fc8-ovsdbserver-nb\") pod \"dnsmasq-dns-5555dfd759-gjw5c\" (UID: \"cca49c71-6d05-4439-86a0-1e920aa27fc8\") " pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.292681 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qmms\" (UniqueName: \"kubernetes.io/projected/cca49c71-6d05-4439-86a0-1e920aa27fc8-kube-api-access-4qmms\") pod \"dnsmasq-dns-5555dfd759-gjw5c\" (UID: \"cca49c71-6d05-4439-86a0-1e920aa27fc8\") " pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.292749 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cca49c71-6d05-4439-86a0-1e920aa27fc8-ovsdbserver-sb\") pod \"dnsmasq-dns-5555dfd759-gjw5c\" (UID: \"cca49c71-6d05-4439-86a0-1e920aa27fc8\") " pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.293004 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cca49c71-6d05-4439-86a0-1e920aa27fc8-config\") pod \"dnsmasq-dns-5555dfd759-gjw5c\" (UID: \"cca49c71-6d05-4439-86a0-1e920aa27fc8\") " pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.293793 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cca49c71-6d05-4439-86a0-1e920aa27fc8-ovsdbserver-sb\") pod \"dnsmasq-dns-5555dfd759-gjw5c\" (UID: \"cca49c71-6d05-4439-86a0-1e920aa27fc8\") " pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.294072 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cca49c71-6d05-4439-86a0-1e920aa27fc8-ovsdbserver-nb\") pod \"dnsmasq-dns-5555dfd759-gjw5c\" (UID: \"cca49c71-6d05-4439-86a0-1e920aa27fc8\") " pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.313080 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qmms\" (UniqueName: \"kubernetes.io/projected/cca49c71-6d05-4439-86a0-1e920aa27fc8-kube-api-access-4qmms\") pod \"dnsmasq-dns-5555dfd759-gjw5c\" (UID: \"cca49c71-6d05-4439-86a0-1e920aa27fc8\") " pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.336881 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.464002 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.497029 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-dns-svc\") pod \"4c524b0f-d099-489b-9675-a44654fc1c53\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.497098 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-ovsdbserver-sb\") pod \"4c524b0f-d099-489b-9675-a44654fc1c53\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.497130 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-config\") pod \"4c524b0f-d099-489b-9675-a44654fc1c53\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.497184 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzf5s\" (UniqueName: \"kubernetes.io/projected/4c524b0f-d099-489b-9675-a44654fc1c53-kube-api-access-qzf5s\") pod \"4c524b0f-d099-489b-9675-a44654fc1c53\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.498052 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-ovsdbserver-nb\") pod \"4c524b0f-d099-489b-9675-a44654fc1c53\" (UID: \"4c524b0f-d099-489b-9675-a44654fc1c53\") " Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.501357 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c524b0f-d099-489b-9675-a44654fc1c53-kube-api-access-qzf5s" (OuterVolumeSpecName: "kube-api-access-qzf5s") pod "4c524b0f-d099-489b-9675-a44654fc1c53" (UID: "4c524b0f-d099-489b-9675-a44654fc1c53"). InnerVolumeSpecName "kube-api-access-qzf5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.570143 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4c524b0f-d099-489b-9675-a44654fc1c53" (UID: "4c524b0f-d099-489b-9675-a44654fc1c53"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.572431 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4c524b0f-d099-489b-9675-a44654fc1c53" (UID: "4c524b0f-d099-489b-9675-a44654fc1c53"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.573401 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-config" (OuterVolumeSpecName: "config") pod "4c524b0f-d099-489b-9675-a44654fc1c53" (UID: "4c524b0f-d099-489b-9675-a44654fc1c53"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.584865 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4c524b0f-d099-489b-9675-a44654fc1c53" (UID: "4c524b0f-d099-489b-9675-a44654fc1c53"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.600753 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.600780 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.600790 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.600811 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c524b0f-d099-489b-9675-a44654fc1c53-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.600821 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzf5s\" (UniqueName: \"kubernetes.io/projected/4c524b0f-d099-489b-9675-a44654fc1c53-kube-api-access-qzf5s\") on node \"crc\" DevicePath \"\"" Dec 01 16:25:22 crc kubenswrapper[4810]: W1201 16:25:22.808993 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcca49c71_6d05_4439_86a0_1e920aa27fc8.slice/crio-f149687387bcae2460b3d5d928fe1418a4da13be7495498fcaf818e254eaf1d6 WatchSource:0}: Error finding container f149687387bcae2460b3d5d928fe1418a4da13be7495498fcaf818e254eaf1d6: Status 404 returned error can't find the container with id f149687387bcae2460b3d5d928fe1418a4da13be7495498fcaf818e254eaf1d6 Dec 01 16:25:22 crc kubenswrapper[4810]: I1201 16:25:22.809752 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5555dfd759-gjw5c"] Dec 01 16:25:23 crc kubenswrapper[4810]: I1201 16:25:23.069018 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" event={"ID":"4c524b0f-d099-489b-9675-a44654fc1c53","Type":"ContainerDied","Data":"2fa2a4a45b28c455713878d35c67cedc9caafed8ac22cf6a3bf2e2e1c394b22f"} Dec 01 16:25:23 crc kubenswrapper[4810]: I1201 16:25:23.069094 4810 scope.go:117] "RemoveContainer" containerID="461e34c10cc81aba237ad62874026ed718e6581607abf907f1cab8e9ae099d1a" Dec 01 16:25:23 crc kubenswrapper[4810]: I1201 16:25:23.069266 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797f6f57f5-n86v7" Dec 01 16:25:23 crc kubenswrapper[4810]: I1201 16:25:23.072443 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" event={"ID":"cca49c71-6d05-4439-86a0-1e920aa27fc8","Type":"ContainerStarted","Data":"f149687387bcae2460b3d5d928fe1418a4da13be7495498fcaf818e254eaf1d6"} Dec 01 16:25:23 crc kubenswrapper[4810]: I1201 16:25:23.096580 4810 scope.go:117] "RemoveContainer" containerID="a4d3a67c7d0b60488a4e6ca4913a6bb709ca4a854614390cf68e4af7b9d97eb2" Dec 01 16:25:23 crc kubenswrapper[4810]: I1201 16:25:23.115611 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-797f6f57f5-n86v7"] Dec 01 16:25:23 crc kubenswrapper[4810]: I1201 16:25:23.127585 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-797f6f57f5-n86v7"] Dec 01 16:25:24 crc kubenswrapper[4810]: I1201 16:25:24.082490 4810 generic.go:334] "Generic (PLEG): container finished" podID="cca49c71-6d05-4439-86a0-1e920aa27fc8" containerID="f9310acfa2dccc1e78169090c9c773d665c527ccb9deac5a204a9e765aa424b6" exitCode=0 Dec 01 16:25:24 crc kubenswrapper[4810]: I1201 16:25:24.082595 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" event={"ID":"cca49c71-6d05-4439-86a0-1e920aa27fc8","Type":"ContainerDied","Data":"f9310acfa2dccc1e78169090c9c773d665c527ccb9deac5a204a9e765aa424b6"} Dec 01 16:25:24 crc kubenswrapper[4810]: I1201 16:25:24.503404 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c524b0f-d099-489b-9675-a44654fc1c53" path="/var/lib/kubelet/pods/4c524b0f-d099-489b-9675-a44654fc1c53/volumes" Dec 01 16:25:25 crc kubenswrapper[4810]: I1201 16:25:25.046846 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2m2vb"] Dec 01 16:25:25 crc kubenswrapper[4810]: I1201 16:25:25.062704 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-2m2vb"] Dec 01 16:25:25 crc kubenswrapper[4810]: I1201 16:25:25.098275 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" event={"ID":"cca49c71-6d05-4439-86a0-1e920aa27fc8","Type":"ContainerStarted","Data":"c0d828b26c4a13faad460e5a1facef901745e96f081316009dc14563d5f1e779"} Dec 01 16:25:25 crc kubenswrapper[4810]: I1201 16:25:25.098483 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:25 crc kubenswrapper[4810]: I1201 16:25:25.120958 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" podStartSLOduration=4.120939178 podStartE2EDuration="4.120939178s" podCreationTimestamp="2025-12-01 16:25:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:25:25.115453789 +0000 UTC m=+6690.878963392" watchObservedRunningTime="2025-12-01 16:25:25.120939178 +0000 UTC m=+6690.884448781" Dec 01 16:25:26 crc kubenswrapper[4810]: I1201 16:25:26.507714 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95f99947-fa7b-43af-9d9d-640e6102771d" path="/var/lib/kubelet/pods/95f99947-fa7b-43af-9d9d-640e6102771d/volumes" Dec 01 16:25:27 crc kubenswrapper[4810]: I1201 16:25:27.032341 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-llq2h"] Dec 01 16:25:27 crc kubenswrapper[4810]: I1201 16:25:27.041781 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-llq2h"] Dec 01 16:25:28 crc kubenswrapper[4810]: I1201 16:25:28.508005 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37eef714-f1dc-47b3-8489-5204d08789bb" path="/var/lib/kubelet/pods/37eef714-f1dc-47b3-8489-5204d08789bb/volumes" Dec 01 16:25:32 crc kubenswrapper[4810]: I1201 16:25:32.338791 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5555dfd759-gjw5c" Dec 01 16:25:32 crc kubenswrapper[4810]: I1201 16:25:32.424490 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bf4588c47-pv8np"] Dec 01 16:25:32 crc kubenswrapper[4810]: I1201 16:25:32.425305 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" podUID="e85267ea-ec74-4851-a062-05d2aa643a30" containerName="dnsmasq-dns" containerID="cri-o://d364219299f1865f970f8e38bf5f9fc34676576785a380ab0de83c9ec63eaf4d" gracePeriod=10 Dec 01 16:25:32 crc kubenswrapper[4810]: I1201 16:25:32.945875 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:32 crc kubenswrapper[4810]: I1201 16:25:32.972934 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:25:32 crc kubenswrapper[4810]: I1201 16:25:32.972997 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.134707 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-ovsdbserver-nb\") pod \"e85267ea-ec74-4851-a062-05d2aa643a30\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.134777 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-ovsdbserver-sb\") pod \"e85267ea-ec74-4851-a062-05d2aa643a30\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.134825 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-openstack-cell1\") pod \"e85267ea-ec74-4851-a062-05d2aa643a30\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.134975 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-dns-svc\") pod \"e85267ea-ec74-4851-a062-05d2aa643a30\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.135031 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-config\") pod \"e85267ea-ec74-4851-a062-05d2aa643a30\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.135060 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xnnb\" (UniqueName: \"kubernetes.io/projected/e85267ea-ec74-4851-a062-05d2aa643a30-kube-api-access-9xnnb\") pod \"e85267ea-ec74-4851-a062-05d2aa643a30\" (UID: \"e85267ea-ec74-4851-a062-05d2aa643a30\") " Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.141587 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e85267ea-ec74-4851-a062-05d2aa643a30-kube-api-access-9xnnb" (OuterVolumeSpecName: "kube-api-access-9xnnb") pod "e85267ea-ec74-4851-a062-05d2aa643a30" (UID: "e85267ea-ec74-4851-a062-05d2aa643a30"). InnerVolumeSpecName "kube-api-access-9xnnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.201261 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-config" (OuterVolumeSpecName: "config") pod "e85267ea-ec74-4851-a062-05d2aa643a30" (UID: "e85267ea-ec74-4851-a062-05d2aa643a30"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.201371 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e85267ea-ec74-4851-a062-05d2aa643a30" (UID: "e85267ea-ec74-4851-a062-05d2aa643a30"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.202821 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "e85267ea-ec74-4851-a062-05d2aa643a30" (UID: "e85267ea-ec74-4851-a062-05d2aa643a30"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.207803 4810 generic.go:334] "Generic (PLEG): container finished" podID="e85267ea-ec74-4851-a062-05d2aa643a30" containerID="d364219299f1865f970f8e38bf5f9fc34676576785a380ab0de83c9ec63eaf4d" exitCode=0 Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.207847 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" event={"ID":"e85267ea-ec74-4851-a062-05d2aa643a30","Type":"ContainerDied","Data":"d364219299f1865f970f8e38bf5f9fc34676576785a380ab0de83c9ec63eaf4d"} Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.207874 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" event={"ID":"e85267ea-ec74-4851-a062-05d2aa643a30","Type":"ContainerDied","Data":"e456bf1f51519c21372b60609eb62a07a88712094ff3ba65205d744fcd9d4ddb"} Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.207904 4810 scope.go:117] "RemoveContainer" containerID="d364219299f1865f970f8e38bf5f9fc34676576785a380ab0de83c9ec63eaf4d" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.207910 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bf4588c47-pv8np" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.220386 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e85267ea-ec74-4851-a062-05d2aa643a30" (UID: "e85267ea-ec74-4851-a062-05d2aa643a30"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.232510 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e85267ea-ec74-4851-a062-05d2aa643a30" (UID: "e85267ea-ec74-4851-a062-05d2aa643a30"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.238171 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.238200 4810 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.238209 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-openstack-cell1\") on node \"crc\" DevicePath \"\"" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.238219 4810 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.238229 4810 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e85267ea-ec74-4851-a062-05d2aa643a30-config\") on node \"crc\" DevicePath \"\"" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.238239 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xnnb\" (UniqueName: \"kubernetes.io/projected/e85267ea-ec74-4851-a062-05d2aa643a30-kube-api-access-9xnnb\") on node \"crc\" DevicePath \"\"" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.314234 4810 scope.go:117] "RemoveContainer" containerID="7ffc1a91a2d3d962a6447e51dafd2a8f25a606374f8237ede2b1d78037e8b50d" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.342662 4810 scope.go:117] "RemoveContainer" containerID="d364219299f1865f970f8e38bf5f9fc34676576785a380ab0de83c9ec63eaf4d" Dec 01 16:25:33 crc kubenswrapper[4810]: E1201 16:25:33.343232 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d364219299f1865f970f8e38bf5f9fc34676576785a380ab0de83c9ec63eaf4d\": container with ID starting with d364219299f1865f970f8e38bf5f9fc34676576785a380ab0de83c9ec63eaf4d not found: ID does not exist" containerID="d364219299f1865f970f8e38bf5f9fc34676576785a380ab0de83c9ec63eaf4d" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.343270 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d364219299f1865f970f8e38bf5f9fc34676576785a380ab0de83c9ec63eaf4d"} err="failed to get container status \"d364219299f1865f970f8e38bf5f9fc34676576785a380ab0de83c9ec63eaf4d\": rpc error: code = NotFound desc = could not find container \"d364219299f1865f970f8e38bf5f9fc34676576785a380ab0de83c9ec63eaf4d\": container with ID starting with d364219299f1865f970f8e38bf5f9fc34676576785a380ab0de83c9ec63eaf4d not found: ID does not exist" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.343294 4810 scope.go:117] "RemoveContainer" containerID="7ffc1a91a2d3d962a6447e51dafd2a8f25a606374f8237ede2b1d78037e8b50d" Dec 01 16:25:33 crc kubenswrapper[4810]: E1201 16:25:33.343662 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ffc1a91a2d3d962a6447e51dafd2a8f25a606374f8237ede2b1d78037e8b50d\": container with ID starting with 7ffc1a91a2d3d962a6447e51dafd2a8f25a606374f8237ede2b1d78037e8b50d not found: ID does not exist" containerID="7ffc1a91a2d3d962a6447e51dafd2a8f25a606374f8237ede2b1d78037e8b50d" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.343690 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ffc1a91a2d3d962a6447e51dafd2a8f25a606374f8237ede2b1d78037e8b50d"} err="failed to get container status \"7ffc1a91a2d3d962a6447e51dafd2a8f25a606374f8237ede2b1d78037e8b50d\": rpc error: code = NotFound desc = could not find container \"7ffc1a91a2d3d962a6447e51dafd2a8f25a606374f8237ede2b1d78037e8b50d\": container with ID starting with 7ffc1a91a2d3d962a6447e51dafd2a8f25a606374f8237ede2b1d78037e8b50d not found: ID does not exist" Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.544107 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bf4588c47-pv8np"] Dec 01 16:25:33 crc kubenswrapper[4810]: I1201 16:25:33.553022 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bf4588c47-pv8np"] Dec 01 16:25:34 crc kubenswrapper[4810]: I1201 16:25:34.501423 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e85267ea-ec74-4851-a062-05d2aa643a30" path="/var/lib/kubelet/pods/e85267ea-ec74-4851-a062-05d2aa643a30/volumes" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.038268 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5"] Dec 01 16:25:43 crc kubenswrapper[4810]: E1201 16:25:43.039364 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e85267ea-ec74-4851-a062-05d2aa643a30" containerName="init" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.039381 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e85267ea-ec74-4851-a062-05d2aa643a30" containerName="init" Dec 01 16:25:43 crc kubenswrapper[4810]: E1201 16:25:43.039411 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c524b0f-d099-489b-9675-a44654fc1c53" containerName="dnsmasq-dns" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.039420 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c524b0f-d099-489b-9675-a44654fc1c53" containerName="dnsmasq-dns" Dec 01 16:25:43 crc kubenswrapper[4810]: E1201 16:25:43.039442 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c524b0f-d099-489b-9675-a44654fc1c53" containerName="init" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.039450 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c524b0f-d099-489b-9675-a44654fc1c53" containerName="init" Dec 01 16:25:43 crc kubenswrapper[4810]: E1201 16:25:43.039466 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e85267ea-ec74-4851-a062-05d2aa643a30" containerName="dnsmasq-dns" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.039496 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e85267ea-ec74-4851-a062-05d2aa643a30" containerName="dnsmasq-dns" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.039727 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c524b0f-d099-489b-9675-a44654fc1c53" containerName="dnsmasq-dns" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.039754 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e85267ea-ec74-4851-a062-05d2aa643a30" containerName="dnsmasq-dns" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.041098 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.043936 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.044421 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.044685 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.044942 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.065123 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5"] Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.165993 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c6da269-a52f-427a-a457-0e6fe712f398-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5\" (UID: \"6c6da269-a52f-427a-a457-0e6fe712f398\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.166440 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjxvt\" (UniqueName: \"kubernetes.io/projected/6c6da269-a52f-427a-a457-0e6fe712f398-kube-api-access-hjxvt\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5\" (UID: \"6c6da269-a52f-427a-a457-0e6fe712f398\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.166732 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c6da269-a52f-427a-a457-0e6fe712f398-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5\" (UID: \"6c6da269-a52f-427a-a457-0e6fe712f398\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.166957 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c6da269-a52f-427a-a457-0e6fe712f398-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5\" (UID: \"6c6da269-a52f-427a-a457-0e6fe712f398\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.269254 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c6da269-a52f-427a-a457-0e6fe712f398-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5\" (UID: \"6c6da269-a52f-427a-a457-0e6fe712f398\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.269346 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c6da269-a52f-427a-a457-0e6fe712f398-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5\" (UID: \"6c6da269-a52f-427a-a457-0e6fe712f398\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.269431 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjxvt\" (UniqueName: \"kubernetes.io/projected/6c6da269-a52f-427a-a457-0e6fe712f398-kube-api-access-hjxvt\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5\" (UID: \"6c6da269-a52f-427a-a457-0e6fe712f398\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.269561 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c6da269-a52f-427a-a457-0e6fe712f398-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5\" (UID: \"6c6da269-a52f-427a-a457-0e6fe712f398\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.275690 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c6da269-a52f-427a-a457-0e6fe712f398-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5\" (UID: \"6c6da269-a52f-427a-a457-0e6fe712f398\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.276426 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c6da269-a52f-427a-a457-0e6fe712f398-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5\" (UID: \"6c6da269-a52f-427a-a457-0e6fe712f398\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.277030 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c6da269-a52f-427a-a457-0e6fe712f398-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5\" (UID: \"6c6da269-a52f-427a-a457-0e6fe712f398\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.285861 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjxvt\" (UniqueName: \"kubernetes.io/projected/6c6da269-a52f-427a-a457-0e6fe712f398-kube-api-access-hjxvt\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5\" (UID: \"6c6da269-a52f-427a-a457-0e6fe712f398\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" Dec 01 16:25:43 crc kubenswrapper[4810]: I1201 16:25:43.366119 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" Dec 01 16:25:44 crc kubenswrapper[4810]: I1201 16:25:44.035563 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5"] Dec 01 16:25:44 crc kubenswrapper[4810]: W1201 16:25:44.048970 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c6da269_a52f_427a_a457_0e6fe712f398.slice/crio-2c99b68bb9ba96e1737e082ac39bd8a2c741f1c8bed7770c9a3764e1cb1c525a WatchSource:0}: Error finding container 2c99b68bb9ba96e1737e082ac39bd8a2c741f1c8bed7770c9a3764e1cb1c525a: Status 404 returned error can't find the container with id 2c99b68bb9ba96e1737e082ac39bd8a2c741f1c8bed7770c9a3764e1cb1c525a Dec 01 16:25:44 crc kubenswrapper[4810]: I1201 16:25:44.051824 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-7pr5v"] Dec 01 16:25:44 crc kubenswrapper[4810]: I1201 16:25:44.066034 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-7pr5v"] Dec 01 16:25:44 crc kubenswrapper[4810]: I1201 16:25:44.330315 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" event={"ID":"6c6da269-a52f-427a-a457-0e6fe712f398","Type":"ContainerStarted","Data":"2c99b68bb9ba96e1737e082ac39bd8a2c741f1c8bed7770c9a3764e1cb1c525a"} Dec 01 16:25:44 crc kubenswrapper[4810]: I1201 16:25:44.505893 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36bd4279-0ed0-4266-953a-7535b1653ec2" path="/var/lib/kubelet/pods/36bd4279-0ed0-4266-953a-7535b1653ec2/volumes" Dec 01 16:25:53 crc kubenswrapper[4810]: I1201 16:25:53.443872 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" event={"ID":"6c6da269-a52f-427a-a457-0e6fe712f398","Type":"ContainerStarted","Data":"8a986f4d7f9803d8ce5ccbd92376835e4f8851d31c0289c838d9400492a6eabc"} Dec 01 16:26:02 crc kubenswrapper[4810]: I1201 16:26:02.972853 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:26:02 crc kubenswrapper[4810]: I1201 16:26:02.973601 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:26:02 crc kubenswrapper[4810]: I1201 16:26:02.973669 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 16:26:02 crc kubenswrapper[4810]: I1201 16:26:02.975013 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:26:02 crc kubenswrapper[4810]: I1201 16:26:02.975130 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" gracePeriod=600 Dec 01 16:26:03 crc kubenswrapper[4810]: E1201 16:26:03.131830 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:26:03 crc kubenswrapper[4810]: I1201 16:26:03.563811 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" exitCode=0 Dec 01 16:26:03 crc kubenswrapper[4810]: I1201 16:26:03.563870 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b"} Dec 01 16:26:03 crc kubenswrapper[4810]: I1201 16:26:03.563907 4810 scope.go:117] "RemoveContainer" containerID="13b8d9a5912bf37c10017c37926041fb03894967f1288822928bc2736df5d8df" Dec 01 16:26:03 crc kubenswrapper[4810]: I1201 16:26:03.564740 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:26:03 crc kubenswrapper[4810]: E1201 16:26:03.565068 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:26:03 crc kubenswrapper[4810]: I1201 16:26:03.593934 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" podStartSLOduration=12.335744855 podStartE2EDuration="20.593913978s" podCreationTimestamp="2025-12-01 16:25:43 +0000 UTC" firstStartedPulling="2025-12-01 16:25:44.055250132 +0000 UTC m=+6709.818759756" lastFinishedPulling="2025-12-01 16:25:52.313419246 +0000 UTC m=+6718.076928879" observedRunningTime="2025-12-01 16:25:53.470316393 +0000 UTC m=+6719.233826086" watchObservedRunningTime="2025-12-01 16:26:03.593913978 +0000 UTC m=+6729.357423581" Dec 01 16:26:05 crc kubenswrapper[4810]: I1201 16:26:05.589792 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c6da269-a52f-427a-a457-0e6fe712f398" containerID="8a986f4d7f9803d8ce5ccbd92376835e4f8851d31c0289c838d9400492a6eabc" exitCode=0 Dec 01 16:26:05 crc kubenswrapper[4810]: I1201 16:26:05.589873 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" event={"ID":"6c6da269-a52f-427a-a457-0e6fe712f398","Type":"ContainerDied","Data":"8a986f4d7f9803d8ce5ccbd92376835e4f8851d31c0289c838d9400492a6eabc"} Dec 01 16:26:07 crc kubenswrapper[4810]: I1201 16:26:07.092260 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" Dec 01 16:26:07 crc kubenswrapper[4810]: I1201 16:26:07.197873 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c6da269-a52f-427a-a457-0e6fe712f398-pre-adoption-validation-combined-ca-bundle\") pod \"6c6da269-a52f-427a-a457-0e6fe712f398\" (UID: \"6c6da269-a52f-427a-a457-0e6fe712f398\") " Dec 01 16:26:07 crc kubenswrapper[4810]: I1201 16:26:07.197972 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjxvt\" (UniqueName: \"kubernetes.io/projected/6c6da269-a52f-427a-a457-0e6fe712f398-kube-api-access-hjxvt\") pod \"6c6da269-a52f-427a-a457-0e6fe712f398\" (UID: \"6c6da269-a52f-427a-a457-0e6fe712f398\") " Dec 01 16:26:07 crc kubenswrapper[4810]: I1201 16:26:07.198457 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c6da269-a52f-427a-a457-0e6fe712f398-ssh-key\") pod \"6c6da269-a52f-427a-a457-0e6fe712f398\" (UID: \"6c6da269-a52f-427a-a457-0e6fe712f398\") " Dec 01 16:26:07 crc kubenswrapper[4810]: I1201 16:26:07.198617 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c6da269-a52f-427a-a457-0e6fe712f398-inventory\") pod \"6c6da269-a52f-427a-a457-0e6fe712f398\" (UID: \"6c6da269-a52f-427a-a457-0e6fe712f398\") " Dec 01 16:26:07 crc kubenswrapper[4810]: I1201 16:26:07.210313 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c6da269-a52f-427a-a457-0e6fe712f398-kube-api-access-hjxvt" (OuterVolumeSpecName: "kube-api-access-hjxvt") pod "6c6da269-a52f-427a-a457-0e6fe712f398" (UID: "6c6da269-a52f-427a-a457-0e6fe712f398"). InnerVolumeSpecName "kube-api-access-hjxvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:26:07 crc kubenswrapper[4810]: I1201 16:26:07.221879 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c6da269-a52f-427a-a457-0e6fe712f398-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "6c6da269-a52f-427a-a457-0e6fe712f398" (UID: "6c6da269-a52f-427a-a457-0e6fe712f398"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:07 crc kubenswrapper[4810]: I1201 16:26:07.231633 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c6da269-a52f-427a-a457-0e6fe712f398-inventory" (OuterVolumeSpecName: "inventory") pod "6c6da269-a52f-427a-a457-0e6fe712f398" (UID: "6c6da269-a52f-427a-a457-0e6fe712f398"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:07 crc kubenswrapper[4810]: I1201 16:26:07.275663 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c6da269-a52f-427a-a457-0e6fe712f398-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6c6da269-a52f-427a-a457-0e6fe712f398" (UID: "6c6da269-a52f-427a-a457-0e6fe712f398"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:26:07 crc kubenswrapper[4810]: I1201 16:26:07.306184 4810 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c6da269-a52f-427a-a457-0e6fe712f398-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:07 crc kubenswrapper[4810]: I1201 16:26:07.306232 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjxvt\" (UniqueName: \"kubernetes.io/projected/6c6da269-a52f-427a-a457-0e6fe712f398-kube-api-access-hjxvt\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:07 crc kubenswrapper[4810]: I1201 16:26:07.306249 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6c6da269-a52f-427a-a457-0e6fe712f398-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:07 crc kubenswrapper[4810]: I1201 16:26:07.306263 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6c6da269-a52f-427a-a457-0e6fe712f398-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:26:07 crc kubenswrapper[4810]: I1201 16:26:07.612326 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" event={"ID":"6c6da269-a52f-427a-a457-0e6fe712f398","Type":"ContainerDied","Data":"2c99b68bb9ba96e1737e082ac39bd8a2c741f1c8bed7770c9a3764e1cb1c525a"} Dec 01 16:26:07 crc kubenswrapper[4810]: I1201 16:26:07.612681 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c99b68bb9ba96e1737e082ac39bd8a2c741f1c8bed7770c9a3764e1cb1c525a" Dec 01 16:26:07 crc kubenswrapper[4810]: I1201 16:26:07.612444 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5" Dec 01 16:26:15 crc kubenswrapper[4810]: I1201 16:26:15.371814 4810 scope.go:117] "RemoveContainer" containerID="2110fa6f8420d02721840cf754957f5be6b1e0af350c6d0523a83fddf56d0682" Dec 01 16:26:15 crc kubenswrapper[4810]: I1201 16:26:15.403788 4810 scope.go:117] "RemoveContainer" containerID="550ed3cf881185aa310513f0a4de4346116cfc8e8b9498af1450fd43e74377fa" Dec 01 16:26:15 crc kubenswrapper[4810]: I1201 16:26:15.460162 4810 scope.go:117] "RemoveContainer" containerID="ff0b3f547bd0fe4827488b9725fc4957403a6bc1f94a84d700e80db21c809c1e" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.027140 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b"] Dec 01 16:26:16 crc kubenswrapper[4810]: E1201 16:26:16.028710 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c6da269-a52f-427a-a457-0e6fe712f398" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.028739 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c6da269-a52f-427a-a457-0e6fe712f398" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.035372 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c6da269-a52f-427a-a457-0e6fe712f398" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.036622 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.038367 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.039987 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.040066 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.040503 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.056343 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b"] Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.178127 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/229207e0-27f2-4caa-b587-8340921ac643-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b\" (UID: \"229207e0-27f2-4caa-b587-8340921ac643\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.178639 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/229207e0-27f2-4caa-b587-8340921ac643-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b\" (UID: \"229207e0-27f2-4caa-b587-8340921ac643\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.178735 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvz4m\" (UniqueName: \"kubernetes.io/projected/229207e0-27f2-4caa-b587-8340921ac643-kube-api-access-nvz4m\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b\" (UID: \"229207e0-27f2-4caa-b587-8340921ac643\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.178885 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/229207e0-27f2-4caa-b587-8340921ac643-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b\" (UID: \"229207e0-27f2-4caa-b587-8340921ac643\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.281843 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/229207e0-27f2-4caa-b587-8340921ac643-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b\" (UID: \"229207e0-27f2-4caa-b587-8340921ac643\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.282062 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/229207e0-27f2-4caa-b587-8340921ac643-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b\" (UID: \"229207e0-27f2-4caa-b587-8340921ac643\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.282131 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvz4m\" (UniqueName: \"kubernetes.io/projected/229207e0-27f2-4caa-b587-8340921ac643-kube-api-access-nvz4m\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b\" (UID: \"229207e0-27f2-4caa-b587-8340921ac643\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.282177 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/229207e0-27f2-4caa-b587-8340921ac643-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b\" (UID: \"229207e0-27f2-4caa-b587-8340921ac643\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.288841 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/229207e0-27f2-4caa-b587-8340921ac643-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b\" (UID: \"229207e0-27f2-4caa-b587-8340921ac643\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.288969 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/229207e0-27f2-4caa-b587-8340921ac643-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b\" (UID: \"229207e0-27f2-4caa-b587-8340921ac643\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.289410 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/229207e0-27f2-4caa-b587-8340921ac643-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b\" (UID: \"229207e0-27f2-4caa-b587-8340921ac643\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.299797 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvz4m\" (UniqueName: \"kubernetes.io/projected/229207e0-27f2-4caa-b587-8340921ac643-kube-api-access-nvz4m\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b\" (UID: \"229207e0-27f2-4caa-b587-8340921ac643\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.360778 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" Dec 01 16:26:16 crc kubenswrapper[4810]: I1201 16:26:16.969249 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b"] Dec 01 16:26:16 crc kubenswrapper[4810]: W1201 16:26:16.979007 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod229207e0_27f2_4caa_b587_8340921ac643.slice/crio-dd65f296efe42ee8cd2c3b9e78114c8c8cd4ecb0f93a493d067a29a8dc5e63ef WatchSource:0}: Error finding container dd65f296efe42ee8cd2c3b9e78114c8c8cd4ecb0f93a493d067a29a8dc5e63ef: Status 404 returned error can't find the container with id dd65f296efe42ee8cd2c3b9e78114c8c8cd4ecb0f93a493d067a29a8dc5e63ef Dec 01 16:26:17 crc kubenswrapper[4810]: I1201 16:26:17.720669 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" event={"ID":"229207e0-27f2-4caa-b587-8340921ac643","Type":"ContainerStarted","Data":"2ecb722d2226dab5e716ac417d6697dec1117117eeac579d3f08010cf0352e74"} Dec 01 16:26:17 crc kubenswrapper[4810]: I1201 16:26:17.721201 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" event={"ID":"229207e0-27f2-4caa-b587-8340921ac643","Type":"ContainerStarted","Data":"dd65f296efe42ee8cd2c3b9e78114c8c8cd4ecb0f93a493d067a29a8dc5e63ef"} Dec 01 16:26:17 crc kubenswrapper[4810]: I1201 16:26:17.744355 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" podStartSLOduration=2.293601571 podStartE2EDuration="2.744338737s" podCreationTimestamp="2025-12-01 16:26:15 +0000 UTC" firstStartedPulling="2025-12-01 16:26:16.982263139 +0000 UTC m=+6742.745772742" lastFinishedPulling="2025-12-01 16:26:17.433000305 +0000 UTC m=+6743.196509908" observedRunningTime="2025-12-01 16:26:17.741360946 +0000 UTC m=+6743.504870569" watchObservedRunningTime="2025-12-01 16:26:17.744338737 +0000 UTC m=+6743.507848340" Dec 01 16:26:18 crc kubenswrapper[4810]: I1201 16:26:18.491601 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:26:18 crc kubenswrapper[4810]: E1201 16:26:18.492027 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:26:31 crc kubenswrapper[4810]: I1201 16:26:31.491197 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:26:31 crc kubenswrapper[4810]: E1201 16:26:31.491926 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:26:46 crc kubenswrapper[4810]: I1201 16:26:46.523297 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:26:46 crc kubenswrapper[4810]: E1201 16:26:46.524303 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:26:57 crc kubenswrapper[4810]: I1201 16:26:57.055712 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tk26t"] Dec 01 16:26:57 crc kubenswrapper[4810]: I1201 16:26:57.059650 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tk26t" Dec 01 16:26:57 crc kubenswrapper[4810]: I1201 16:26:57.074175 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tk26t"] Dec 01 16:26:57 crc kubenswrapper[4810]: I1201 16:26:57.204284 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm94z\" (UniqueName: \"kubernetes.io/projected/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294-kube-api-access-zm94z\") pod \"community-operators-tk26t\" (UID: \"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294\") " pod="openshift-marketplace/community-operators-tk26t" Dec 01 16:26:57 crc kubenswrapper[4810]: I1201 16:26:57.204414 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294-catalog-content\") pod \"community-operators-tk26t\" (UID: \"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294\") " pod="openshift-marketplace/community-operators-tk26t" Dec 01 16:26:57 crc kubenswrapper[4810]: I1201 16:26:57.204443 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294-utilities\") pod \"community-operators-tk26t\" (UID: \"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294\") " pod="openshift-marketplace/community-operators-tk26t" Dec 01 16:26:57 crc kubenswrapper[4810]: I1201 16:26:57.306570 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm94z\" (UniqueName: \"kubernetes.io/projected/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294-kube-api-access-zm94z\") pod \"community-operators-tk26t\" (UID: \"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294\") " pod="openshift-marketplace/community-operators-tk26t" Dec 01 16:26:57 crc kubenswrapper[4810]: I1201 16:26:57.307060 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294-catalog-content\") pod \"community-operators-tk26t\" (UID: \"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294\") " pod="openshift-marketplace/community-operators-tk26t" Dec 01 16:26:57 crc kubenswrapper[4810]: I1201 16:26:57.307172 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294-utilities\") pod \"community-operators-tk26t\" (UID: \"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294\") " pod="openshift-marketplace/community-operators-tk26t" Dec 01 16:26:57 crc kubenswrapper[4810]: I1201 16:26:57.307601 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294-catalog-content\") pod \"community-operators-tk26t\" (UID: \"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294\") " pod="openshift-marketplace/community-operators-tk26t" Dec 01 16:26:57 crc kubenswrapper[4810]: I1201 16:26:57.307673 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294-utilities\") pod \"community-operators-tk26t\" (UID: \"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294\") " pod="openshift-marketplace/community-operators-tk26t" Dec 01 16:26:57 crc kubenswrapper[4810]: I1201 16:26:57.328260 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm94z\" (UniqueName: \"kubernetes.io/projected/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294-kube-api-access-zm94z\") pod \"community-operators-tk26t\" (UID: \"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294\") " pod="openshift-marketplace/community-operators-tk26t" Dec 01 16:26:57 crc kubenswrapper[4810]: I1201 16:26:57.390549 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tk26t" Dec 01 16:26:57 crc kubenswrapper[4810]: I1201 16:26:57.491401 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:26:57 crc kubenswrapper[4810]: E1201 16:26:57.491779 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:26:58 crc kubenswrapper[4810]: I1201 16:26:58.036062 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tk26t"] Dec 01 16:26:58 crc kubenswrapper[4810]: I1201 16:26:58.131463 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tk26t" event={"ID":"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294","Type":"ContainerStarted","Data":"fc8c8e65ed9702afefd6dc4f4a1c2b17878727012a04dca2314c72d9c522893f"} Dec 01 16:26:59 crc kubenswrapper[4810]: I1201 16:26:59.141233 4810 generic.go:334] "Generic (PLEG): container finished" podID="f81e6c36-a33b-4fe5-b8b5-8dac22d2b294" containerID="8e1e5a8499f3aa9f99b42a7ba70317e61e3a44304043401205662bce97cd81ed" exitCode=0 Dec 01 16:26:59 crc kubenswrapper[4810]: I1201 16:26:59.141345 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tk26t" event={"ID":"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294","Type":"ContainerDied","Data":"8e1e5a8499f3aa9f99b42a7ba70317e61e3a44304043401205662bce97cd81ed"} Dec 01 16:27:04 crc kubenswrapper[4810]: I1201 16:27:04.041116 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-mggf6"] Dec 01 16:27:04 crc kubenswrapper[4810]: I1201 16:27:04.054906 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-mggf6"] Dec 01 16:27:04 crc kubenswrapper[4810]: I1201 16:27:04.506865 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7db4342d-31f0-48c3-8c8c-eccc2a1324ad" path="/var/lib/kubelet/pods/7db4342d-31f0-48c3-8c8c-eccc2a1324ad/volumes" Dec 01 16:27:05 crc kubenswrapper[4810]: I1201 16:27:05.253636 4810 generic.go:334] "Generic (PLEG): container finished" podID="f81e6c36-a33b-4fe5-b8b5-8dac22d2b294" containerID="0e27029a7946e008d5d71e53dafcf6dba211589fcfefa68211bfa645dd869121" exitCode=0 Dec 01 16:27:05 crc kubenswrapper[4810]: I1201 16:27:05.253986 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tk26t" event={"ID":"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294","Type":"ContainerDied","Data":"0e27029a7946e008d5d71e53dafcf6dba211589fcfefa68211bfa645dd869121"} Dec 01 16:27:06 crc kubenswrapper[4810]: I1201 16:27:06.035187 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-1e8e-account-create-update-9mvxm"] Dec 01 16:27:06 crc kubenswrapper[4810]: I1201 16:27:06.047908 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-1e8e-account-create-update-9mvxm"] Dec 01 16:27:06 crc kubenswrapper[4810]: I1201 16:27:06.266974 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tk26t" event={"ID":"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294","Type":"ContainerStarted","Data":"74dbdf8d08b57b7688e6482d0e8774d960ad696ee84fc73c595f66a8de30e45a"} Dec 01 16:27:06 crc kubenswrapper[4810]: I1201 16:27:06.286034 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tk26t" podStartSLOduration=2.427612378 podStartE2EDuration="9.286015823s" podCreationTimestamp="2025-12-01 16:26:57 +0000 UTC" firstStartedPulling="2025-12-01 16:26:59.143326661 +0000 UTC m=+6784.906836264" lastFinishedPulling="2025-12-01 16:27:06.001730096 +0000 UTC m=+6791.765239709" observedRunningTime="2025-12-01 16:27:06.283875565 +0000 UTC m=+6792.047385178" watchObservedRunningTime="2025-12-01 16:27:06.286015823 +0000 UTC m=+6792.049525426" Dec 01 16:27:06 crc kubenswrapper[4810]: I1201 16:27:06.502566 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c02cfaaf-96b5-444f-a03f-e4421bf3b8a1" path="/var/lib/kubelet/pods/c02cfaaf-96b5-444f-a03f-e4421bf3b8a1/volumes" Dec 01 16:27:07 crc kubenswrapper[4810]: I1201 16:27:07.391184 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tk26t" Dec 01 16:27:07 crc kubenswrapper[4810]: I1201 16:27:07.391603 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tk26t" Dec 01 16:27:08 crc kubenswrapper[4810]: I1201 16:27:08.449168 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-tk26t" podUID="f81e6c36-a33b-4fe5-b8b5-8dac22d2b294" containerName="registry-server" probeResult="failure" output=< Dec 01 16:27:08 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Dec 01 16:27:08 crc kubenswrapper[4810]: > Dec 01 16:27:11 crc kubenswrapper[4810]: I1201 16:27:11.491387 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:27:11 crc kubenswrapper[4810]: E1201 16:27:11.491939 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:27:12 crc kubenswrapper[4810]: I1201 16:27:12.027830 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-pxplj"] Dec 01 16:27:12 crc kubenswrapper[4810]: I1201 16:27:12.037297 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-pxplj"] Dec 01 16:27:12 crc kubenswrapper[4810]: I1201 16:27:12.506092 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9edab071-f11b-4afe-a2a9-b63b96c4069c" path="/var/lib/kubelet/pods/9edab071-f11b-4afe-a2a9-b63b96c4069c/volumes" Dec 01 16:27:13 crc kubenswrapper[4810]: I1201 16:27:13.031898 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-11c5-account-create-update-b6n7h"] Dec 01 16:27:13 crc kubenswrapper[4810]: I1201 16:27:13.044411 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-11c5-account-create-update-b6n7h"] Dec 01 16:27:14 crc kubenswrapper[4810]: I1201 16:27:14.510795 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="188b6e42-b278-4777-9be3-d86ffe3c16a5" path="/var/lib/kubelet/pods/188b6e42-b278-4777-9be3-d86ffe3c16a5/volumes" Dec 01 16:27:15 crc kubenswrapper[4810]: I1201 16:27:15.602410 4810 scope.go:117] "RemoveContainer" containerID="cdb7bdf10e9e9ee7431f3d188d36e0aab4e07e9981b9931f20479e6933e688f4" Dec 01 16:27:15 crc kubenswrapper[4810]: I1201 16:27:15.641694 4810 scope.go:117] "RemoveContainer" containerID="eb880792a3c2fbc18060174543da6970f79bc73c96c926f83aa10ec402bbbd6f" Dec 01 16:27:15 crc kubenswrapper[4810]: I1201 16:27:15.706710 4810 scope.go:117] "RemoveContainer" containerID="f3878656db426ae1896e62188b3c9738ffa6c6a41583f5419dd40bb9018b9ad2" Dec 01 16:27:15 crc kubenswrapper[4810]: I1201 16:27:15.766360 4810 scope.go:117] "RemoveContainer" containerID="9b1bdb512c45ec6b07574c4ce36454471f78dd7ab8ff4a6e6fbf03b8f26a6008" Dec 01 16:27:17 crc kubenswrapper[4810]: I1201 16:27:17.442268 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tk26t" Dec 01 16:27:17 crc kubenswrapper[4810]: I1201 16:27:17.496458 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tk26t" Dec 01 16:27:17 crc kubenswrapper[4810]: I1201 16:27:17.574379 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tk26t"] Dec 01 16:27:17 crc kubenswrapper[4810]: I1201 16:27:17.685418 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gd5zz"] Dec 01 16:27:17 crc kubenswrapper[4810]: I1201 16:27:17.685947 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gd5zz" podUID="9e8daf3c-58ed-4feb-b574-8bda66dd9c79" containerName="registry-server" containerID="cri-o://44632dd9584a65c8447e93ce8db198c67c23e6cb0e0e4361c92967632df77581" gracePeriod=2 Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.207306 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gd5zz" Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.282826 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jt72v\" (UniqueName: \"kubernetes.io/projected/9e8daf3c-58ed-4feb-b574-8bda66dd9c79-kube-api-access-jt72v\") pod \"9e8daf3c-58ed-4feb-b574-8bda66dd9c79\" (UID: \"9e8daf3c-58ed-4feb-b574-8bda66dd9c79\") " Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.282933 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e8daf3c-58ed-4feb-b574-8bda66dd9c79-catalog-content\") pod \"9e8daf3c-58ed-4feb-b574-8bda66dd9c79\" (UID: \"9e8daf3c-58ed-4feb-b574-8bda66dd9c79\") " Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.282965 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e8daf3c-58ed-4feb-b574-8bda66dd9c79-utilities\") pod \"9e8daf3c-58ed-4feb-b574-8bda66dd9c79\" (UID: \"9e8daf3c-58ed-4feb-b574-8bda66dd9c79\") " Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.283782 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e8daf3c-58ed-4feb-b574-8bda66dd9c79-utilities" (OuterVolumeSpecName: "utilities") pod "9e8daf3c-58ed-4feb-b574-8bda66dd9c79" (UID: "9e8daf3c-58ed-4feb-b574-8bda66dd9c79"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.291391 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e8daf3c-58ed-4feb-b574-8bda66dd9c79-kube-api-access-jt72v" (OuterVolumeSpecName: "kube-api-access-jt72v") pod "9e8daf3c-58ed-4feb-b574-8bda66dd9c79" (UID: "9e8daf3c-58ed-4feb-b574-8bda66dd9c79"). InnerVolumeSpecName "kube-api-access-jt72v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.325522 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e8daf3c-58ed-4feb-b574-8bda66dd9c79-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e8daf3c-58ed-4feb-b574-8bda66dd9c79" (UID: "9e8daf3c-58ed-4feb-b574-8bda66dd9c79"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.385019 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jt72v\" (UniqueName: \"kubernetes.io/projected/9e8daf3c-58ed-4feb-b574-8bda66dd9c79-kube-api-access-jt72v\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.385058 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e8daf3c-58ed-4feb-b574-8bda66dd9c79-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.385072 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e8daf3c-58ed-4feb-b574-8bda66dd9c79-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.387542 4810 generic.go:334] "Generic (PLEG): container finished" podID="9e8daf3c-58ed-4feb-b574-8bda66dd9c79" containerID="44632dd9584a65c8447e93ce8db198c67c23e6cb0e0e4361c92967632df77581" exitCode=0 Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.387575 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gd5zz" event={"ID":"9e8daf3c-58ed-4feb-b574-8bda66dd9c79","Type":"ContainerDied","Data":"44632dd9584a65c8447e93ce8db198c67c23e6cb0e0e4361c92967632df77581"} Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.387628 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gd5zz" event={"ID":"9e8daf3c-58ed-4feb-b574-8bda66dd9c79","Type":"ContainerDied","Data":"95b782f30c7ed0b4957bc75b4b410c18537f624234aa421300686434963c1ee6"} Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.387650 4810 scope.go:117] "RemoveContainer" containerID="44632dd9584a65c8447e93ce8db198c67c23e6cb0e0e4361c92967632df77581" Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.387922 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gd5zz" Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.407079 4810 scope.go:117] "RemoveContainer" containerID="7f3937cd12a9163825ad2bf7b02f451ce92a5016d602be0051815d390550e6e2" Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.424144 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gd5zz"] Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.431671 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gd5zz"] Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.441593 4810 scope.go:117] "RemoveContainer" containerID="a867b5b0fcfc7f3ba0b7f6938f28527326a38e9eee6fae8b4a7b30a8a0f62d67" Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.481500 4810 scope.go:117] "RemoveContainer" containerID="44632dd9584a65c8447e93ce8db198c67c23e6cb0e0e4361c92967632df77581" Dec 01 16:27:18 crc kubenswrapper[4810]: E1201 16:27:18.481977 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44632dd9584a65c8447e93ce8db198c67c23e6cb0e0e4361c92967632df77581\": container with ID starting with 44632dd9584a65c8447e93ce8db198c67c23e6cb0e0e4361c92967632df77581 not found: ID does not exist" containerID="44632dd9584a65c8447e93ce8db198c67c23e6cb0e0e4361c92967632df77581" Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.482036 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44632dd9584a65c8447e93ce8db198c67c23e6cb0e0e4361c92967632df77581"} err="failed to get container status \"44632dd9584a65c8447e93ce8db198c67c23e6cb0e0e4361c92967632df77581\": rpc error: code = NotFound desc = could not find container \"44632dd9584a65c8447e93ce8db198c67c23e6cb0e0e4361c92967632df77581\": container with ID starting with 44632dd9584a65c8447e93ce8db198c67c23e6cb0e0e4361c92967632df77581 not found: ID does not exist" Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.482058 4810 scope.go:117] "RemoveContainer" containerID="7f3937cd12a9163825ad2bf7b02f451ce92a5016d602be0051815d390550e6e2" Dec 01 16:27:18 crc kubenswrapper[4810]: E1201 16:27:18.482525 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f3937cd12a9163825ad2bf7b02f451ce92a5016d602be0051815d390550e6e2\": container with ID starting with 7f3937cd12a9163825ad2bf7b02f451ce92a5016d602be0051815d390550e6e2 not found: ID does not exist" containerID="7f3937cd12a9163825ad2bf7b02f451ce92a5016d602be0051815d390550e6e2" Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.482562 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f3937cd12a9163825ad2bf7b02f451ce92a5016d602be0051815d390550e6e2"} err="failed to get container status \"7f3937cd12a9163825ad2bf7b02f451ce92a5016d602be0051815d390550e6e2\": rpc error: code = NotFound desc = could not find container \"7f3937cd12a9163825ad2bf7b02f451ce92a5016d602be0051815d390550e6e2\": container with ID starting with 7f3937cd12a9163825ad2bf7b02f451ce92a5016d602be0051815d390550e6e2 not found: ID does not exist" Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.482590 4810 scope.go:117] "RemoveContainer" containerID="a867b5b0fcfc7f3ba0b7f6938f28527326a38e9eee6fae8b4a7b30a8a0f62d67" Dec 01 16:27:18 crc kubenswrapper[4810]: E1201 16:27:18.482862 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a867b5b0fcfc7f3ba0b7f6938f28527326a38e9eee6fae8b4a7b30a8a0f62d67\": container with ID starting with a867b5b0fcfc7f3ba0b7f6938f28527326a38e9eee6fae8b4a7b30a8a0f62d67 not found: ID does not exist" containerID="a867b5b0fcfc7f3ba0b7f6938f28527326a38e9eee6fae8b4a7b30a8a0f62d67" Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.482893 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a867b5b0fcfc7f3ba0b7f6938f28527326a38e9eee6fae8b4a7b30a8a0f62d67"} err="failed to get container status \"a867b5b0fcfc7f3ba0b7f6938f28527326a38e9eee6fae8b4a7b30a8a0f62d67\": rpc error: code = NotFound desc = could not find container \"a867b5b0fcfc7f3ba0b7f6938f28527326a38e9eee6fae8b4a7b30a8a0f62d67\": container with ID starting with a867b5b0fcfc7f3ba0b7f6938f28527326a38e9eee6fae8b4a7b30a8a0f62d67 not found: ID does not exist" Dec 01 16:27:18 crc kubenswrapper[4810]: I1201 16:27:18.504443 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e8daf3c-58ed-4feb-b574-8bda66dd9c79" path="/var/lib/kubelet/pods/9e8daf3c-58ed-4feb-b574-8bda66dd9c79/volumes" Dec 01 16:27:24 crc kubenswrapper[4810]: I1201 16:27:24.505715 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:27:24 crc kubenswrapper[4810]: E1201 16:27:24.506221 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:27:37 crc kubenswrapper[4810]: I1201 16:27:37.492352 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:27:37 crc kubenswrapper[4810]: E1201 16:27:37.493702 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:27:50 crc kubenswrapper[4810]: I1201 16:27:50.491816 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:27:50 crc kubenswrapper[4810]: E1201 16:27:50.492609 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:27:53 crc kubenswrapper[4810]: I1201 16:27:53.049021 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jbg9l"] Dec 01 16:27:53 crc kubenswrapper[4810]: E1201 16:27:53.050023 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e8daf3c-58ed-4feb-b574-8bda66dd9c79" containerName="extract-utilities" Dec 01 16:27:53 crc kubenswrapper[4810]: I1201 16:27:53.050048 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e8daf3c-58ed-4feb-b574-8bda66dd9c79" containerName="extract-utilities" Dec 01 16:27:53 crc kubenswrapper[4810]: E1201 16:27:53.050095 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e8daf3c-58ed-4feb-b574-8bda66dd9c79" containerName="registry-server" Dec 01 16:27:53 crc kubenswrapper[4810]: I1201 16:27:53.050107 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e8daf3c-58ed-4feb-b574-8bda66dd9c79" containerName="registry-server" Dec 01 16:27:53 crc kubenswrapper[4810]: E1201 16:27:53.050162 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e8daf3c-58ed-4feb-b574-8bda66dd9c79" containerName="extract-content" Dec 01 16:27:53 crc kubenswrapper[4810]: I1201 16:27:53.050176 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e8daf3c-58ed-4feb-b574-8bda66dd9c79" containerName="extract-content" Dec 01 16:27:53 crc kubenswrapper[4810]: I1201 16:27:53.050501 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e8daf3c-58ed-4feb-b574-8bda66dd9c79" containerName="registry-server" Dec 01 16:27:53 crc kubenswrapper[4810]: I1201 16:27:53.052634 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbg9l" Dec 01 16:27:53 crc kubenswrapper[4810]: I1201 16:27:53.066652 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jbg9l"] Dec 01 16:27:53 crc kubenswrapper[4810]: I1201 16:27:53.069398 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e5f7163-51e1-41e1-92f2-39a887f390b6-catalog-content\") pod \"redhat-operators-jbg9l\" (UID: \"6e5f7163-51e1-41e1-92f2-39a887f390b6\") " pod="openshift-marketplace/redhat-operators-jbg9l" Dec 01 16:27:53 crc kubenswrapper[4810]: I1201 16:27:53.069498 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lkcb\" (UniqueName: \"kubernetes.io/projected/6e5f7163-51e1-41e1-92f2-39a887f390b6-kube-api-access-8lkcb\") pod \"redhat-operators-jbg9l\" (UID: \"6e5f7163-51e1-41e1-92f2-39a887f390b6\") " pod="openshift-marketplace/redhat-operators-jbg9l" Dec 01 16:27:53 crc kubenswrapper[4810]: I1201 16:27:53.069631 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e5f7163-51e1-41e1-92f2-39a887f390b6-utilities\") pod \"redhat-operators-jbg9l\" (UID: \"6e5f7163-51e1-41e1-92f2-39a887f390b6\") " pod="openshift-marketplace/redhat-operators-jbg9l" Dec 01 16:27:53 crc kubenswrapper[4810]: I1201 16:27:53.171884 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e5f7163-51e1-41e1-92f2-39a887f390b6-catalog-content\") pod \"redhat-operators-jbg9l\" (UID: \"6e5f7163-51e1-41e1-92f2-39a887f390b6\") " pod="openshift-marketplace/redhat-operators-jbg9l" Dec 01 16:27:53 crc kubenswrapper[4810]: I1201 16:27:53.171949 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lkcb\" (UniqueName: \"kubernetes.io/projected/6e5f7163-51e1-41e1-92f2-39a887f390b6-kube-api-access-8lkcb\") pod \"redhat-operators-jbg9l\" (UID: \"6e5f7163-51e1-41e1-92f2-39a887f390b6\") " pod="openshift-marketplace/redhat-operators-jbg9l" Dec 01 16:27:53 crc kubenswrapper[4810]: I1201 16:27:53.171995 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e5f7163-51e1-41e1-92f2-39a887f390b6-utilities\") pod \"redhat-operators-jbg9l\" (UID: \"6e5f7163-51e1-41e1-92f2-39a887f390b6\") " pod="openshift-marketplace/redhat-operators-jbg9l" Dec 01 16:27:53 crc kubenswrapper[4810]: I1201 16:27:53.172447 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e5f7163-51e1-41e1-92f2-39a887f390b6-catalog-content\") pod \"redhat-operators-jbg9l\" (UID: \"6e5f7163-51e1-41e1-92f2-39a887f390b6\") " pod="openshift-marketplace/redhat-operators-jbg9l" Dec 01 16:27:53 crc kubenswrapper[4810]: I1201 16:27:53.172700 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e5f7163-51e1-41e1-92f2-39a887f390b6-utilities\") pod \"redhat-operators-jbg9l\" (UID: \"6e5f7163-51e1-41e1-92f2-39a887f390b6\") " pod="openshift-marketplace/redhat-operators-jbg9l" Dec 01 16:27:53 crc kubenswrapper[4810]: I1201 16:27:53.189891 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lkcb\" (UniqueName: \"kubernetes.io/projected/6e5f7163-51e1-41e1-92f2-39a887f390b6-kube-api-access-8lkcb\") pod \"redhat-operators-jbg9l\" (UID: \"6e5f7163-51e1-41e1-92f2-39a887f390b6\") " pod="openshift-marketplace/redhat-operators-jbg9l" Dec 01 16:27:53 crc kubenswrapper[4810]: I1201 16:27:53.383441 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbg9l" Dec 01 16:27:53 crc kubenswrapper[4810]: I1201 16:27:53.842841 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jbg9l"] Dec 01 16:27:54 crc kubenswrapper[4810]: I1201 16:27:54.799247 4810 generic.go:334] "Generic (PLEG): container finished" podID="6e5f7163-51e1-41e1-92f2-39a887f390b6" containerID="16a983b41169cb6b48045957b3e110e4ac69117c2ccddf1e81b89642039c55a8" exitCode=0 Dec 01 16:27:54 crc kubenswrapper[4810]: I1201 16:27:54.799334 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbg9l" event={"ID":"6e5f7163-51e1-41e1-92f2-39a887f390b6","Type":"ContainerDied","Data":"16a983b41169cb6b48045957b3e110e4ac69117c2ccddf1e81b89642039c55a8"} Dec 01 16:27:54 crc kubenswrapper[4810]: I1201 16:27:54.799627 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbg9l" event={"ID":"6e5f7163-51e1-41e1-92f2-39a887f390b6","Type":"ContainerStarted","Data":"fb45afe5289991e84d9d03dd7b0942dd8139b40ccdd43702fa4e4755cac1c1e2"} Dec 01 16:27:54 crc kubenswrapper[4810]: I1201 16:27:54.801874 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:27:55 crc kubenswrapper[4810]: I1201 16:27:55.812093 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbg9l" event={"ID":"6e5f7163-51e1-41e1-92f2-39a887f390b6","Type":"ContainerStarted","Data":"eb4fa0be257c7293f8d5b2ab77c2645e5eb9a91d86da7cde0c06e8872c50a418"} Dec 01 16:27:58 crc kubenswrapper[4810]: E1201 16:27:58.286263 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e5f7163_51e1_41e1_92f2_39a887f390b6.slice/crio-eb4fa0be257c7293f8d5b2ab77c2645e5eb9a91d86da7cde0c06e8872c50a418.scope\": RecentStats: unable to find data in memory cache]" Dec 01 16:27:59 crc kubenswrapper[4810]: I1201 16:27:59.043264 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-bcw9j"] Dec 01 16:27:59 crc kubenswrapper[4810]: I1201 16:27:59.053815 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-bcw9j"] Dec 01 16:27:59 crc kubenswrapper[4810]: I1201 16:27:59.845601 4810 generic.go:334] "Generic (PLEG): container finished" podID="6e5f7163-51e1-41e1-92f2-39a887f390b6" containerID="eb4fa0be257c7293f8d5b2ab77c2645e5eb9a91d86da7cde0c06e8872c50a418" exitCode=0 Dec 01 16:27:59 crc kubenswrapper[4810]: I1201 16:27:59.845654 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbg9l" event={"ID":"6e5f7163-51e1-41e1-92f2-39a887f390b6","Type":"ContainerDied","Data":"eb4fa0be257c7293f8d5b2ab77c2645e5eb9a91d86da7cde0c06e8872c50a418"} Dec 01 16:28:00 crc kubenswrapper[4810]: I1201 16:28:00.505246 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89486609-b664-4b57-8776-c9b421186fab" path="/var/lib/kubelet/pods/89486609-b664-4b57-8776-c9b421186fab/volumes" Dec 01 16:28:00 crc kubenswrapper[4810]: I1201 16:28:00.865551 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbg9l" event={"ID":"6e5f7163-51e1-41e1-92f2-39a887f390b6","Type":"ContainerStarted","Data":"af46ae7631afccf83f4741cdf99c2c5017f04b217439188dea553e4ae408a4f1"} Dec 01 16:28:00 crc kubenswrapper[4810]: I1201 16:28:00.890006 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jbg9l" podStartSLOduration=2.365539339 podStartE2EDuration="7.88998593s" podCreationTimestamp="2025-12-01 16:27:53 +0000 UTC" firstStartedPulling="2025-12-01 16:27:54.80160808 +0000 UTC m=+6840.565117683" lastFinishedPulling="2025-12-01 16:28:00.326054671 +0000 UTC m=+6846.089564274" observedRunningTime="2025-12-01 16:28:00.884015719 +0000 UTC m=+6846.647525362" watchObservedRunningTime="2025-12-01 16:28:00.88998593 +0000 UTC m=+6846.653495543" Dec 01 16:28:03 crc kubenswrapper[4810]: I1201 16:28:03.384271 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jbg9l" Dec 01 16:28:03 crc kubenswrapper[4810]: I1201 16:28:03.384697 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jbg9l" Dec 01 16:28:04 crc kubenswrapper[4810]: I1201 16:28:04.434232 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jbg9l" podUID="6e5f7163-51e1-41e1-92f2-39a887f390b6" containerName="registry-server" probeResult="failure" output=< Dec 01 16:28:04 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Dec 01 16:28:04 crc kubenswrapper[4810]: > Dec 01 16:28:04 crc kubenswrapper[4810]: I1201 16:28:04.498322 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:28:04 crc kubenswrapper[4810]: E1201 16:28:04.498686 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:28:13 crc kubenswrapper[4810]: I1201 16:28:13.434693 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jbg9l" Dec 01 16:28:13 crc kubenswrapper[4810]: I1201 16:28:13.489848 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jbg9l" Dec 01 16:28:13 crc kubenswrapper[4810]: I1201 16:28:13.675966 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jbg9l"] Dec 01 16:28:14 crc kubenswrapper[4810]: I1201 16:28:14.997748 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jbg9l" podUID="6e5f7163-51e1-41e1-92f2-39a887f390b6" containerName="registry-server" containerID="cri-o://af46ae7631afccf83f4741cdf99c2c5017f04b217439188dea553e4ae408a4f1" gracePeriod=2 Dec 01 16:28:15 crc kubenswrapper[4810]: I1201 16:28:15.453154 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbg9l" Dec 01 16:28:15 crc kubenswrapper[4810]: I1201 16:28:15.477523 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lkcb\" (UniqueName: \"kubernetes.io/projected/6e5f7163-51e1-41e1-92f2-39a887f390b6-kube-api-access-8lkcb\") pod \"6e5f7163-51e1-41e1-92f2-39a887f390b6\" (UID: \"6e5f7163-51e1-41e1-92f2-39a887f390b6\") " Dec 01 16:28:15 crc kubenswrapper[4810]: I1201 16:28:15.477640 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e5f7163-51e1-41e1-92f2-39a887f390b6-utilities\") pod \"6e5f7163-51e1-41e1-92f2-39a887f390b6\" (UID: \"6e5f7163-51e1-41e1-92f2-39a887f390b6\") " Dec 01 16:28:15 crc kubenswrapper[4810]: I1201 16:28:15.477742 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e5f7163-51e1-41e1-92f2-39a887f390b6-catalog-content\") pod \"6e5f7163-51e1-41e1-92f2-39a887f390b6\" (UID: \"6e5f7163-51e1-41e1-92f2-39a887f390b6\") " Dec 01 16:28:15 crc kubenswrapper[4810]: I1201 16:28:15.478556 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e5f7163-51e1-41e1-92f2-39a887f390b6-utilities" (OuterVolumeSpecName: "utilities") pod "6e5f7163-51e1-41e1-92f2-39a887f390b6" (UID: "6e5f7163-51e1-41e1-92f2-39a887f390b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:28:15 crc kubenswrapper[4810]: I1201 16:28:15.492748 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e5f7163-51e1-41e1-92f2-39a887f390b6-kube-api-access-8lkcb" (OuterVolumeSpecName: "kube-api-access-8lkcb") pod "6e5f7163-51e1-41e1-92f2-39a887f390b6" (UID: "6e5f7163-51e1-41e1-92f2-39a887f390b6"). InnerVolumeSpecName "kube-api-access-8lkcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:28:15 crc kubenswrapper[4810]: I1201 16:28:15.580685 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e5f7163-51e1-41e1-92f2-39a887f390b6-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:28:15 crc kubenswrapper[4810]: I1201 16:28:15.580718 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lkcb\" (UniqueName: \"kubernetes.io/projected/6e5f7163-51e1-41e1-92f2-39a887f390b6-kube-api-access-8lkcb\") on node \"crc\" DevicePath \"\"" Dec 01 16:28:15 crc kubenswrapper[4810]: I1201 16:28:15.582289 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e5f7163-51e1-41e1-92f2-39a887f390b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e5f7163-51e1-41e1-92f2-39a887f390b6" (UID: "6e5f7163-51e1-41e1-92f2-39a887f390b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:28:15 crc kubenswrapper[4810]: I1201 16:28:15.682885 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e5f7163-51e1-41e1-92f2-39a887f390b6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:28:15 crc kubenswrapper[4810]: I1201 16:28:15.891535 4810 scope.go:117] "RemoveContainer" containerID="5780a34684dc0843aa275ef4948d4f619741a71249fff8e5310309fa0b40399f" Dec 01 16:28:15 crc kubenswrapper[4810]: I1201 16:28:15.935753 4810 scope.go:117] "RemoveContainer" containerID="e6ceddeec41cb81433b0b0fd627186faef0eba71d3692c4eaad6035e83b4bba2" Dec 01 16:28:16 crc kubenswrapper[4810]: I1201 16:28:16.024406 4810 generic.go:334] "Generic (PLEG): container finished" podID="6e5f7163-51e1-41e1-92f2-39a887f390b6" containerID="af46ae7631afccf83f4741cdf99c2c5017f04b217439188dea553e4ae408a4f1" exitCode=0 Dec 01 16:28:16 crc kubenswrapper[4810]: I1201 16:28:16.024628 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbg9l" Dec 01 16:28:16 crc kubenswrapper[4810]: I1201 16:28:16.024653 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbg9l" event={"ID":"6e5f7163-51e1-41e1-92f2-39a887f390b6","Type":"ContainerDied","Data":"af46ae7631afccf83f4741cdf99c2c5017f04b217439188dea553e4ae408a4f1"} Dec 01 16:28:16 crc kubenswrapper[4810]: I1201 16:28:16.035854 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbg9l" event={"ID":"6e5f7163-51e1-41e1-92f2-39a887f390b6","Type":"ContainerDied","Data":"fb45afe5289991e84d9d03dd7b0942dd8139b40ccdd43702fa4e4755cac1c1e2"} Dec 01 16:28:16 crc kubenswrapper[4810]: I1201 16:28:16.035913 4810 scope.go:117] "RemoveContainer" containerID="af46ae7631afccf83f4741cdf99c2c5017f04b217439188dea553e4ae408a4f1" Dec 01 16:28:16 crc kubenswrapper[4810]: I1201 16:28:16.071340 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jbg9l"] Dec 01 16:28:16 crc kubenswrapper[4810]: I1201 16:28:16.082763 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jbg9l"] Dec 01 16:28:16 crc kubenswrapper[4810]: I1201 16:28:16.089033 4810 scope.go:117] "RemoveContainer" containerID="eb4fa0be257c7293f8d5b2ab77c2645e5eb9a91d86da7cde0c06e8872c50a418" Dec 01 16:28:16 crc kubenswrapper[4810]: I1201 16:28:16.127690 4810 scope.go:117] "RemoveContainer" containerID="16a983b41169cb6b48045957b3e110e4ac69117c2ccddf1e81b89642039c55a8" Dec 01 16:28:16 crc kubenswrapper[4810]: I1201 16:28:16.210025 4810 scope.go:117] "RemoveContainer" containerID="af46ae7631afccf83f4741cdf99c2c5017f04b217439188dea553e4ae408a4f1" Dec 01 16:28:16 crc kubenswrapper[4810]: E1201 16:28:16.210480 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af46ae7631afccf83f4741cdf99c2c5017f04b217439188dea553e4ae408a4f1\": container with ID starting with af46ae7631afccf83f4741cdf99c2c5017f04b217439188dea553e4ae408a4f1 not found: ID does not exist" containerID="af46ae7631afccf83f4741cdf99c2c5017f04b217439188dea553e4ae408a4f1" Dec 01 16:28:16 crc kubenswrapper[4810]: I1201 16:28:16.210511 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af46ae7631afccf83f4741cdf99c2c5017f04b217439188dea553e4ae408a4f1"} err="failed to get container status \"af46ae7631afccf83f4741cdf99c2c5017f04b217439188dea553e4ae408a4f1\": rpc error: code = NotFound desc = could not find container \"af46ae7631afccf83f4741cdf99c2c5017f04b217439188dea553e4ae408a4f1\": container with ID starting with af46ae7631afccf83f4741cdf99c2c5017f04b217439188dea553e4ae408a4f1 not found: ID does not exist" Dec 01 16:28:16 crc kubenswrapper[4810]: I1201 16:28:16.210533 4810 scope.go:117] "RemoveContainer" containerID="eb4fa0be257c7293f8d5b2ab77c2645e5eb9a91d86da7cde0c06e8872c50a418" Dec 01 16:28:16 crc kubenswrapper[4810]: E1201 16:28:16.211016 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb4fa0be257c7293f8d5b2ab77c2645e5eb9a91d86da7cde0c06e8872c50a418\": container with ID starting with eb4fa0be257c7293f8d5b2ab77c2645e5eb9a91d86da7cde0c06e8872c50a418 not found: ID does not exist" containerID="eb4fa0be257c7293f8d5b2ab77c2645e5eb9a91d86da7cde0c06e8872c50a418" Dec 01 16:28:16 crc kubenswrapper[4810]: I1201 16:28:16.211074 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb4fa0be257c7293f8d5b2ab77c2645e5eb9a91d86da7cde0c06e8872c50a418"} err="failed to get container status \"eb4fa0be257c7293f8d5b2ab77c2645e5eb9a91d86da7cde0c06e8872c50a418\": rpc error: code = NotFound desc = could not find container \"eb4fa0be257c7293f8d5b2ab77c2645e5eb9a91d86da7cde0c06e8872c50a418\": container with ID starting with eb4fa0be257c7293f8d5b2ab77c2645e5eb9a91d86da7cde0c06e8872c50a418 not found: ID does not exist" Dec 01 16:28:16 crc kubenswrapper[4810]: I1201 16:28:16.211121 4810 scope.go:117] "RemoveContainer" containerID="16a983b41169cb6b48045957b3e110e4ac69117c2ccddf1e81b89642039c55a8" Dec 01 16:28:16 crc kubenswrapper[4810]: E1201 16:28:16.211412 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16a983b41169cb6b48045957b3e110e4ac69117c2ccddf1e81b89642039c55a8\": container with ID starting with 16a983b41169cb6b48045957b3e110e4ac69117c2ccddf1e81b89642039c55a8 not found: ID does not exist" containerID="16a983b41169cb6b48045957b3e110e4ac69117c2ccddf1e81b89642039c55a8" Dec 01 16:28:16 crc kubenswrapper[4810]: I1201 16:28:16.211491 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16a983b41169cb6b48045957b3e110e4ac69117c2ccddf1e81b89642039c55a8"} err="failed to get container status \"16a983b41169cb6b48045957b3e110e4ac69117c2ccddf1e81b89642039c55a8\": rpc error: code = NotFound desc = could not find container \"16a983b41169cb6b48045957b3e110e4ac69117c2ccddf1e81b89642039c55a8\": container with ID starting with 16a983b41169cb6b48045957b3e110e4ac69117c2ccddf1e81b89642039c55a8 not found: ID does not exist" Dec 01 16:28:16 crc kubenswrapper[4810]: I1201 16:28:16.504786 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e5f7163-51e1-41e1-92f2-39a887f390b6" path="/var/lib/kubelet/pods/6e5f7163-51e1-41e1-92f2-39a887f390b6/volumes" Dec 01 16:28:19 crc kubenswrapper[4810]: I1201 16:28:19.502208 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:28:19 crc kubenswrapper[4810]: E1201 16:28:19.503451 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:28:31 crc kubenswrapper[4810]: I1201 16:28:31.491221 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:28:31 crc kubenswrapper[4810]: E1201 16:28:31.491975 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:28:43 crc kubenswrapper[4810]: I1201 16:28:43.492198 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:28:43 crc kubenswrapper[4810]: E1201 16:28:43.493306 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:28:55 crc kubenswrapper[4810]: I1201 16:28:55.491804 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:28:55 crc kubenswrapper[4810]: E1201 16:28:55.492587 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:29:06 crc kubenswrapper[4810]: I1201 16:29:06.491562 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:29:06 crc kubenswrapper[4810]: E1201 16:29:06.492252 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:29:16 crc kubenswrapper[4810]: I1201 16:29:16.117307 4810 scope.go:117] "RemoveContainer" containerID="2474303523ecd36c3c4c15fd281847fc3259faaaf1bcd7a17867ffa92b23bb70" Dec 01 16:29:21 crc kubenswrapper[4810]: I1201 16:29:21.490707 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:29:21 crc kubenswrapper[4810]: E1201 16:29:21.492581 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:29:34 crc kubenswrapper[4810]: I1201 16:29:34.499819 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:29:34 crc kubenswrapper[4810]: E1201 16:29:34.500557 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:29:45 crc kubenswrapper[4810]: I1201 16:29:45.492754 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:29:45 crc kubenswrapper[4810]: E1201 16:29:45.493535 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:29:57 crc kubenswrapper[4810]: I1201 16:29:57.492282 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:29:57 crc kubenswrapper[4810]: E1201 16:29:57.493206 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.160642 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw"] Dec 01 16:30:00 crc kubenswrapper[4810]: E1201 16:30:00.161716 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e5f7163-51e1-41e1-92f2-39a887f390b6" containerName="extract-utilities" Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.161736 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e5f7163-51e1-41e1-92f2-39a887f390b6" containerName="extract-utilities" Dec 01 16:30:00 crc kubenswrapper[4810]: E1201 16:30:00.161766 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e5f7163-51e1-41e1-92f2-39a887f390b6" containerName="registry-server" Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.161774 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e5f7163-51e1-41e1-92f2-39a887f390b6" containerName="registry-server" Dec 01 16:30:00 crc kubenswrapper[4810]: E1201 16:30:00.161788 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e5f7163-51e1-41e1-92f2-39a887f390b6" containerName="extract-content" Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.161796 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e5f7163-51e1-41e1-92f2-39a887f390b6" containerName="extract-content" Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.162059 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e5f7163-51e1-41e1-92f2-39a887f390b6" containerName="registry-server" Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.163178 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw" Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.165280 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.165334 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.172046 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw"] Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.294897 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbpl4\" (UniqueName: \"kubernetes.io/projected/5af27677-6662-4102-882b-38d36c928031-kube-api-access-qbpl4\") pod \"collect-profiles-29410110-47pfw\" (UID: \"5af27677-6662-4102-882b-38d36c928031\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw" Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.294983 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5af27677-6662-4102-882b-38d36c928031-secret-volume\") pod \"collect-profiles-29410110-47pfw\" (UID: \"5af27677-6662-4102-882b-38d36c928031\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw" Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.295049 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5af27677-6662-4102-882b-38d36c928031-config-volume\") pod \"collect-profiles-29410110-47pfw\" (UID: \"5af27677-6662-4102-882b-38d36c928031\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw" Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.396807 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbpl4\" (UniqueName: \"kubernetes.io/projected/5af27677-6662-4102-882b-38d36c928031-kube-api-access-qbpl4\") pod \"collect-profiles-29410110-47pfw\" (UID: \"5af27677-6662-4102-882b-38d36c928031\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw" Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.396875 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5af27677-6662-4102-882b-38d36c928031-secret-volume\") pod \"collect-profiles-29410110-47pfw\" (UID: \"5af27677-6662-4102-882b-38d36c928031\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw" Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.396944 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5af27677-6662-4102-882b-38d36c928031-config-volume\") pod \"collect-profiles-29410110-47pfw\" (UID: \"5af27677-6662-4102-882b-38d36c928031\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw" Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.397877 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5af27677-6662-4102-882b-38d36c928031-config-volume\") pod \"collect-profiles-29410110-47pfw\" (UID: \"5af27677-6662-4102-882b-38d36c928031\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw" Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.403313 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5af27677-6662-4102-882b-38d36c928031-secret-volume\") pod \"collect-profiles-29410110-47pfw\" (UID: \"5af27677-6662-4102-882b-38d36c928031\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw" Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.417284 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbpl4\" (UniqueName: \"kubernetes.io/projected/5af27677-6662-4102-882b-38d36c928031-kube-api-access-qbpl4\") pod \"collect-profiles-29410110-47pfw\" (UID: \"5af27677-6662-4102-882b-38d36c928031\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw" Dec 01 16:30:00 crc kubenswrapper[4810]: I1201 16:30:00.493823 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw" Dec 01 16:30:01 crc kubenswrapper[4810]: I1201 16:30:01.014856 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw"] Dec 01 16:30:01 crc kubenswrapper[4810]: I1201 16:30:01.103919 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw" event={"ID":"5af27677-6662-4102-882b-38d36c928031","Type":"ContainerStarted","Data":"5a83bb24afb9268977367f876b8d592565aad6ea9d9d4ccfb8ac9be85174a0e5"} Dec 01 16:30:02 crc kubenswrapper[4810]: I1201 16:30:02.113491 4810 generic.go:334] "Generic (PLEG): container finished" podID="5af27677-6662-4102-882b-38d36c928031" containerID="9c2b483d543b590039596544bbe7e243cff494dd2fd6f4d6d63290643ab73bb0" exitCode=0 Dec 01 16:30:02 crc kubenswrapper[4810]: I1201 16:30:02.113556 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw" event={"ID":"5af27677-6662-4102-882b-38d36c928031","Type":"ContainerDied","Data":"9c2b483d543b590039596544bbe7e243cff494dd2fd6f4d6d63290643ab73bb0"} Dec 01 16:30:03 crc kubenswrapper[4810]: I1201 16:30:03.477925 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw" Dec 01 16:30:03 crc kubenswrapper[4810]: I1201 16:30:03.585441 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbpl4\" (UniqueName: \"kubernetes.io/projected/5af27677-6662-4102-882b-38d36c928031-kube-api-access-qbpl4\") pod \"5af27677-6662-4102-882b-38d36c928031\" (UID: \"5af27677-6662-4102-882b-38d36c928031\") " Dec 01 16:30:03 crc kubenswrapper[4810]: I1201 16:30:03.585757 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5af27677-6662-4102-882b-38d36c928031-secret-volume\") pod \"5af27677-6662-4102-882b-38d36c928031\" (UID: \"5af27677-6662-4102-882b-38d36c928031\") " Dec 01 16:30:03 crc kubenswrapper[4810]: I1201 16:30:03.585844 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5af27677-6662-4102-882b-38d36c928031-config-volume\") pod \"5af27677-6662-4102-882b-38d36c928031\" (UID: \"5af27677-6662-4102-882b-38d36c928031\") " Dec 01 16:30:03 crc kubenswrapper[4810]: I1201 16:30:03.586610 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5af27677-6662-4102-882b-38d36c928031-config-volume" (OuterVolumeSpecName: "config-volume") pod "5af27677-6662-4102-882b-38d36c928031" (UID: "5af27677-6662-4102-882b-38d36c928031"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:30:03 crc kubenswrapper[4810]: I1201 16:30:03.598180 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5af27677-6662-4102-882b-38d36c928031-kube-api-access-qbpl4" (OuterVolumeSpecName: "kube-api-access-qbpl4") pod "5af27677-6662-4102-882b-38d36c928031" (UID: "5af27677-6662-4102-882b-38d36c928031"). InnerVolumeSpecName "kube-api-access-qbpl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:30:03 crc kubenswrapper[4810]: I1201 16:30:03.599504 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af27677-6662-4102-882b-38d36c928031-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5af27677-6662-4102-882b-38d36c928031" (UID: "5af27677-6662-4102-882b-38d36c928031"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:30:03 crc kubenswrapper[4810]: I1201 16:30:03.688325 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5af27677-6662-4102-882b-38d36c928031-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 16:30:03 crc kubenswrapper[4810]: I1201 16:30:03.688371 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5af27677-6662-4102-882b-38d36c928031-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 16:30:03 crc kubenswrapper[4810]: I1201 16:30:03.688383 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbpl4\" (UniqueName: \"kubernetes.io/projected/5af27677-6662-4102-882b-38d36c928031-kube-api-access-qbpl4\") on node \"crc\" DevicePath \"\"" Dec 01 16:30:04 crc kubenswrapper[4810]: I1201 16:30:04.135168 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw" event={"ID":"5af27677-6662-4102-882b-38d36c928031","Type":"ContainerDied","Data":"5a83bb24afb9268977367f876b8d592565aad6ea9d9d4ccfb8ac9be85174a0e5"} Dec 01 16:30:04 crc kubenswrapper[4810]: I1201 16:30:04.135209 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a83bb24afb9268977367f876b8d592565aad6ea9d9d4ccfb8ac9be85174a0e5" Dec 01 16:30:04 crc kubenswrapper[4810]: I1201 16:30:04.135263 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw" Dec 01 16:30:04 crc kubenswrapper[4810]: I1201 16:30:04.555349 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv"] Dec 01 16:30:04 crc kubenswrapper[4810]: I1201 16:30:04.564956 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410065-knjcv"] Dec 01 16:30:06 crc kubenswrapper[4810]: I1201 16:30:06.505402 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7a3c1b8-3d1c-4884-b887-c13edcaacd53" path="/var/lib/kubelet/pods/a7a3c1b8-3d1c-4884-b887-c13edcaacd53/volumes" Dec 01 16:30:09 crc kubenswrapper[4810]: I1201 16:30:09.491709 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:30:09 crc kubenswrapper[4810]: E1201 16:30:09.492439 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:30:16 crc kubenswrapper[4810]: I1201 16:30:16.204079 4810 scope.go:117] "RemoveContainer" containerID="d7e73dcfe83a0bbd79d836bdea642db498fc3588517c95a2e7be6b62f3b1bd91" Dec 01 16:30:24 crc kubenswrapper[4810]: I1201 16:30:24.500316 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:30:24 crc kubenswrapper[4810]: E1201 16:30:24.500926 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:30:37 crc kubenswrapper[4810]: I1201 16:30:37.491275 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:30:37 crc kubenswrapper[4810]: E1201 16:30:37.492029 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:30:48 crc kubenswrapper[4810]: I1201 16:30:48.492081 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:30:48 crc kubenswrapper[4810]: E1201 16:30:48.493188 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:31:03 crc kubenswrapper[4810]: I1201 16:31:03.492048 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:31:03 crc kubenswrapper[4810]: I1201 16:31:03.830556 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"71549b0dadf7e5851cf8c3dca9e008d0d9bb20d1aa5c39eae94356aa087f1657"} Dec 01 16:31:22 crc kubenswrapper[4810]: I1201 16:31:22.049682 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-4f6vj"] Dec 01 16:31:22 crc kubenswrapper[4810]: I1201 16:31:22.060613 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-4dbf-account-create-update-l4q8c"] Dec 01 16:31:22 crc kubenswrapper[4810]: I1201 16:31:22.071196 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-4f6vj"] Dec 01 16:31:22 crc kubenswrapper[4810]: I1201 16:31:22.078798 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-4dbf-account-create-update-l4q8c"] Dec 01 16:31:22 crc kubenswrapper[4810]: I1201 16:31:22.510825 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f36f4f1-7565-45e8-b96d-5b52449a4bb4" path="/var/lib/kubelet/pods/6f36f4f1-7565-45e8-b96d-5b52449a4bb4/volumes" Dec 01 16:31:22 crc kubenswrapper[4810]: I1201 16:31:22.512390 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4ded436-383b-493b-a03b-c65f14063d69" path="/var/lib/kubelet/pods/d4ded436-383b-493b-a03b-c65f14063d69/volumes" Dec 01 16:31:35 crc kubenswrapper[4810]: I1201 16:31:35.051760 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-z84cx"] Dec 01 16:31:35 crc kubenswrapper[4810]: I1201 16:31:35.063533 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-z84cx"] Dec 01 16:31:36 crc kubenswrapper[4810]: I1201 16:31:36.503287 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd707047-2f0e-4029-846b-9b5a4de8cad7" path="/var/lib/kubelet/pods/bd707047-2f0e-4029-846b-9b5a4de8cad7/volumes" Dec 01 16:32:16 crc kubenswrapper[4810]: I1201 16:32:16.387454 4810 scope.go:117] "RemoveContainer" containerID="83803efd30b35fd5969eb155df297e46804d7964eb683c52cea0d0fa6962f5c1" Dec 01 16:32:16 crc kubenswrapper[4810]: I1201 16:32:16.432683 4810 scope.go:117] "RemoveContainer" containerID="1ff4a064bacfef6ce9cf508b9b97be081917f14446589ff03b147b9176491865" Dec 01 16:32:16 crc kubenswrapper[4810]: I1201 16:32:16.505904 4810 scope.go:117] "RemoveContainer" containerID="5a341e04cd35522af14e63af6a7a386eba582fd84fe4bf09171e73d6c627bd0e" Dec 01 16:33:32 crc kubenswrapper[4810]: I1201 16:33:32.971971 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:33:32 crc kubenswrapper[4810]: I1201 16:33:32.972705 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:34:02 crc kubenswrapper[4810]: I1201 16:34:02.972543 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:34:02 crc kubenswrapper[4810]: I1201 16:34:02.973327 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:34:08 crc kubenswrapper[4810]: I1201 16:34:08.063866 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-2833-account-create-update-x4f4n"] Dec 01 16:34:08 crc kubenswrapper[4810]: I1201 16:34:08.073931 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-2833-account-create-update-x4f4n"] Dec 01 16:34:08 crc kubenswrapper[4810]: I1201 16:34:08.083314 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-vgp88"] Dec 01 16:34:08 crc kubenswrapper[4810]: I1201 16:34:08.092665 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-vgp88"] Dec 01 16:34:08 crc kubenswrapper[4810]: I1201 16:34:08.508963 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0b4462a-f09b-4d97-9069-c8b248855311" path="/var/lib/kubelet/pods/b0b4462a-f09b-4d97-9069-c8b248855311/volumes" Dec 01 16:34:08 crc kubenswrapper[4810]: I1201 16:34:08.510812 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e" path="/var/lib/kubelet/pods/bbcd2b18-92ab-425c-bcc9-00c8eacd7c6e/volumes" Dec 01 16:34:16 crc kubenswrapper[4810]: I1201 16:34:16.654676 4810 scope.go:117] "RemoveContainer" containerID="09a9240e4da76772e26368cb6bc54742aa9b3279565331075191dc6d9408af3b" Dec 01 16:34:16 crc kubenswrapper[4810]: I1201 16:34:16.690012 4810 scope.go:117] "RemoveContainer" containerID="e53c4f4ffada9a1df56e799ff33e717732604ff8999d584f19a7aeff4729156e" Dec 01 16:34:19 crc kubenswrapper[4810]: I1201 16:34:19.059834 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-t8xft"] Dec 01 16:34:19 crc kubenswrapper[4810]: I1201 16:34:19.068309 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-t8xft"] Dec 01 16:34:20 crc kubenswrapper[4810]: I1201 16:34:20.505634 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85bae6bb-65c7-4004-99a3-d41d528f1be3" path="/var/lib/kubelet/pods/85bae6bb-65c7-4004-99a3-d41d528f1be3/volumes" Dec 01 16:34:32 crc kubenswrapper[4810]: I1201 16:34:32.972656 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:34:32 crc kubenswrapper[4810]: I1201 16:34:32.973234 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:34:32 crc kubenswrapper[4810]: I1201 16:34:32.973287 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 16:34:32 crc kubenswrapper[4810]: I1201 16:34:32.974150 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"71549b0dadf7e5851cf8c3dca9e008d0d9bb20d1aa5c39eae94356aa087f1657"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:34:32 crc kubenswrapper[4810]: I1201 16:34:32.974211 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://71549b0dadf7e5851cf8c3dca9e008d0d9bb20d1aa5c39eae94356aa087f1657" gracePeriod=600 Dec 01 16:34:33 crc kubenswrapper[4810]: I1201 16:34:33.147795 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="71549b0dadf7e5851cf8c3dca9e008d0d9bb20d1aa5c39eae94356aa087f1657" exitCode=0 Dec 01 16:34:33 crc kubenswrapper[4810]: I1201 16:34:33.147873 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"71549b0dadf7e5851cf8c3dca9e008d0d9bb20d1aa5c39eae94356aa087f1657"} Dec 01 16:34:33 crc kubenswrapper[4810]: I1201 16:34:33.147939 4810 scope.go:117] "RemoveContainer" containerID="6811fd0b48443172abc844773d0194086ad7dce77253258a3b520f87f582324b" Dec 01 16:34:34 crc kubenswrapper[4810]: I1201 16:34:34.160192 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b"} Dec 01 16:35:16 crc kubenswrapper[4810]: I1201 16:35:16.794051 4810 scope.go:117] "RemoveContainer" containerID="04a2372a0b8b8c0e8f34fafcd609ae79e2b5c96f79520cbcc7412d88f535df6f" Dec 01 16:35:48 crc kubenswrapper[4810]: I1201 16:35:48.071839 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mn97w"] Dec 01 16:35:48 crc kubenswrapper[4810]: E1201 16:35:48.072765 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af27677-6662-4102-882b-38d36c928031" containerName="collect-profiles" Dec 01 16:35:48 crc kubenswrapper[4810]: I1201 16:35:48.072776 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af27677-6662-4102-882b-38d36c928031" containerName="collect-profiles" Dec 01 16:35:48 crc kubenswrapper[4810]: I1201 16:35:48.072997 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af27677-6662-4102-882b-38d36c928031" containerName="collect-profiles" Dec 01 16:35:48 crc kubenswrapper[4810]: I1201 16:35:48.074560 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mn97w" Dec 01 16:35:48 crc kubenswrapper[4810]: I1201 16:35:48.082295 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba60559e-fda1-4366-b7ee-eabb2b6c2324-catalog-content\") pod \"certified-operators-mn97w\" (UID: \"ba60559e-fda1-4366-b7ee-eabb2b6c2324\") " pod="openshift-marketplace/certified-operators-mn97w" Dec 01 16:35:48 crc kubenswrapper[4810]: I1201 16:35:48.082363 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba60559e-fda1-4366-b7ee-eabb2b6c2324-utilities\") pod \"certified-operators-mn97w\" (UID: \"ba60559e-fda1-4366-b7ee-eabb2b6c2324\") " pod="openshift-marketplace/certified-operators-mn97w" Dec 01 16:35:48 crc kubenswrapper[4810]: I1201 16:35:48.082432 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5sqd\" (UniqueName: \"kubernetes.io/projected/ba60559e-fda1-4366-b7ee-eabb2b6c2324-kube-api-access-t5sqd\") pod \"certified-operators-mn97w\" (UID: \"ba60559e-fda1-4366-b7ee-eabb2b6c2324\") " pod="openshift-marketplace/certified-operators-mn97w" Dec 01 16:35:48 crc kubenswrapper[4810]: I1201 16:35:48.099294 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mn97w"] Dec 01 16:35:48 crc kubenswrapper[4810]: I1201 16:35:48.183566 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5sqd\" (UniqueName: \"kubernetes.io/projected/ba60559e-fda1-4366-b7ee-eabb2b6c2324-kube-api-access-t5sqd\") pod \"certified-operators-mn97w\" (UID: \"ba60559e-fda1-4366-b7ee-eabb2b6c2324\") " pod="openshift-marketplace/certified-operators-mn97w" Dec 01 16:35:48 crc kubenswrapper[4810]: I1201 16:35:48.183743 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba60559e-fda1-4366-b7ee-eabb2b6c2324-catalog-content\") pod \"certified-operators-mn97w\" (UID: \"ba60559e-fda1-4366-b7ee-eabb2b6c2324\") " pod="openshift-marketplace/certified-operators-mn97w" Dec 01 16:35:48 crc kubenswrapper[4810]: I1201 16:35:48.183805 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba60559e-fda1-4366-b7ee-eabb2b6c2324-utilities\") pod \"certified-operators-mn97w\" (UID: \"ba60559e-fda1-4366-b7ee-eabb2b6c2324\") " pod="openshift-marketplace/certified-operators-mn97w" Dec 01 16:35:48 crc kubenswrapper[4810]: I1201 16:35:48.184343 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba60559e-fda1-4366-b7ee-eabb2b6c2324-catalog-content\") pod \"certified-operators-mn97w\" (UID: \"ba60559e-fda1-4366-b7ee-eabb2b6c2324\") " pod="openshift-marketplace/certified-operators-mn97w" Dec 01 16:35:48 crc kubenswrapper[4810]: I1201 16:35:48.184657 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba60559e-fda1-4366-b7ee-eabb2b6c2324-utilities\") pod \"certified-operators-mn97w\" (UID: \"ba60559e-fda1-4366-b7ee-eabb2b6c2324\") " pod="openshift-marketplace/certified-operators-mn97w" Dec 01 16:35:48 crc kubenswrapper[4810]: I1201 16:35:48.202730 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5sqd\" (UniqueName: \"kubernetes.io/projected/ba60559e-fda1-4366-b7ee-eabb2b6c2324-kube-api-access-t5sqd\") pod \"certified-operators-mn97w\" (UID: \"ba60559e-fda1-4366-b7ee-eabb2b6c2324\") " pod="openshift-marketplace/certified-operators-mn97w" Dec 01 16:35:48 crc kubenswrapper[4810]: I1201 16:35:48.406528 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mn97w" Dec 01 16:35:48 crc kubenswrapper[4810]: I1201 16:35:48.910867 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mn97w"] Dec 01 16:35:49 crc kubenswrapper[4810]: I1201 16:35:49.874579 4810 generic.go:334] "Generic (PLEG): container finished" podID="ba60559e-fda1-4366-b7ee-eabb2b6c2324" containerID="6577b1cd20524512143ca60d502784ca9adffd31cdfd2c76ea965dfe7a81bcdc" exitCode=0 Dec 01 16:35:49 crc kubenswrapper[4810]: I1201 16:35:49.874652 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mn97w" event={"ID":"ba60559e-fda1-4366-b7ee-eabb2b6c2324","Type":"ContainerDied","Data":"6577b1cd20524512143ca60d502784ca9adffd31cdfd2c76ea965dfe7a81bcdc"} Dec 01 16:35:49 crc kubenswrapper[4810]: I1201 16:35:49.874910 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mn97w" event={"ID":"ba60559e-fda1-4366-b7ee-eabb2b6c2324","Type":"ContainerStarted","Data":"782b1ff4c30356964d649e7b4516b431aaed6ab02b8ffb47c4cc0261eed6c7f5"} Dec 01 16:35:49 crc kubenswrapper[4810]: I1201 16:35:49.877891 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:35:55 crc kubenswrapper[4810]: I1201 16:35:55.942713 4810 generic.go:334] "Generic (PLEG): container finished" podID="ba60559e-fda1-4366-b7ee-eabb2b6c2324" containerID="27d23d562153b9fb4a81fbf91230e791e9aaba2c2dad5fbf8d649e01f9ddfd8d" exitCode=0 Dec 01 16:35:55 crc kubenswrapper[4810]: I1201 16:35:55.944877 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mn97w" event={"ID":"ba60559e-fda1-4366-b7ee-eabb2b6c2324","Type":"ContainerDied","Data":"27d23d562153b9fb4a81fbf91230e791e9aaba2c2dad5fbf8d649e01f9ddfd8d"} Dec 01 16:35:56 crc kubenswrapper[4810]: I1201 16:35:56.953960 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mn97w" event={"ID":"ba60559e-fda1-4366-b7ee-eabb2b6c2324","Type":"ContainerStarted","Data":"b0a1534c51edfdc3a3a5acdbad91e66ea5b8396a37b2e770cedb71ae2328892d"} Dec 01 16:35:56 crc kubenswrapper[4810]: I1201 16:35:56.978622 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mn97w" podStartSLOduration=2.429834908 podStartE2EDuration="8.978595957s" podCreationTimestamp="2025-12-01 16:35:48 +0000 UTC" firstStartedPulling="2025-12-01 16:35:49.877655032 +0000 UTC m=+7315.641164635" lastFinishedPulling="2025-12-01 16:35:56.426416071 +0000 UTC m=+7322.189925684" observedRunningTime="2025-12-01 16:35:56.974979699 +0000 UTC m=+7322.738489302" watchObservedRunningTime="2025-12-01 16:35:56.978595957 +0000 UTC m=+7322.742105560" Dec 01 16:35:58 crc kubenswrapper[4810]: I1201 16:35:58.407197 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mn97w" Dec 01 16:35:58 crc kubenswrapper[4810]: I1201 16:35:58.407554 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mn97w" Dec 01 16:35:58 crc kubenswrapper[4810]: I1201 16:35:58.453226 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mn97w" Dec 01 16:36:08 crc kubenswrapper[4810]: I1201 16:36:08.460251 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mn97w" Dec 01 16:36:08 crc kubenswrapper[4810]: I1201 16:36:08.534258 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mn97w"] Dec 01 16:36:08 crc kubenswrapper[4810]: I1201 16:36:08.583957 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lxcfp"] Dec 01 16:36:08 crc kubenswrapper[4810]: I1201 16:36:08.584223 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lxcfp" podUID="61fdf03d-257d-4d46-93c4-8b772371cf61" containerName="registry-server" containerID="cri-o://16e1a603cc9aa502a52d917fae64c2a401a08e03eea47afdcbc47a89b6271eba" gracePeriod=2 Dec 01 16:36:09 crc kubenswrapper[4810]: I1201 16:36:09.083049 4810 generic.go:334] "Generic (PLEG): container finished" podID="61fdf03d-257d-4d46-93c4-8b772371cf61" containerID="16e1a603cc9aa502a52d917fae64c2a401a08e03eea47afdcbc47a89b6271eba" exitCode=0 Dec 01 16:36:09 crc kubenswrapper[4810]: I1201 16:36:09.083135 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxcfp" event={"ID":"61fdf03d-257d-4d46-93c4-8b772371cf61","Type":"ContainerDied","Data":"16e1a603cc9aa502a52d917fae64c2a401a08e03eea47afdcbc47a89b6271eba"} Dec 01 16:36:09 crc kubenswrapper[4810]: I1201 16:36:09.083644 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxcfp" event={"ID":"61fdf03d-257d-4d46-93c4-8b772371cf61","Type":"ContainerDied","Data":"7171653cf638f9ab8a14f4bbe3cbb95168145bb484eea8eefc0c2d662ac7fecf"} Dec 01 16:36:09 crc kubenswrapper[4810]: I1201 16:36:09.083670 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7171653cf638f9ab8a14f4bbe3cbb95168145bb484eea8eefc0c2d662ac7fecf" Dec 01 16:36:09 crc kubenswrapper[4810]: I1201 16:36:09.155398 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lxcfp" Dec 01 16:36:09 crc kubenswrapper[4810]: I1201 16:36:09.251040 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61fdf03d-257d-4d46-93c4-8b772371cf61-utilities\") pod \"61fdf03d-257d-4d46-93c4-8b772371cf61\" (UID: \"61fdf03d-257d-4d46-93c4-8b772371cf61\") " Dec 01 16:36:09 crc kubenswrapper[4810]: I1201 16:36:09.251565 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzchj\" (UniqueName: \"kubernetes.io/projected/61fdf03d-257d-4d46-93c4-8b772371cf61-kube-api-access-hzchj\") pod \"61fdf03d-257d-4d46-93c4-8b772371cf61\" (UID: \"61fdf03d-257d-4d46-93c4-8b772371cf61\") " Dec 01 16:36:09 crc kubenswrapper[4810]: I1201 16:36:09.251710 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61fdf03d-257d-4d46-93c4-8b772371cf61-catalog-content\") pod \"61fdf03d-257d-4d46-93c4-8b772371cf61\" (UID: \"61fdf03d-257d-4d46-93c4-8b772371cf61\") " Dec 01 16:36:09 crc kubenswrapper[4810]: I1201 16:36:09.252360 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61fdf03d-257d-4d46-93c4-8b772371cf61-utilities" (OuterVolumeSpecName: "utilities") pod "61fdf03d-257d-4d46-93c4-8b772371cf61" (UID: "61fdf03d-257d-4d46-93c4-8b772371cf61"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:36:09 crc kubenswrapper[4810]: I1201 16:36:09.258455 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61fdf03d-257d-4d46-93c4-8b772371cf61-kube-api-access-hzchj" (OuterVolumeSpecName: "kube-api-access-hzchj") pod "61fdf03d-257d-4d46-93c4-8b772371cf61" (UID: "61fdf03d-257d-4d46-93c4-8b772371cf61"). InnerVolumeSpecName "kube-api-access-hzchj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:36:09 crc kubenswrapper[4810]: I1201 16:36:09.346274 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61fdf03d-257d-4d46-93c4-8b772371cf61-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "61fdf03d-257d-4d46-93c4-8b772371cf61" (UID: "61fdf03d-257d-4d46-93c4-8b772371cf61"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:36:09 crc kubenswrapper[4810]: I1201 16:36:09.354763 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61fdf03d-257d-4d46-93c4-8b772371cf61-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:36:09 crc kubenswrapper[4810]: I1201 16:36:09.354800 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzchj\" (UniqueName: \"kubernetes.io/projected/61fdf03d-257d-4d46-93c4-8b772371cf61-kube-api-access-hzchj\") on node \"crc\" DevicePath \"\"" Dec 01 16:36:09 crc kubenswrapper[4810]: I1201 16:36:09.354811 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61fdf03d-257d-4d46-93c4-8b772371cf61-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:36:10 crc kubenswrapper[4810]: I1201 16:36:10.091605 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lxcfp" Dec 01 16:36:10 crc kubenswrapper[4810]: I1201 16:36:10.126234 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lxcfp"] Dec 01 16:36:10 crc kubenswrapper[4810]: I1201 16:36:10.136266 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lxcfp"] Dec 01 16:36:10 crc kubenswrapper[4810]: I1201 16:36:10.514224 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61fdf03d-257d-4d46-93c4-8b772371cf61" path="/var/lib/kubelet/pods/61fdf03d-257d-4d46-93c4-8b772371cf61/volumes" Dec 01 16:36:16 crc kubenswrapper[4810]: I1201 16:36:16.892528 4810 scope.go:117] "RemoveContainer" containerID="b2e5141706467610d79bea0fc62e605112035d2280a1cf83a3f9a340827bd720" Dec 01 16:36:16 crc kubenswrapper[4810]: I1201 16:36:16.922346 4810 scope.go:117] "RemoveContainer" containerID="782e1c99d910749de0d78a3fb2da995a2e63d505efda016f54c41bf8c7eb5152" Dec 01 16:36:17 crc kubenswrapper[4810]: I1201 16:36:17.043873 4810 scope.go:117] "RemoveContainer" containerID="16e1a603cc9aa502a52d917fae64c2a401a08e03eea47afdcbc47a89b6271eba" Dec 01 16:37:02 crc kubenswrapper[4810]: I1201 16:37:02.971845 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:37:02 crc kubenswrapper[4810]: I1201 16:37:02.972448 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:37:09 crc kubenswrapper[4810]: I1201 16:37:09.781737 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fwd2c"] Dec 01 16:37:09 crc kubenswrapper[4810]: E1201 16:37:09.783231 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fdf03d-257d-4d46-93c4-8b772371cf61" containerName="registry-server" Dec 01 16:37:09 crc kubenswrapper[4810]: I1201 16:37:09.783267 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fdf03d-257d-4d46-93c4-8b772371cf61" containerName="registry-server" Dec 01 16:37:09 crc kubenswrapper[4810]: E1201 16:37:09.783372 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fdf03d-257d-4d46-93c4-8b772371cf61" containerName="extract-content" Dec 01 16:37:09 crc kubenswrapper[4810]: I1201 16:37:09.783391 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fdf03d-257d-4d46-93c4-8b772371cf61" containerName="extract-content" Dec 01 16:37:09 crc kubenswrapper[4810]: E1201 16:37:09.783422 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fdf03d-257d-4d46-93c4-8b772371cf61" containerName="extract-utilities" Dec 01 16:37:09 crc kubenswrapper[4810]: I1201 16:37:09.783438 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fdf03d-257d-4d46-93c4-8b772371cf61" containerName="extract-utilities" Dec 01 16:37:09 crc kubenswrapper[4810]: I1201 16:37:09.783893 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="61fdf03d-257d-4d46-93c4-8b772371cf61" containerName="registry-server" Dec 01 16:37:09 crc kubenswrapper[4810]: I1201 16:37:09.787001 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fwd2c" Dec 01 16:37:09 crc kubenswrapper[4810]: I1201 16:37:09.805720 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fwd2c"] Dec 01 16:37:09 crc kubenswrapper[4810]: I1201 16:37:09.945876 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrjmt\" (UniqueName: \"kubernetes.io/projected/08085514-b92c-49cf-af2c-aaf19c75af3f-kube-api-access-vrjmt\") pod \"community-operators-fwd2c\" (UID: \"08085514-b92c-49cf-af2c-aaf19c75af3f\") " pod="openshift-marketplace/community-operators-fwd2c" Dec 01 16:37:09 crc kubenswrapper[4810]: I1201 16:37:09.946207 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08085514-b92c-49cf-af2c-aaf19c75af3f-catalog-content\") pod \"community-operators-fwd2c\" (UID: \"08085514-b92c-49cf-af2c-aaf19c75af3f\") " pod="openshift-marketplace/community-operators-fwd2c" Dec 01 16:37:09 crc kubenswrapper[4810]: I1201 16:37:09.946338 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08085514-b92c-49cf-af2c-aaf19c75af3f-utilities\") pod \"community-operators-fwd2c\" (UID: \"08085514-b92c-49cf-af2c-aaf19c75af3f\") " pod="openshift-marketplace/community-operators-fwd2c" Dec 01 16:37:10 crc kubenswrapper[4810]: I1201 16:37:10.051394 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrjmt\" (UniqueName: \"kubernetes.io/projected/08085514-b92c-49cf-af2c-aaf19c75af3f-kube-api-access-vrjmt\") pod \"community-operators-fwd2c\" (UID: \"08085514-b92c-49cf-af2c-aaf19c75af3f\") " pod="openshift-marketplace/community-operators-fwd2c" Dec 01 16:37:10 crc kubenswrapper[4810]: I1201 16:37:10.051582 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08085514-b92c-49cf-af2c-aaf19c75af3f-catalog-content\") pod \"community-operators-fwd2c\" (UID: \"08085514-b92c-49cf-af2c-aaf19c75af3f\") " pod="openshift-marketplace/community-operators-fwd2c" Dec 01 16:37:10 crc kubenswrapper[4810]: I1201 16:37:10.051684 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08085514-b92c-49cf-af2c-aaf19c75af3f-utilities\") pod \"community-operators-fwd2c\" (UID: \"08085514-b92c-49cf-af2c-aaf19c75af3f\") " pod="openshift-marketplace/community-operators-fwd2c" Dec 01 16:37:10 crc kubenswrapper[4810]: I1201 16:37:10.052394 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08085514-b92c-49cf-af2c-aaf19c75af3f-utilities\") pod \"community-operators-fwd2c\" (UID: \"08085514-b92c-49cf-af2c-aaf19c75af3f\") " pod="openshift-marketplace/community-operators-fwd2c" Dec 01 16:37:10 crc kubenswrapper[4810]: I1201 16:37:10.052523 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08085514-b92c-49cf-af2c-aaf19c75af3f-catalog-content\") pod \"community-operators-fwd2c\" (UID: \"08085514-b92c-49cf-af2c-aaf19c75af3f\") " pod="openshift-marketplace/community-operators-fwd2c" Dec 01 16:37:10 crc kubenswrapper[4810]: I1201 16:37:10.089600 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrjmt\" (UniqueName: \"kubernetes.io/projected/08085514-b92c-49cf-af2c-aaf19c75af3f-kube-api-access-vrjmt\") pod \"community-operators-fwd2c\" (UID: \"08085514-b92c-49cf-af2c-aaf19c75af3f\") " pod="openshift-marketplace/community-operators-fwd2c" Dec 01 16:37:10 crc kubenswrapper[4810]: I1201 16:37:10.121125 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fwd2c" Dec 01 16:37:10 crc kubenswrapper[4810]: I1201 16:37:10.675141 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fwd2c"] Dec 01 16:37:10 crc kubenswrapper[4810]: I1201 16:37:10.696280 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fwd2c" event={"ID":"08085514-b92c-49cf-af2c-aaf19c75af3f","Type":"ContainerStarted","Data":"d86171ab98ebce5f101b6300505d1eaace412a738515552a9eeada1fc19b55d3"} Dec 01 16:37:11 crc kubenswrapper[4810]: I1201 16:37:11.709976 4810 generic.go:334] "Generic (PLEG): container finished" podID="08085514-b92c-49cf-af2c-aaf19c75af3f" containerID="a608051c2b6b92b93105fde040804e99a49b6abfa61bf4a2c3f18a99b07c3af0" exitCode=0 Dec 01 16:37:11 crc kubenswrapper[4810]: I1201 16:37:11.710086 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fwd2c" event={"ID":"08085514-b92c-49cf-af2c-aaf19c75af3f","Type":"ContainerDied","Data":"a608051c2b6b92b93105fde040804e99a49b6abfa61bf4a2c3f18a99b07c3af0"} Dec 01 16:37:13 crc kubenswrapper[4810]: I1201 16:37:13.731001 4810 generic.go:334] "Generic (PLEG): container finished" podID="08085514-b92c-49cf-af2c-aaf19c75af3f" containerID="521b3d847b30ee1d67f4a3d90b9f1b22fa2dec9272799de915303cc2ad076476" exitCode=0 Dec 01 16:37:13 crc kubenswrapper[4810]: I1201 16:37:13.731061 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fwd2c" event={"ID":"08085514-b92c-49cf-af2c-aaf19c75af3f","Type":"ContainerDied","Data":"521b3d847b30ee1d67f4a3d90b9f1b22fa2dec9272799de915303cc2ad076476"} Dec 01 16:37:13 crc kubenswrapper[4810]: E1201 16:37:13.831028 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08085514_b92c_49cf_af2c_aaf19c75af3f.slice/crio-conmon-521b3d847b30ee1d67f4a3d90b9f1b22fa2dec9272799de915303cc2ad076476.scope\": RecentStats: unable to find data in memory cache]" Dec 01 16:37:14 crc kubenswrapper[4810]: I1201 16:37:14.742609 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fwd2c" event={"ID":"08085514-b92c-49cf-af2c-aaf19c75af3f","Type":"ContainerStarted","Data":"d1286f853ee88d751a0b3bf1cf06cfa83ac8a6fef4e3bb4f6576ee4c267c94cf"} Dec 01 16:37:14 crc kubenswrapper[4810]: I1201 16:37:14.765731 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fwd2c" podStartSLOduration=3.263989344 podStartE2EDuration="5.765706786s" podCreationTimestamp="2025-12-01 16:37:09 +0000 UTC" firstStartedPulling="2025-12-01 16:37:11.711798326 +0000 UTC m=+7397.475307929" lastFinishedPulling="2025-12-01 16:37:14.213515768 +0000 UTC m=+7399.977025371" observedRunningTime="2025-12-01 16:37:14.756409553 +0000 UTC m=+7400.519919166" watchObservedRunningTime="2025-12-01 16:37:14.765706786 +0000 UTC m=+7400.529216389" Dec 01 16:37:20 crc kubenswrapper[4810]: I1201 16:37:20.121751 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fwd2c" Dec 01 16:37:20 crc kubenswrapper[4810]: I1201 16:37:20.122220 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fwd2c" Dec 01 16:37:20 crc kubenswrapper[4810]: I1201 16:37:20.174170 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fwd2c" Dec 01 16:37:20 crc kubenswrapper[4810]: I1201 16:37:20.852115 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fwd2c" Dec 01 16:37:20 crc kubenswrapper[4810]: I1201 16:37:20.901355 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fwd2c"] Dec 01 16:37:22 crc kubenswrapper[4810]: I1201 16:37:22.819906 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fwd2c" podUID="08085514-b92c-49cf-af2c-aaf19c75af3f" containerName="registry-server" containerID="cri-o://d1286f853ee88d751a0b3bf1cf06cfa83ac8a6fef4e3bb4f6576ee4c267c94cf" gracePeriod=2 Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.407447 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fwd2c" Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.496216 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08085514-b92c-49cf-af2c-aaf19c75af3f-utilities\") pod \"08085514-b92c-49cf-af2c-aaf19c75af3f\" (UID: \"08085514-b92c-49cf-af2c-aaf19c75af3f\") " Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.496354 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrjmt\" (UniqueName: \"kubernetes.io/projected/08085514-b92c-49cf-af2c-aaf19c75af3f-kube-api-access-vrjmt\") pod \"08085514-b92c-49cf-af2c-aaf19c75af3f\" (UID: \"08085514-b92c-49cf-af2c-aaf19c75af3f\") " Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.496395 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08085514-b92c-49cf-af2c-aaf19c75af3f-catalog-content\") pod \"08085514-b92c-49cf-af2c-aaf19c75af3f\" (UID: \"08085514-b92c-49cf-af2c-aaf19c75af3f\") " Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.498371 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08085514-b92c-49cf-af2c-aaf19c75af3f-utilities" (OuterVolumeSpecName: "utilities") pod "08085514-b92c-49cf-af2c-aaf19c75af3f" (UID: "08085514-b92c-49cf-af2c-aaf19c75af3f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.504728 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08085514-b92c-49cf-af2c-aaf19c75af3f-kube-api-access-vrjmt" (OuterVolumeSpecName: "kube-api-access-vrjmt") pod "08085514-b92c-49cf-af2c-aaf19c75af3f" (UID: "08085514-b92c-49cf-af2c-aaf19c75af3f"). InnerVolumeSpecName "kube-api-access-vrjmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.566762 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08085514-b92c-49cf-af2c-aaf19c75af3f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08085514-b92c-49cf-af2c-aaf19c75af3f" (UID: "08085514-b92c-49cf-af2c-aaf19c75af3f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.599758 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08085514-b92c-49cf-af2c-aaf19c75af3f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.600026 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08085514-b92c-49cf-af2c-aaf19c75af3f-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.600117 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrjmt\" (UniqueName: \"kubernetes.io/projected/08085514-b92c-49cf-af2c-aaf19c75af3f-kube-api-access-vrjmt\") on node \"crc\" DevicePath \"\"" Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.832427 4810 generic.go:334] "Generic (PLEG): container finished" podID="08085514-b92c-49cf-af2c-aaf19c75af3f" containerID="d1286f853ee88d751a0b3bf1cf06cfa83ac8a6fef4e3bb4f6576ee4c267c94cf" exitCode=0 Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.832485 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fwd2c" event={"ID":"08085514-b92c-49cf-af2c-aaf19c75af3f","Type":"ContainerDied","Data":"d1286f853ee88d751a0b3bf1cf06cfa83ac8a6fef4e3bb4f6576ee4c267c94cf"} Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.832532 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fwd2c" event={"ID":"08085514-b92c-49cf-af2c-aaf19c75af3f","Type":"ContainerDied","Data":"d86171ab98ebce5f101b6300505d1eaace412a738515552a9eeada1fc19b55d3"} Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.832550 4810 scope.go:117] "RemoveContainer" containerID="d1286f853ee88d751a0b3bf1cf06cfa83ac8a6fef4e3bb4f6576ee4c267c94cf" Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.832569 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fwd2c" Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.875610 4810 scope.go:117] "RemoveContainer" containerID="521b3d847b30ee1d67f4a3d90b9f1b22fa2dec9272799de915303cc2ad076476" Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.899899 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fwd2c"] Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.914728 4810 scope.go:117] "RemoveContainer" containerID="a608051c2b6b92b93105fde040804e99a49b6abfa61bf4a2c3f18a99b07c3af0" Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.918632 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fwd2c"] Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.962343 4810 scope.go:117] "RemoveContainer" containerID="d1286f853ee88d751a0b3bf1cf06cfa83ac8a6fef4e3bb4f6576ee4c267c94cf" Dec 01 16:37:23 crc kubenswrapper[4810]: E1201 16:37:23.962836 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1286f853ee88d751a0b3bf1cf06cfa83ac8a6fef4e3bb4f6576ee4c267c94cf\": container with ID starting with d1286f853ee88d751a0b3bf1cf06cfa83ac8a6fef4e3bb4f6576ee4c267c94cf not found: ID does not exist" containerID="d1286f853ee88d751a0b3bf1cf06cfa83ac8a6fef4e3bb4f6576ee4c267c94cf" Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.962873 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1286f853ee88d751a0b3bf1cf06cfa83ac8a6fef4e3bb4f6576ee4c267c94cf"} err="failed to get container status \"d1286f853ee88d751a0b3bf1cf06cfa83ac8a6fef4e3bb4f6576ee4c267c94cf\": rpc error: code = NotFound desc = could not find container \"d1286f853ee88d751a0b3bf1cf06cfa83ac8a6fef4e3bb4f6576ee4c267c94cf\": container with ID starting with d1286f853ee88d751a0b3bf1cf06cfa83ac8a6fef4e3bb4f6576ee4c267c94cf not found: ID does not exist" Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.962902 4810 scope.go:117] "RemoveContainer" containerID="521b3d847b30ee1d67f4a3d90b9f1b22fa2dec9272799de915303cc2ad076476" Dec 01 16:37:23 crc kubenswrapper[4810]: E1201 16:37:23.964601 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"521b3d847b30ee1d67f4a3d90b9f1b22fa2dec9272799de915303cc2ad076476\": container with ID starting with 521b3d847b30ee1d67f4a3d90b9f1b22fa2dec9272799de915303cc2ad076476 not found: ID does not exist" containerID="521b3d847b30ee1d67f4a3d90b9f1b22fa2dec9272799de915303cc2ad076476" Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.964632 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"521b3d847b30ee1d67f4a3d90b9f1b22fa2dec9272799de915303cc2ad076476"} err="failed to get container status \"521b3d847b30ee1d67f4a3d90b9f1b22fa2dec9272799de915303cc2ad076476\": rpc error: code = NotFound desc = could not find container \"521b3d847b30ee1d67f4a3d90b9f1b22fa2dec9272799de915303cc2ad076476\": container with ID starting with 521b3d847b30ee1d67f4a3d90b9f1b22fa2dec9272799de915303cc2ad076476 not found: ID does not exist" Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.964651 4810 scope.go:117] "RemoveContainer" containerID="a608051c2b6b92b93105fde040804e99a49b6abfa61bf4a2c3f18a99b07c3af0" Dec 01 16:37:23 crc kubenswrapper[4810]: E1201 16:37:23.964881 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a608051c2b6b92b93105fde040804e99a49b6abfa61bf4a2c3f18a99b07c3af0\": container with ID starting with a608051c2b6b92b93105fde040804e99a49b6abfa61bf4a2c3f18a99b07c3af0 not found: ID does not exist" containerID="a608051c2b6b92b93105fde040804e99a49b6abfa61bf4a2c3f18a99b07c3af0" Dec 01 16:37:23 crc kubenswrapper[4810]: I1201 16:37:23.964903 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a608051c2b6b92b93105fde040804e99a49b6abfa61bf4a2c3f18a99b07c3af0"} err="failed to get container status \"a608051c2b6b92b93105fde040804e99a49b6abfa61bf4a2c3f18a99b07c3af0\": rpc error: code = NotFound desc = could not find container \"a608051c2b6b92b93105fde040804e99a49b6abfa61bf4a2c3f18a99b07c3af0\": container with ID starting with a608051c2b6b92b93105fde040804e99a49b6abfa61bf4a2c3f18a99b07c3af0 not found: ID does not exist" Dec 01 16:37:24 crc kubenswrapper[4810]: I1201 16:37:24.502358 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08085514-b92c-49cf-af2c-aaf19c75af3f" path="/var/lib/kubelet/pods/08085514-b92c-49cf-af2c-aaf19c75af3f/volumes" Dec 01 16:37:32 crc kubenswrapper[4810]: I1201 16:37:32.971870 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:37:32 crc kubenswrapper[4810]: I1201 16:37:32.972360 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:38:02 crc kubenswrapper[4810]: I1201 16:38:02.972628 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:38:02 crc kubenswrapper[4810]: I1201 16:38:02.973223 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:38:02 crc kubenswrapper[4810]: I1201 16:38:02.973269 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 16:38:02 crc kubenswrapper[4810]: I1201 16:38:02.974101 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:38:02 crc kubenswrapper[4810]: I1201 16:38:02.974169 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" gracePeriod=600 Dec 01 16:38:03 crc kubenswrapper[4810]: E1201 16:38:03.097778 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:38:03 crc kubenswrapper[4810]: I1201 16:38:03.245882 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" exitCode=0 Dec 01 16:38:03 crc kubenswrapper[4810]: I1201 16:38:03.245922 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b"} Dec 01 16:38:03 crc kubenswrapper[4810]: I1201 16:38:03.245954 4810 scope.go:117] "RemoveContainer" containerID="71549b0dadf7e5851cf8c3dca9e008d0d9bb20d1aa5c39eae94356aa087f1657" Dec 01 16:38:03 crc kubenswrapper[4810]: I1201 16:38:03.246634 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:38:03 crc kubenswrapper[4810]: E1201 16:38:03.246875 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:38:17 crc kubenswrapper[4810]: I1201 16:38:17.491676 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:38:17 crc kubenswrapper[4810]: E1201 16:38:17.492514 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:38:32 crc kubenswrapper[4810]: I1201 16:38:32.491774 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:38:32 crc kubenswrapper[4810]: E1201 16:38:32.493746 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:38:44 crc kubenswrapper[4810]: I1201 16:38:44.671419 4810 generic.go:334] "Generic (PLEG): container finished" podID="229207e0-27f2-4caa-b587-8340921ac643" containerID="2ecb722d2226dab5e716ac417d6697dec1117117eeac579d3f08010cf0352e74" exitCode=0 Dec 01 16:38:44 crc kubenswrapper[4810]: I1201 16:38:44.671509 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" event={"ID":"229207e0-27f2-4caa-b587-8340921ac643","Type":"ContainerDied","Data":"2ecb722d2226dab5e716ac417d6697dec1117117eeac579d3f08010cf0352e74"} Dec 01 16:38:46 crc kubenswrapper[4810]: I1201 16:38:46.092893 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" Dec 01 16:38:46 crc kubenswrapper[4810]: I1201 16:38:46.243360 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/229207e0-27f2-4caa-b587-8340921ac643-inventory\") pod \"229207e0-27f2-4caa-b587-8340921ac643\" (UID: \"229207e0-27f2-4caa-b587-8340921ac643\") " Dec 01 16:38:46 crc kubenswrapper[4810]: I1201 16:38:46.243828 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvz4m\" (UniqueName: \"kubernetes.io/projected/229207e0-27f2-4caa-b587-8340921ac643-kube-api-access-nvz4m\") pod \"229207e0-27f2-4caa-b587-8340921ac643\" (UID: \"229207e0-27f2-4caa-b587-8340921ac643\") " Dec 01 16:38:46 crc kubenswrapper[4810]: I1201 16:38:46.243889 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/229207e0-27f2-4caa-b587-8340921ac643-tripleo-cleanup-combined-ca-bundle\") pod \"229207e0-27f2-4caa-b587-8340921ac643\" (UID: \"229207e0-27f2-4caa-b587-8340921ac643\") " Dec 01 16:38:46 crc kubenswrapper[4810]: I1201 16:38:46.243978 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/229207e0-27f2-4caa-b587-8340921ac643-ssh-key\") pod \"229207e0-27f2-4caa-b587-8340921ac643\" (UID: \"229207e0-27f2-4caa-b587-8340921ac643\") " Dec 01 16:38:46 crc kubenswrapper[4810]: I1201 16:38:46.256913 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/229207e0-27f2-4caa-b587-8340921ac643-kube-api-access-nvz4m" (OuterVolumeSpecName: "kube-api-access-nvz4m") pod "229207e0-27f2-4caa-b587-8340921ac643" (UID: "229207e0-27f2-4caa-b587-8340921ac643"). InnerVolumeSpecName "kube-api-access-nvz4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:38:46 crc kubenswrapper[4810]: I1201 16:38:46.256834 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/229207e0-27f2-4caa-b587-8340921ac643-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "229207e0-27f2-4caa-b587-8340921ac643" (UID: "229207e0-27f2-4caa-b587-8340921ac643"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:38:46 crc kubenswrapper[4810]: I1201 16:38:46.272933 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/229207e0-27f2-4caa-b587-8340921ac643-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "229207e0-27f2-4caa-b587-8340921ac643" (UID: "229207e0-27f2-4caa-b587-8340921ac643"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:38:46 crc kubenswrapper[4810]: I1201 16:38:46.275464 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/229207e0-27f2-4caa-b587-8340921ac643-inventory" (OuterVolumeSpecName: "inventory") pod "229207e0-27f2-4caa-b587-8340921ac643" (UID: "229207e0-27f2-4caa-b587-8340921ac643"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:38:46 crc kubenswrapper[4810]: I1201 16:38:46.346275 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/229207e0-27f2-4caa-b587-8340921ac643-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:38:46 crc kubenswrapper[4810]: I1201 16:38:46.346307 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvz4m\" (UniqueName: \"kubernetes.io/projected/229207e0-27f2-4caa-b587-8340921ac643-kube-api-access-nvz4m\") on node \"crc\" DevicePath \"\"" Dec 01 16:38:46 crc kubenswrapper[4810]: I1201 16:38:46.346318 4810 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/229207e0-27f2-4caa-b587-8340921ac643-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:38:46 crc kubenswrapper[4810]: I1201 16:38:46.346329 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/229207e0-27f2-4caa-b587-8340921ac643-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:38:46 crc kubenswrapper[4810]: I1201 16:38:46.690410 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" event={"ID":"229207e0-27f2-4caa-b587-8340921ac643","Type":"ContainerDied","Data":"dd65f296efe42ee8cd2c3b9e78114c8c8cd4ecb0f93a493d067a29a8dc5e63ef"} Dec 01 16:38:46 crc kubenswrapper[4810]: I1201 16:38:46.690455 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd65f296efe42ee8cd2c3b9e78114c8c8cd4ecb0f93a493d067a29a8dc5e63ef" Dec 01 16:38:46 crc kubenswrapper[4810]: I1201 16:38:46.690461 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b" Dec 01 16:38:47 crc kubenswrapper[4810]: I1201 16:38:47.490668 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:38:47 crc kubenswrapper[4810]: E1201 16:38:47.491251 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.581391 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-2fqfj"] Dec 01 16:38:51 crc kubenswrapper[4810]: E1201 16:38:51.582392 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="229207e0-27f2-4caa-b587-8340921ac643" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.582407 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="229207e0-27f2-4caa-b587-8340921ac643" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 01 16:38:51 crc kubenswrapper[4810]: E1201 16:38:51.582423 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08085514-b92c-49cf-af2c-aaf19c75af3f" containerName="extract-content" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.582429 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="08085514-b92c-49cf-af2c-aaf19c75af3f" containerName="extract-content" Dec 01 16:38:51 crc kubenswrapper[4810]: E1201 16:38:51.582450 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08085514-b92c-49cf-af2c-aaf19c75af3f" containerName="extract-utilities" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.582456 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="08085514-b92c-49cf-af2c-aaf19c75af3f" containerName="extract-utilities" Dec 01 16:38:51 crc kubenswrapper[4810]: E1201 16:38:51.582486 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08085514-b92c-49cf-af2c-aaf19c75af3f" containerName="registry-server" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.582493 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="08085514-b92c-49cf-af2c-aaf19c75af3f" containerName="registry-server" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.582761 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="08085514-b92c-49cf-af2c-aaf19c75af3f" containerName="registry-server" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.582790 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="229207e0-27f2-4caa-b587-8340921ac643" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.583668 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.586508 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.586513 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.586926 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.589718 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.591358 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-2fqfj"] Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.761568 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-inventory\") pod \"bootstrap-openstack-openstack-cell1-2fqfj\" (UID: \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\") " pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.761859 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-2fqfj\" (UID: \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\") " pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.762038 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6rrk\" (UniqueName: \"kubernetes.io/projected/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-kube-api-access-f6rrk\") pod \"bootstrap-openstack-openstack-cell1-2fqfj\" (UID: \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\") " pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.762313 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-2fqfj\" (UID: \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\") " pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.863950 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6rrk\" (UniqueName: \"kubernetes.io/projected/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-kube-api-access-f6rrk\") pod \"bootstrap-openstack-openstack-cell1-2fqfj\" (UID: \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\") " pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.864154 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-2fqfj\" (UID: \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\") " pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.864189 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-inventory\") pod \"bootstrap-openstack-openstack-cell1-2fqfj\" (UID: \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\") " pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.864216 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-2fqfj\" (UID: \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\") " pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.869793 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-inventory\") pod \"bootstrap-openstack-openstack-cell1-2fqfj\" (UID: \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\") " pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.876388 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-2fqfj\" (UID: \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\") " pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.878440 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-2fqfj\" (UID: \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\") " pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.881245 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6rrk\" (UniqueName: \"kubernetes.io/projected/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-kube-api-access-f6rrk\") pod \"bootstrap-openstack-openstack-cell1-2fqfj\" (UID: \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\") " pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" Dec 01 16:38:51 crc kubenswrapper[4810]: I1201 16:38:51.913617 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" Dec 01 16:38:52 crc kubenswrapper[4810]: I1201 16:38:52.552248 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-2fqfj"] Dec 01 16:38:52 crc kubenswrapper[4810]: I1201 16:38:52.750186 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" event={"ID":"2f0d06d5-7eea-4c8d-8700-439ea72af7cc","Type":"ContainerStarted","Data":"3789e858b368dc85a19ded5a1d2538435e8dfee2690003d6098b4fc7f8cff262"} Dec 01 16:38:54 crc kubenswrapper[4810]: I1201 16:38:54.781023 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" event={"ID":"2f0d06d5-7eea-4c8d-8700-439ea72af7cc","Type":"ContainerStarted","Data":"d6091cb176357686927d9c2a2e133edacb65459fe4416658ab97ebf37929c029"} Dec 01 16:38:54 crc kubenswrapper[4810]: I1201 16:38:54.809550 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" podStartSLOduration=2.149374335 podStartE2EDuration="3.809526942s" podCreationTimestamp="2025-12-01 16:38:51 +0000 UTC" firstStartedPulling="2025-12-01 16:38:52.554710047 +0000 UTC m=+7498.318219650" lastFinishedPulling="2025-12-01 16:38:54.214862654 +0000 UTC m=+7499.978372257" observedRunningTime="2025-12-01 16:38:54.797101175 +0000 UTC m=+7500.560610788" watchObservedRunningTime="2025-12-01 16:38:54.809526942 +0000 UTC m=+7500.573036545" Dec 01 16:38:58 crc kubenswrapper[4810]: I1201 16:38:58.491279 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:38:58 crc kubenswrapper[4810]: E1201 16:38:58.492143 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:39:10 crc kubenswrapper[4810]: I1201 16:39:10.492348 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:39:10 crc kubenswrapper[4810]: E1201 16:39:10.493394 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:39:24 crc kubenswrapper[4810]: I1201 16:39:24.074708 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9qwgr"] Dec 01 16:39:24 crc kubenswrapper[4810]: I1201 16:39:24.078077 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qwgr" Dec 01 16:39:24 crc kubenswrapper[4810]: I1201 16:39:24.092963 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9qwgr"] Dec 01 16:39:24 crc kubenswrapper[4810]: I1201 16:39:24.174706 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsj8v\" (UniqueName: \"kubernetes.io/projected/b5c248e7-bfad-4689-9ba7-18efc91ed6e7-kube-api-access-gsj8v\") pod \"redhat-operators-9qwgr\" (UID: \"b5c248e7-bfad-4689-9ba7-18efc91ed6e7\") " pod="openshift-marketplace/redhat-operators-9qwgr" Dec 01 16:39:24 crc kubenswrapper[4810]: I1201 16:39:24.175141 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5c248e7-bfad-4689-9ba7-18efc91ed6e7-catalog-content\") pod \"redhat-operators-9qwgr\" (UID: \"b5c248e7-bfad-4689-9ba7-18efc91ed6e7\") " pod="openshift-marketplace/redhat-operators-9qwgr" Dec 01 16:39:24 crc kubenswrapper[4810]: I1201 16:39:24.175166 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5c248e7-bfad-4689-9ba7-18efc91ed6e7-utilities\") pod \"redhat-operators-9qwgr\" (UID: \"b5c248e7-bfad-4689-9ba7-18efc91ed6e7\") " pod="openshift-marketplace/redhat-operators-9qwgr" Dec 01 16:39:24 crc kubenswrapper[4810]: I1201 16:39:24.276504 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5c248e7-bfad-4689-9ba7-18efc91ed6e7-utilities\") pod \"redhat-operators-9qwgr\" (UID: \"b5c248e7-bfad-4689-9ba7-18efc91ed6e7\") " pod="openshift-marketplace/redhat-operators-9qwgr" Dec 01 16:39:24 crc kubenswrapper[4810]: I1201 16:39:24.276541 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5c248e7-bfad-4689-9ba7-18efc91ed6e7-catalog-content\") pod \"redhat-operators-9qwgr\" (UID: \"b5c248e7-bfad-4689-9ba7-18efc91ed6e7\") " pod="openshift-marketplace/redhat-operators-9qwgr" Dec 01 16:39:24 crc kubenswrapper[4810]: I1201 16:39:24.277027 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5c248e7-bfad-4689-9ba7-18efc91ed6e7-catalog-content\") pod \"redhat-operators-9qwgr\" (UID: \"b5c248e7-bfad-4689-9ba7-18efc91ed6e7\") " pod="openshift-marketplace/redhat-operators-9qwgr" Dec 01 16:39:24 crc kubenswrapper[4810]: I1201 16:39:24.277090 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsj8v\" (UniqueName: \"kubernetes.io/projected/b5c248e7-bfad-4689-9ba7-18efc91ed6e7-kube-api-access-gsj8v\") pod \"redhat-operators-9qwgr\" (UID: \"b5c248e7-bfad-4689-9ba7-18efc91ed6e7\") " pod="openshift-marketplace/redhat-operators-9qwgr" Dec 01 16:39:24 crc kubenswrapper[4810]: I1201 16:39:24.277632 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5c248e7-bfad-4689-9ba7-18efc91ed6e7-utilities\") pod \"redhat-operators-9qwgr\" (UID: \"b5c248e7-bfad-4689-9ba7-18efc91ed6e7\") " pod="openshift-marketplace/redhat-operators-9qwgr" Dec 01 16:39:24 crc kubenswrapper[4810]: I1201 16:39:24.296653 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsj8v\" (UniqueName: \"kubernetes.io/projected/b5c248e7-bfad-4689-9ba7-18efc91ed6e7-kube-api-access-gsj8v\") pod \"redhat-operators-9qwgr\" (UID: \"b5c248e7-bfad-4689-9ba7-18efc91ed6e7\") " pod="openshift-marketplace/redhat-operators-9qwgr" Dec 01 16:39:24 crc kubenswrapper[4810]: I1201 16:39:24.435949 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qwgr" Dec 01 16:39:24 crc kubenswrapper[4810]: I1201 16:39:24.508146 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:39:24 crc kubenswrapper[4810]: E1201 16:39:24.508538 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:39:24 crc kubenswrapper[4810]: I1201 16:39:24.934705 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9qwgr"] Dec 01 16:39:25 crc kubenswrapper[4810]: I1201 16:39:25.062750 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qwgr" event={"ID":"b5c248e7-bfad-4689-9ba7-18efc91ed6e7","Type":"ContainerStarted","Data":"84054bf1bde7265e06cf562c0a599fe6a2f469a035d35016851665af29a3ceac"} Dec 01 16:39:26 crc kubenswrapper[4810]: I1201 16:39:26.101505 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qwgr" event={"ID":"b5c248e7-bfad-4689-9ba7-18efc91ed6e7","Type":"ContainerDied","Data":"95615a98fb2179a95ffb71253c4e351789cb6a187e9593f326b6a281e7b88fca"} Dec 01 16:39:26 crc kubenswrapper[4810]: I1201 16:39:26.102517 4810 generic.go:334] "Generic (PLEG): container finished" podID="b5c248e7-bfad-4689-9ba7-18efc91ed6e7" containerID="95615a98fb2179a95ffb71253c4e351789cb6a187e9593f326b6a281e7b88fca" exitCode=0 Dec 01 16:39:28 crc kubenswrapper[4810]: I1201 16:39:28.121037 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qwgr" event={"ID":"b5c248e7-bfad-4689-9ba7-18efc91ed6e7","Type":"ContainerStarted","Data":"67a64323611548bbbe767f0ee23950d1f1b745d6112c5c075e79ea9732277e8a"} Dec 01 16:39:30 crc kubenswrapper[4810]: I1201 16:39:30.142584 4810 generic.go:334] "Generic (PLEG): container finished" podID="b5c248e7-bfad-4689-9ba7-18efc91ed6e7" containerID="67a64323611548bbbe767f0ee23950d1f1b745d6112c5c075e79ea9732277e8a" exitCode=0 Dec 01 16:39:30 crc kubenswrapper[4810]: I1201 16:39:30.142696 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qwgr" event={"ID":"b5c248e7-bfad-4689-9ba7-18efc91ed6e7","Type":"ContainerDied","Data":"67a64323611548bbbe767f0ee23950d1f1b745d6112c5c075e79ea9732277e8a"} Dec 01 16:39:31 crc kubenswrapper[4810]: I1201 16:39:31.157675 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qwgr" event={"ID":"b5c248e7-bfad-4689-9ba7-18efc91ed6e7","Type":"ContainerStarted","Data":"b8e7e287c15ed922bfd2811f2a966ab65718a3171fd13f345715c0a177dac0c0"} Dec 01 16:39:31 crc kubenswrapper[4810]: I1201 16:39:31.188556 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9qwgr" podStartSLOduration=2.483741414 podStartE2EDuration="7.18853499s" podCreationTimestamp="2025-12-01 16:39:24 +0000 UTC" firstStartedPulling="2025-12-01 16:39:26.103906112 +0000 UTC m=+7531.867415735" lastFinishedPulling="2025-12-01 16:39:30.808699668 +0000 UTC m=+7536.572209311" observedRunningTime="2025-12-01 16:39:31.185310122 +0000 UTC m=+7536.948819725" watchObservedRunningTime="2025-12-01 16:39:31.18853499 +0000 UTC m=+7536.952044593" Dec 01 16:39:32 crc kubenswrapper[4810]: I1201 16:39:32.062323 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tq4n2"] Dec 01 16:39:32 crc kubenswrapper[4810]: I1201 16:39:32.064645 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tq4n2" Dec 01 16:39:32 crc kubenswrapper[4810]: I1201 16:39:32.075165 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tq4n2"] Dec 01 16:39:32 crc kubenswrapper[4810]: I1201 16:39:32.115904 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94e9a64a-8058-4096-b35d-9ce576d0aea7-utilities\") pod \"redhat-marketplace-tq4n2\" (UID: \"94e9a64a-8058-4096-b35d-9ce576d0aea7\") " pod="openshift-marketplace/redhat-marketplace-tq4n2" Dec 01 16:39:32 crc kubenswrapper[4810]: I1201 16:39:32.116325 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj786\" (UniqueName: \"kubernetes.io/projected/94e9a64a-8058-4096-b35d-9ce576d0aea7-kube-api-access-dj786\") pod \"redhat-marketplace-tq4n2\" (UID: \"94e9a64a-8058-4096-b35d-9ce576d0aea7\") " pod="openshift-marketplace/redhat-marketplace-tq4n2" Dec 01 16:39:32 crc kubenswrapper[4810]: I1201 16:39:32.116539 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94e9a64a-8058-4096-b35d-9ce576d0aea7-catalog-content\") pod \"redhat-marketplace-tq4n2\" (UID: \"94e9a64a-8058-4096-b35d-9ce576d0aea7\") " pod="openshift-marketplace/redhat-marketplace-tq4n2" Dec 01 16:39:32 crc kubenswrapper[4810]: I1201 16:39:32.218491 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94e9a64a-8058-4096-b35d-9ce576d0aea7-utilities\") pod \"redhat-marketplace-tq4n2\" (UID: \"94e9a64a-8058-4096-b35d-9ce576d0aea7\") " pod="openshift-marketplace/redhat-marketplace-tq4n2" Dec 01 16:39:32 crc kubenswrapper[4810]: I1201 16:39:32.218604 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj786\" (UniqueName: \"kubernetes.io/projected/94e9a64a-8058-4096-b35d-9ce576d0aea7-kube-api-access-dj786\") pod \"redhat-marketplace-tq4n2\" (UID: \"94e9a64a-8058-4096-b35d-9ce576d0aea7\") " pod="openshift-marketplace/redhat-marketplace-tq4n2" Dec 01 16:39:32 crc kubenswrapper[4810]: I1201 16:39:32.218651 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94e9a64a-8058-4096-b35d-9ce576d0aea7-catalog-content\") pod \"redhat-marketplace-tq4n2\" (UID: \"94e9a64a-8058-4096-b35d-9ce576d0aea7\") " pod="openshift-marketplace/redhat-marketplace-tq4n2" Dec 01 16:39:32 crc kubenswrapper[4810]: I1201 16:39:32.219196 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94e9a64a-8058-4096-b35d-9ce576d0aea7-utilities\") pod \"redhat-marketplace-tq4n2\" (UID: \"94e9a64a-8058-4096-b35d-9ce576d0aea7\") " pod="openshift-marketplace/redhat-marketplace-tq4n2" Dec 01 16:39:32 crc kubenswrapper[4810]: I1201 16:39:32.219262 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94e9a64a-8058-4096-b35d-9ce576d0aea7-catalog-content\") pod \"redhat-marketplace-tq4n2\" (UID: \"94e9a64a-8058-4096-b35d-9ce576d0aea7\") " pod="openshift-marketplace/redhat-marketplace-tq4n2" Dec 01 16:39:32 crc kubenswrapper[4810]: I1201 16:39:32.237842 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj786\" (UniqueName: \"kubernetes.io/projected/94e9a64a-8058-4096-b35d-9ce576d0aea7-kube-api-access-dj786\") pod \"redhat-marketplace-tq4n2\" (UID: \"94e9a64a-8058-4096-b35d-9ce576d0aea7\") " pod="openshift-marketplace/redhat-marketplace-tq4n2" Dec 01 16:39:32 crc kubenswrapper[4810]: I1201 16:39:32.422882 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tq4n2" Dec 01 16:39:32 crc kubenswrapper[4810]: I1201 16:39:32.924415 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tq4n2"] Dec 01 16:39:32 crc kubenswrapper[4810]: W1201 16:39:32.926748 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94e9a64a_8058_4096_b35d_9ce576d0aea7.slice/crio-aa1dbf1259d8dc68a88781500550b6af3e03653d6bae60b7c65f67e5502a4c91 WatchSource:0}: Error finding container aa1dbf1259d8dc68a88781500550b6af3e03653d6bae60b7c65f67e5502a4c91: Status 404 returned error can't find the container with id aa1dbf1259d8dc68a88781500550b6af3e03653d6bae60b7c65f67e5502a4c91 Dec 01 16:39:33 crc kubenswrapper[4810]: I1201 16:39:33.179691 4810 generic.go:334] "Generic (PLEG): container finished" podID="94e9a64a-8058-4096-b35d-9ce576d0aea7" containerID="f75d83fa3a1518dc26802256be348377bb4d438a7945f243137a3d53e62d7ff8" exitCode=0 Dec 01 16:39:33 crc kubenswrapper[4810]: I1201 16:39:33.179844 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tq4n2" event={"ID":"94e9a64a-8058-4096-b35d-9ce576d0aea7","Type":"ContainerDied","Data":"f75d83fa3a1518dc26802256be348377bb4d438a7945f243137a3d53e62d7ff8"} Dec 01 16:39:33 crc kubenswrapper[4810]: I1201 16:39:33.180037 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tq4n2" event={"ID":"94e9a64a-8058-4096-b35d-9ce576d0aea7","Type":"ContainerStarted","Data":"aa1dbf1259d8dc68a88781500550b6af3e03653d6bae60b7c65f67e5502a4c91"} Dec 01 16:39:34 crc kubenswrapper[4810]: I1201 16:39:34.191227 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tq4n2" event={"ID":"94e9a64a-8058-4096-b35d-9ce576d0aea7","Type":"ContainerStarted","Data":"62a3d87150fb5a3d57882c290ee466c4080679dbfdfe4e1ffcbd1513bb25d538"} Dec 01 16:39:34 crc kubenswrapper[4810]: I1201 16:39:34.436650 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9qwgr" Dec 01 16:39:34 crc kubenswrapper[4810]: I1201 16:39:34.437656 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9qwgr" Dec 01 16:39:35 crc kubenswrapper[4810]: I1201 16:39:35.202265 4810 generic.go:334] "Generic (PLEG): container finished" podID="94e9a64a-8058-4096-b35d-9ce576d0aea7" containerID="62a3d87150fb5a3d57882c290ee466c4080679dbfdfe4e1ffcbd1513bb25d538" exitCode=0 Dec 01 16:39:35 crc kubenswrapper[4810]: I1201 16:39:35.202374 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tq4n2" event={"ID":"94e9a64a-8058-4096-b35d-9ce576d0aea7","Type":"ContainerDied","Data":"62a3d87150fb5a3d57882c290ee466c4080679dbfdfe4e1ffcbd1513bb25d538"} Dec 01 16:39:35 crc kubenswrapper[4810]: I1201 16:39:35.599490 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9qwgr" podUID="b5c248e7-bfad-4689-9ba7-18efc91ed6e7" containerName="registry-server" probeResult="failure" output=< Dec 01 16:39:35 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Dec 01 16:39:35 crc kubenswrapper[4810]: > Dec 01 16:39:37 crc kubenswrapper[4810]: I1201 16:39:37.222121 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tq4n2" event={"ID":"94e9a64a-8058-4096-b35d-9ce576d0aea7","Type":"ContainerStarted","Data":"f670e354b67221a6e7809c8e4521fd0194b89c287ba0908d5e94850b18610459"} Dec 01 16:39:37 crc kubenswrapper[4810]: I1201 16:39:37.252560 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tq4n2" podStartSLOduration=2.6996551909999997 podStartE2EDuration="5.252538391s" podCreationTimestamp="2025-12-01 16:39:32 +0000 UTC" firstStartedPulling="2025-12-01 16:39:33.181625497 +0000 UTC m=+7538.945135100" lastFinishedPulling="2025-12-01 16:39:35.734508697 +0000 UTC m=+7541.498018300" observedRunningTime="2025-12-01 16:39:37.246312012 +0000 UTC m=+7543.009821615" watchObservedRunningTime="2025-12-01 16:39:37.252538391 +0000 UTC m=+7543.016047994" Dec 01 16:39:38 crc kubenswrapper[4810]: I1201 16:39:38.494524 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:39:38 crc kubenswrapper[4810]: E1201 16:39:38.495328 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:39:42 crc kubenswrapper[4810]: I1201 16:39:42.423934 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tq4n2" Dec 01 16:39:42 crc kubenswrapper[4810]: I1201 16:39:42.424458 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tq4n2" Dec 01 16:39:42 crc kubenswrapper[4810]: I1201 16:39:42.478736 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tq4n2" Dec 01 16:39:43 crc kubenswrapper[4810]: I1201 16:39:43.353023 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tq4n2" Dec 01 16:39:43 crc kubenswrapper[4810]: I1201 16:39:43.418023 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tq4n2"] Dec 01 16:39:44 crc kubenswrapper[4810]: I1201 16:39:44.483649 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9qwgr" Dec 01 16:39:44 crc kubenswrapper[4810]: I1201 16:39:44.541096 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9qwgr" Dec 01 16:39:45 crc kubenswrapper[4810]: I1201 16:39:45.118690 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9qwgr"] Dec 01 16:39:45 crc kubenswrapper[4810]: I1201 16:39:45.308998 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tq4n2" podUID="94e9a64a-8058-4096-b35d-9ce576d0aea7" containerName="registry-server" containerID="cri-o://f670e354b67221a6e7809c8e4521fd0194b89c287ba0908d5e94850b18610459" gracePeriod=2 Dec 01 16:39:45 crc kubenswrapper[4810]: I1201 16:39:45.802024 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tq4n2" Dec 01 16:39:45 crc kubenswrapper[4810]: I1201 16:39:45.960152 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94e9a64a-8058-4096-b35d-9ce576d0aea7-catalog-content\") pod \"94e9a64a-8058-4096-b35d-9ce576d0aea7\" (UID: \"94e9a64a-8058-4096-b35d-9ce576d0aea7\") " Dec 01 16:39:45 crc kubenswrapper[4810]: I1201 16:39:45.960239 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94e9a64a-8058-4096-b35d-9ce576d0aea7-utilities\") pod \"94e9a64a-8058-4096-b35d-9ce576d0aea7\" (UID: \"94e9a64a-8058-4096-b35d-9ce576d0aea7\") " Dec 01 16:39:45 crc kubenswrapper[4810]: I1201 16:39:45.960444 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dj786\" (UniqueName: \"kubernetes.io/projected/94e9a64a-8058-4096-b35d-9ce576d0aea7-kube-api-access-dj786\") pod \"94e9a64a-8058-4096-b35d-9ce576d0aea7\" (UID: \"94e9a64a-8058-4096-b35d-9ce576d0aea7\") " Dec 01 16:39:45 crc kubenswrapper[4810]: I1201 16:39:45.962441 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94e9a64a-8058-4096-b35d-9ce576d0aea7-utilities" (OuterVolumeSpecName: "utilities") pod "94e9a64a-8058-4096-b35d-9ce576d0aea7" (UID: "94e9a64a-8058-4096-b35d-9ce576d0aea7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:39:45 crc kubenswrapper[4810]: I1201 16:39:45.967834 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94e9a64a-8058-4096-b35d-9ce576d0aea7-kube-api-access-dj786" (OuterVolumeSpecName: "kube-api-access-dj786") pod "94e9a64a-8058-4096-b35d-9ce576d0aea7" (UID: "94e9a64a-8058-4096-b35d-9ce576d0aea7"). InnerVolumeSpecName "kube-api-access-dj786". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:39:45 crc kubenswrapper[4810]: I1201 16:39:45.988550 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94e9a64a-8058-4096-b35d-9ce576d0aea7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94e9a64a-8058-4096-b35d-9ce576d0aea7" (UID: "94e9a64a-8058-4096-b35d-9ce576d0aea7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.062509 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94e9a64a-8058-4096-b35d-9ce576d0aea7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.062544 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94e9a64a-8058-4096-b35d-9ce576d0aea7-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.062554 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dj786\" (UniqueName: \"kubernetes.io/projected/94e9a64a-8058-4096-b35d-9ce576d0aea7-kube-api-access-dj786\") on node \"crc\" DevicePath \"\"" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.319179 4810 generic.go:334] "Generic (PLEG): container finished" podID="94e9a64a-8058-4096-b35d-9ce576d0aea7" containerID="f670e354b67221a6e7809c8e4521fd0194b89c287ba0908d5e94850b18610459" exitCode=0 Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.319263 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tq4n2" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.319294 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tq4n2" event={"ID":"94e9a64a-8058-4096-b35d-9ce576d0aea7","Type":"ContainerDied","Data":"f670e354b67221a6e7809c8e4521fd0194b89c287ba0908d5e94850b18610459"} Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.319361 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tq4n2" event={"ID":"94e9a64a-8058-4096-b35d-9ce576d0aea7","Type":"ContainerDied","Data":"aa1dbf1259d8dc68a88781500550b6af3e03653d6bae60b7c65f67e5502a4c91"} Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.319382 4810 scope.go:117] "RemoveContainer" containerID="f670e354b67221a6e7809c8e4521fd0194b89c287ba0908d5e94850b18610459" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.319389 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9qwgr" podUID="b5c248e7-bfad-4689-9ba7-18efc91ed6e7" containerName="registry-server" containerID="cri-o://b8e7e287c15ed922bfd2811f2a966ab65718a3171fd13f345715c0a177dac0c0" gracePeriod=2 Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.359508 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tq4n2"] Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.371422 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tq4n2"] Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.373790 4810 scope.go:117] "RemoveContainer" containerID="62a3d87150fb5a3d57882c290ee466c4080679dbfdfe4e1ffcbd1513bb25d538" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.398054 4810 scope.go:117] "RemoveContainer" containerID="f75d83fa3a1518dc26802256be348377bb4d438a7945f243137a3d53e62d7ff8" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.503831 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94e9a64a-8058-4096-b35d-9ce576d0aea7" path="/var/lib/kubelet/pods/94e9a64a-8058-4096-b35d-9ce576d0aea7/volumes" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.585992 4810 scope.go:117] "RemoveContainer" containerID="f670e354b67221a6e7809c8e4521fd0194b89c287ba0908d5e94850b18610459" Dec 01 16:39:46 crc kubenswrapper[4810]: E1201 16:39:46.586491 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f670e354b67221a6e7809c8e4521fd0194b89c287ba0908d5e94850b18610459\": container with ID starting with f670e354b67221a6e7809c8e4521fd0194b89c287ba0908d5e94850b18610459 not found: ID does not exist" containerID="f670e354b67221a6e7809c8e4521fd0194b89c287ba0908d5e94850b18610459" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.586547 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f670e354b67221a6e7809c8e4521fd0194b89c287ba0908d5e94850b18610459"} err="failed to get container status \"f670e354b67221a6e7809c8e4521fd0194b89c287ba0908d5e94850b18610459\": rpc error: code = NotFound desc = could not find container \"f670e354b67221a6e7809c8e4521fd0194b89c287ba0908d5e94850b18610459\": container with ID starting with f670e354b67221a6e7809c8e4521fd0194b89c287ba0908d5e94850b18610459 not found: ID does not exist" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.586569 4810 scope.go:117] "RemoveContainer" containerID="62a3d87150fb5a3d57882c290ee466c4080679dbfdfe4e1ffcbd1513bb25d538" Dec 01 16:39:46 crc kubenswrapper[4810]: E1201 16:39:46.586814 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62a3d87150fb5a3d57882c290ee466c4080679dbfdfe4e1ffcbd1513bb25d538\": container with ID starting with 62a3d87150fb5a3d57882c290ee466c4080679dbfdfe4e1ffcbd1513bb25d538 not found: ID does not exist" containerID="62a3d87150fb5a3d57882c290ee466c4080679dbfdfe4e1ffcbd1513bb25d538" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.586832 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62a3d87150fb5a3d57882c290ee466c4080679dbfdfe4e1ffcbd1513bb25d538"} err="failed to get container status \"62a3d87150fb5a3d57882c290ee466c4080679dbfdfe4e1ffcbd1513bb25d538\": rpc error: code = NotFound desc = could not find container \"62a3d87150fb5a3d57882c290ee466c4080679dbfdfe4e1ffcbd1513bb25d538\": container with ID starting with 62a3d87150fb5a3d57882c290ee466c4080679dbfdfe4e1ffcbd1513bb25d538 not found: ID does not exist" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.586844 4810 scope.go:117] "RemoveContainer" containerID="f75d83fa3a1518dc26802256be348377bb4d438a7945f243137a3d53e62d7ff8" Dec 01 16:39:46 crc kubenswrapper[4810]: E1201 16:39:46.587122 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f75d83fa3a1518dc26802256be348377bb4d438a7945f243137a3d53e62d7ff8\": container with ID starting with f75d83fa3a1518dc26802256be348377bb4d438a7945f243137a3d53e62d7ff8 not found: ID does not exist" containerID="f75d83fa3a1518dc26802256be348377bb4d438a7945f243137a3d53e62d7ff8" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.587140 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f75d83fa3a1518dc26802256be348377bb4d438a7945f243137a3d53e62d7ff8"} err="failed to get container status \"f75d83fa3a1518dc26802256be348377bb4d438a7945f243137a3d53e62d7ff8\": rpc error: code = NotFound desc = could not find container \"f75d83fa3a1518dc26802256be348377bb4d438a7945f243137a3d53e62d7ff8\": container with ID starting with f75d83fa3a1518dc26802256be348377bb4d438a7945f243137a3d53e62d7ff8 not found: ID does not exist" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.772614 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qwgr" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.881308 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsj8v\" (UniqueName: \"kubernetes.io/projected/b5c248e7-bfad-4689-9ba7-18efc91ed6e7-kube-api-access-gsj8v\") pod \"b5c248e7-bfad-4689-9ba7-18efc91ed6e7\" (UID: \"b5c248e7-bfad-4689-9ba7-18efc91ed6e7\") " Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.881416 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5c248e7-bfad-4689-9ba7-18efc91ed6e7-catalog-content\") pod \"b5c248e7-bfad-4689-9ba7-18efc91ed6e7\" (UID: \"b5c248e7-bfad-4689-9ba7-18efc91ed6e7\") " Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.881464 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5c248e7-bfad-4689-9ba7-18efc91ed6e7-utilities\") pod \"b5c248e7-bfad-4689-9ba7-18efc91ed6e7\" (UID: \"b5c248e7-bfad-4689-9ba7-18efc91ed6e7\") " Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.882417 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5c248e7-bfad-4689-9ba7-18efc91ed6e7-utilities" (OuterVolumeSpecName: "utilities") pod "b5c248e7-bfad-4689-9ba7-18efc91ed6e7" (UID: "b5c248e7-bfad-4689-9ba7-18efc91ed6e7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.891813 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5c248e7-bfad-4689-9ba7-18efc91ed6e7-kube-api-access-gsj8v" (OuterVolumeSpecName: "kube-api-access-gsj8v") pod "b5c248e7-bfad-4689-9ba7-18efc91ed6e7" (UID: "b5c248e7-bfad-4689-9ba7-18efc91ed6e7"). InnerVolumeSpecName "kube-api-access-gsj8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.985831 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsj8v\" (UniqueName: \"kubernetes.io/projected/b5c248e7-bfad-4689-9ba7-18efc91ed6e7-kube-api-access-gsj8v\") on node \"crc\" DevicePath \"\"" Dec 01 16:39:46 crc kubenswrapper[4810]: I1201 16:39:46.985873 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5c248e7-bfad-4689-9ba7-18efc91ed6e7-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:39:47 crc kubenswrapper[4810]: I1201 16:39:47.037583 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5c248e7-bfad-4689-9ba7-18efc91ed6e7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5c248e7-bfad-4689-9ba7-18efc91ed6e7" (UID: "b5c248e7-bfad-4689-9ba7-18efc91ed6e7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:39:47 crc kubenswrapper[4810]: I1201 16:39:47.087974 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5c248e7-bfad-4689-9ba7-18efc91ed6e7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:39:47 crc kubenswrapper[4810]: I1201 16:39:47.371687 4810 generic.go:334] "Generic (PLEG): container finished" podID="b5c248e7-bfad-4689-9ba7-18efc91ed6e7" containerID="b8e7e287c15ed922bfd2811f2a966ab65718a3171fd13f345715c0a177dac0c0" exitCode=0 Dec 01 16:39:47 crc kubenswrapper[4810]: I1201 16:39:47.371751 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qwgr" event={"ID":"b5c248e7-bfad-4689-9ba7-18efc91ed6e7","Type":"ContainerDied","Data":"b8e7e287c15ed922bfd2811f2a966ab65718a3171fd13f345715c0a177dac0c0"} Dec 01 16:39:47 crc kubenswrapper[4810]: I1201 16:39:47.371774 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9qwgr" Dec 01 16:39:47 crc kubenswrapper[4810]: I1201 16:39:47.371795 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9qwgr" event={"ID":"b5c248e7-bfad-4689-9ba7-18efc91ed6e7","Type":"ContainerDied","Data":"84054bf1bde7265e06cf562c0a599fe6a2f469a035d35016851665af29a3ceac"} Dec 01 16:39:47 crc kubenswrapper[4810]: I1201 16:39:47.371817 4810 scope.go:117] "RemoveContainer" containerID="b8e7e287c15ed922bfd2811f2a966ab65718a3171fd13f345715c0a177dac0c0" Dec 01 16:39:47 crc kubenswrapper[4810]: I1201 16:39:47.397591 4810 scope.go:117] "RemoveContainer" containerID="67a64323611548bbbe767f0ee23950d1f1b745d6112c5c075e79ea9732277e8a" Dec 01 16:39:47 crc kubenswrapper[4810]: I1201 16:39:47.412994 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9qwgr"] Dec 01 16:39:47 crc kubenswrapper[4810]: I1201 16:39:47.431056 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9qwgr"] Dec 01 16:39:47 crc kubenswrapper[4810]: I1201 16:39:47.462490 4810 scope.go:117] "RemoveContainer" containerID="95615a98fb2179a95ffb71253c4e351789cb6a187e9593f326b6a281e7b88fca" Dec 01 16:39:47 crc kubenswrapper[4810]: I1201 16:39:47.502298 4810 scope.go:117] "RemoveContainer" containerID="b8e7e287c15ed922bfd2811f2a966ab65718a3171fd13f345715c0a177dac0c0" Dec 01 16:39:47 crc kubenswrapper[4810]: E1201 16:39:47.502953 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8e7e287c15ed922bfd2811f2a966ab65718a3171fd13f345715c0a177dac0c0\": container with ID starting with b8e7e287c15ed922bfd2811f2a966ab65718a3171fd13f345715c0a177dac0c0 not found: ID does not exist" containerID="b8e7e287c15ed922bfd2811f2a966ab65718a3171fd13f345715c0a177dac0c0" Dec 01 16:39:47 crc kubenswrapper[4810]: I1201 16:39:47.503011 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8e7e287c15ed922bfd2811f2a966ab65718a3171fd13f345715c0a177dac0c0"} err="failed to get container status \"b8e7e287c15ed922bfd2811f2a966ab65718a3171fd13f345715c0a177dac0c0\": rpc error: code = NotFound desc = could not find container \"b8e7e287c15ed922bfd2811f2a966ab65718a3171fd13f345715c0a177dac0c0\": container with ID starting with b8e7e287c15ed922bfd2811f2a966ab65718a3171fd13f345715c0a177dac0c0 not found: ID does not exist" Dec 01 16:39:47 crc kubenswrapper[4810]: I1201 16:39:47.503046 4810 scope.go:117] "RemoveContainer" containerID="67a64323611548bbbe767f0ee23950d1f1b745d6112c5c075e79ea9732277e8a" Dec 01 16:39:47 crc kubenswrapper[4810]: E1201 16:39:47.503399 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67a64323611548bbbe767f0ee23950d1f1b745d6112c5c075e79ea9732277e8a\": container with ID starting with 67a64323611548bbbe767f0ee23950d1f1b745d6112c5c075e79ea9732277e8a not found: ID does not exist" containerID="67a64323611548bbbe767f0ee23950d1f1b745d6112c5c075e79ea9732277e8a" Dec 01 16:39:47 crc kubenswrapper[4810]: I1201 16:39:47.503451 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67a64323611548bbbe767f0ee23950d1f1b745d6112c5c075e79ea9732277e8a"} err="failed to get container status \"67a64323611548bbbe767f0ee23950d1f1b745d6112c5c075e79ea9732277e8a\": rpc error: code = NotFound desc = could not find container \"67a64323611548bbbe767f0ee23950d1f1b745d6112c5c075e79ea9732277e8a\": container with ID starting with 67a64323611548bbbe767f0ee23950d1f1b745d6112c5c075e79ea9732277e8a not found: ID does not exist" Dec 01 16:39:47 crc kubenswrapper[4810]: I1201 16:39:47.503492 4810 scope.go:117] "RemoveContainer" containerID="95615a98fb2179a95ffb71253c4e351789cb6a187e9593f326b6a281e7b88fca" Dec 01 16:39:47 crc kubenswrapper[4810]: E1201 16:39:47.503856 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95615a98fb2179a95ffb71253c4e351789cb6a187e9593f326b6a281e7b88fca\": container with ID starting with 95615a98fb2179a95ffb71253c4e351789cb6a187e9593f326b6a281e7b88fca not found: ID does not exist" containerID="95615a98fb2179a95ffb71253c4e351789cb6a187e9593f326b6a281e7b88fca" Dec 01 16:39:47 crc kubenswrapper[4810]: I1201 16:39:47.503912 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95615a98fb2179a95ffb71253c4e351789cb6a187e9593f326b6a281e7b88fca"} err="failed to get container status \"95615a98fb2179a95ffb71253c4e351789cb6a187e9593f326b6a281e7b88fca\": rpc error: code = NotFound desc = could not find container \"95615a98fb2179a95ffb71253c4e351789cb6a187e9593f326b6a281e7b88fca\": container with ID starting with 95615a98fb2179a95ffb71253c4e351789cb6a187e9593f326b6a281e7b88fca not found: ID does not exist" Dec 01 16:39:48 crc kubenswrapper[4810]: I1201 16:39:48.503631 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5c248e7-bfad-4689-9ba7-18efc91ed6e7" path="/var/lib/kubelet/pods/b5c248e7-bfad-4689-9ba7-18efc91ed6e7/volumes" Dec 01 16:39:49 crc kubenswrapper[4810]: I1201 16:39:49.491713 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:39:49 crc kubenswrapper[4810]: E1201 16:39:49.492163 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:40:00 crc kubenswrapper[4810]: I1201 16:40:00.493943 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:40:00 crc kubenswrapper[4810]: E1201 16:40:00.495703 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:40:13 crc kubenswrapper[4810]: I1201 16:40:13.491376 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:40:13 crc kubenswrapper[4810]: E1201 16:40:13.492280 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:40:26 crc kubenswrapper[4810]: I1201 16:40:26.491839 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:40:26 crc kubenswrapper[4810]: E1201 16:40:26.492823 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:40:39 crc kubenswrapper[4810]: I1201 16:40:39.491349 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:40:39 crc kubenswrapper[4810]: E1201 16:40:39.492340 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:40:51 crc kubenswrapper[4810]: I1201 16:40:51.491599 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:40:51 crc kubenswrapper[4810]: E1201 16:40:51.492229 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:41:05 crc kubenswrapper[4810]: I1201 16:41:05.491115 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:41:05 crc kubenswrapper[4810]: E1201 16:41:05.491904 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:41:17 crc kubenswrapper[4810]: I1201 16:41:17.491710 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:41:17 crc kubenswrapper[4810]: E1201 16:41:17.492664 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:41:31 crc kubenswrapper[4810]: I1201 16:41:31.491751 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:41:31 crc kubenswrapper[4810]: E1201 16:41:31.492535 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:41:43 crc kubenswrapper[4810]: I1201 16:41:43.490653 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:41:43 crc kubenswrapper[4810]: E1201 16:41:43.491432 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:41:54 crc kubenswrapper[4810]: I1201 16:41:54.498933 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:41:54 crc kubenswrapper[4810]: E1201 16:41:54.499724 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:41:59 crc kubenswrapper[4810]: I1201 16:41:59.678189 4810 generic.go:334] "Generic (PLEG): container finished" podID="2f0d06d5-7eea-4c8d-8700-439ea72af7cc" containerID="d6091cb176357686927d9c2a2e133edacb65459fe4416658ab97ebf37929c029" exitCode=0 Dec 01 16:41:59 crc kubenswrapper[4810]: I1201 16:41:59.678259 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" event={"ID":"2f0d06d5-7eea-4c8d-8700-439ea72af7cc","Type":"ContainerDied","Data":"d6091cb176357686927d9c2a2e133edacb65459fe4416658ab97ebf37929c029"} Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.178705 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.311390 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6rrk\" (UniqueName: \"kubernetes.io/projected/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-kube-api-access-f6rrk\") pod \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\" (UID: \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\") " Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.311867 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-ssh-key\") pod \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\" (UID: \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\") " Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.312535 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-bootstrap-combined-ca-bundle\") pod \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\" (UID: \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\") " Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.312802 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-inventory\") pod \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\" (UID: \"2f0d06d5-7eea-4c8d-8700-439ea72af7cc\") " Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.320455 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-kube-api-access-f6rrk" (OuterVolumeSpecName: "kube-api-access-f6rrk") pod "2f0d06d5-7eea-4c8d-8700-439ea72af7cc" (UID: "2f0d06d5-7eea-4c8d-8700-439ea72af7cc"). InnerVolumeSpecName "kube-api-access-f6rrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.323065 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "2f0d06d5-7eea-4c8d-8700-439ea72af7cc" (UID: "2f0d06d5-7eea-4c8d-8700-439ea72af7cc"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.351953 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2f0d06d5-7eea-4c8d-8700-439ea72af7cc" (UID: "2f0d06d5-7eea-4c8d-8700-439ea72af7cc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.354743 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-inventory" (OuterVolumeSpecName: "inventory") pod "2f0d06d5-7eea-4c8d-8700-439ea72af7cc" (UID: "2f0d06d5-7eea-4c8d-8700-439ea72af7cc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.415094 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6rrk\" (UniqueName: \"kubernetes.io/projected/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-kube-api-access-f6rrk\") on node \"crc\" DevicePath \"\"" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.415126 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.415136 4810 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.415146 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f0d06d5-7eea-4c8d-8700-439ea72af7cc-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.703940 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" event={"ID":"2f0d06d5-7eea-4c8d-8700-439ea72af7cc","Type":"ContainerDied","Data":"3789e858b368dc85a19ded5a1d2538435e8dfee2690003d6098b4fc7f8cff262"} Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.704287 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3789e858b368dc85a19ded5a1d2538435e8dfee2690003d6098b4fc7f8cff262" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.703973 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-2fqfj" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.795423 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-ptllz"] Dec 01 16:42:01 crc kubenswrapper[4810]: E1201 16:42:01.795912 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94e9a64a-8058-4096-b35d-9ce576d0aea7" containerName="extract-utilities" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.795931 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="94e9a64a-8058-4096-b35d-9ce576d0aea7" containerName="extract-utilities" Dec 01 16:42:01 crc kubenswrapper[4810]: E1201 16:42:01.795948 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94e9a64a-8058-4096-b35d-9ce576d0aea7" containerName="extract-content" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.795957 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="94e9a64a-8058-4096-b35d-9ce576d0aea7" containerName="extract-content" Dec 01 16:42:01 crc kubenswrapper[4810]: E1201 16:42:01.795981 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94e9a64a-8058-4096-b35d-9ce576d0aea7" containerName="registry-server" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.795987 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="94e9a64a-8058-4096-b35d-9ce576d0aea7" containerName="registry-server" Dec 01 16:42:01 crc kubenswrapper[4810]: E1201 16:42:01.796004 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5c248e7-bfad-4689-9ba7-18efc91ed6e7" containerName="extract-utilities" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.796009 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5c248e7-bfad-4689-9ba7-18efc91ed6e7" containerName="extract-utilities" Dec 01 16:42:01 crc kubenswrapper[4810]: E1201 16:42:01.796021 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5c248e7-bfad-4689-9ba7-18efc91ed6e7" containerName="registry-server" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.796027 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5c248e7-bfad-4689-9ba7-18efc91ed6e7" containerName="registry-server" Dec 01 16:42:01 crc kubenswrapper[4810]: E1201 16:42:01.796049 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f0d06d5-7eea-4c8d-8700-439ea72af7cc" containerName="bootstrap-openstack-openstack-cell1" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.796056 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f0d06d5-7eea-4c8d-8700-439ea72af7cc" containerName="bootstrap-openstack-openstack-cell1" Dec 01 16:42:01 crc kubenswrapper[4810]: E1201 16:42:01.796068 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5c248e7-bfad-4689-9ba7-18efc91ed6e7" containerName="extract-content" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.796074 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5c248e7-bfad-4689-9ba7-18efc91ed6e7" containerName="extract-content" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.796280 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f0d06d5-7eea-4c8d-8700-439ea72af7cc" containerName="bootstrap-openstack-openstack-cell1" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.796302 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5c248e7-bfad-4689-9ba7-18efc91ed6e7" containerName="registry-server" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.796325 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="94e9a64a-8058-4096-b35d-9ce576d0aea7" containerName="registry-server" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.797037 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-ptllz" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.799402 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.799740 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.801181 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.801407 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.804038 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-ptllz"] Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.825864 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8ngk\" (UniqueName: \"kubernetes.io/projected/c1bcac83-4d6b-46bc-83bc-3b5328c1c021-kube-api-access-n8ngk\") pod \"download-cache-openstack-openstack-cell1-ptllz\" (UID: \"c1bcac83-4d6b-46bc-83bc-3b5328c1c021\") " pod="openstack/download-cache-openstack-openstack-cell1-ptllz" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.826003 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1bcac83-4d6b-46bc-83bc-3b5328c1c021-inventory\") pod \"download-cache-openstack-openstack-cell1-ptllz\" (UID: \"c1bcac83-4d6b-46bc-83bc-3b5328c1c021\") " pod="openstack/download-cache-openstack-openstack-cell1-ptllz" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.826149 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1bcac83-4d6b-46bc-83bc-3b5328c1c021-ssh-key\") pod \"download-cache-openstack-openstack-cell1-ptllz\" (UID: \"c1bcac83-4d6b-46bc-83bc-3b5328c1c021\") " pod="openstack/download-cache-openstack-openstack-cell1-ptllz" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.927531 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1bcac83-4d6b-46bc-83bc-3b5328c1c021-ssh-key\") pod \"download-cache-openstack-openstack-cell1-ptllz\" (UID: \"c1bcac83-4d6b-46bc-83bc-3b5328c1c021\") " pod="openstack/download-cache-openstack-openstack-cell1-ptllz" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.928173 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8ngk\" (UniqueName: \"kubernetes.io/projected/c1bcac83-4d6b-46bc-83bc-3b5328c1c021-kube-api-access-n8ngk\") pod \"download-cache-openstack-openstack-cell1-ptllz\" (UID: \"c1bcac83-4d6b-46bc-83bc-3b5328c1c021\") " pod="openstack/download-cache-openstack-openstack-cell1-ptllz" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.928299 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1bcac83-4d6b-46bc-83bc-3b5328c1c021-inventory\") pod \"download-cache-openstack-openstack-cell1-ptllz\" (UID: \"c1bcac83-4d6b-46bc-83bc-3b5328c1c021\") " pod="openstack/download-cache-openstack-openstack-cell1-ptllz" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.943387 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1bcac83-4d6b-46bc-83bc-3b5328c1c021-inventory\") pod \"download-cache-openstack-openstack-cell1-ptllz\" (UID: \"c1bcac83-4d6b-46bc-83bc-3b5328c1c021\") " pod="openstack/download-cache-openstack-openstack-cell1-ptllz" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.943797 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8ngk\" (UniqueName: \"kubernetes.io/projected/c1bcac83-4d6b-46bc-83bc-3b5328c1c021-kube-api-access-n8ngk\") pod \"download-cache-openstack-openstack-cell1-ptllz\" (UID: \"c1bcac83-4d6b-46bc-83bc-3b5328c1c021\") " pod="openstack/download-cache-openstack-openstack-cell1-ptllz" Dec 01 16:42:01 crc kubenswrapper[4810]: I1201 16:42:01.953517 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1bcac83-4d6b-46bc-83bc-3b5328c1c021-ssh-key\") pod \"download-cache-openstack-openstack-cell1-ptllz\" (UID: \"c1bcac83-4d6b-46bc-83bc-3b5328c1c021\") " pod="openstack/download-cache-openstack-openstack-cell1-ptllz" Dec 01 16:42:01 crc kubenswrapper[4810]: E1201 16:42:01.973201 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f0d06d5_7eea_4c8d_8700_439ea72af7cc.slice/crio-3789e858b368dc85a19ded5a1d2538435e8dfee2690003d6098b4fc7f8cff262\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f0d06d5_7eea_4c8d_8700_439ea72af7cc.slice\": RecentStats: unable to find data in memory cache]" Dec 01 16:42:02 crc kubenswrapper[4810]: I1201 16:42:02.181566 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-ptllz" Dec 01 16:42:02 crc kubenswrapper[4810]: I1201 16:42:02.763217 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-ptllz"] Dec 01 16:42:02 crc kubenswrapper[4810]: I1201 16:42:02.775592 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:42:03 crc kubenswrapper[4810]: I1201 16:42:03.723154 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-ptllz" event={"ID":"c1bcac83-4d6b-46bc-83bc-3b5328c1c021","Type":"ContainerStarted","Data":"9c78b1a9ab0c16bdafb99a058d081b177c4de180e77dcbe29795568fce10da9e"} Dec 01 16:42:03 crc kubenswrapper[4810]: I1201 16:42:03.723505 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-ptllz" event={"ID":"c1bcac83-4d6b-46bc-83bc-3b5328c1c021","Type":"ContainerStarted","Data":"a7525cac9ee14a4f3c05cbda89a835a96a5ca059d5c6b4e8910f498f4e4619cf"} Dec 01 16:42:03 crc kubenswrapper[4810]: I1201 16:42:03.745596 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-ptllz" podStartSLOduration=2.210506345 podStartE2EDuration="2.745575208s" podCreationTimestamp="2025-12-01 16:42:01 +0000 UTC" firstStartedPulling="2025-12-01 16:42:02.775090516 +0000 UTC m=+7688.538600139" lastFinishedPulling="2025-12-01 16:42:03.310159399 +0000 UTC m=+7689.073669002" observedRunningTime="2025-12-01 16:42:03.741256561 +0000 UTC m=+7689.504766174" watchObservedRunningTime="2025-12-01 16:42:03.745575208 +0000 UTC m=+7689.509084811" Dec 01 16:42:05 crc kubenswrapper[4810]: I1201 16:42:05.491251 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:42:05 crc kubenswrapper[4810]: E1201 16:42:05.491777 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:42:17 crc kubenswrapper[4810]: I1201 16:42:17.490942 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:42:17 crc kubenswrapper[4810]: E1201 16:42:17.493295 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:42:30 crc kubenswrapper[4810]: I1201 16:42:30.491667 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:42:30 crc kubenswrapper[4810]: E1201 16:42:30.492844 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:42:41 crc kubenswrapper[4810]: I1201 16:42:41.491654 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:42:41 crc kubenswrapper[4810]: E1201 16:42:41.492607 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:42:53 crc kubenswrapper[4810]: I1201 16:42:53.494582 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:42:53 crc kubenswrapper[4810]: E1201 16:42:53.497894 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:43:04 crc kubenswrapper[4810]: I1201 16:43:04.501169 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:43:05 crc kubenswrapper[4810]: I1201 16:43:05.329391 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"f65292529d9bba3480ced8eb81902012c9eb05b1dba8694d8ec12fc39a3e4b89"} Dec 01 16:43:34 crc kubenswrapper[4810]: I1201 16:43:34.613915 4810 generic.go:334] "Generic (PLEG): container finished" podID="c1bcac83-4d6b-46bc-83bc-3b5328c1c021" containerID="9c78b1a9ab0c16bdafb99a058d081b177c4de180e77dcbe29795568fce10da9e" exitCode=0 Dec 01 16:43:34 crc kubenswrapper[4810]: I1201 16:43:34.614048 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-ptllz" event={"ID":"c1bcac83-4d6b-46bc-83bc-3b5328c1c021","Type":"ContainerDied","Data":"9c78b1a9ab0c16bdafb99a058d081b177c4de180e77dcbe29795568fce10da9e"} Dec 01 16:43:34 crc kubenswrapper[4810]: E1201 16:43:34.697825 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1bcac83_4d6b_46bc_83bc_3b5328c1c021.slice/crio-conmon-9c78b1a9ab0c16bdafb99a058d081b177c4de180e77dcbe29795568fce10da9e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1bcac83_4d6b_46bc_83bc_3b5328c1c021.slice/crio-9c78b1a9ab0c16bdafb99a058d081b177c4de180e77dcbe29795568fce10da9e.scope\": RecentStats: unable to find data in memory cache]" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.056832 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-ptllz" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.195762 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1bcac83-4d6b-46bc-83bc-3b5328c1c021-ssh-key\") pod \"c1bcac83-4d6b-46bc-83bc-3b5328c1c021\" (UID: \"c1bcac83-4d6b-46bc-83bc-3b5328c1c021\") " Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.196023 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8ngk\" (UniqueName: \"kubernetes.io/projected/c1bcac83-4d6b-46bc-83bc-3b5328c1c021-kube-api-access-n8ngk\") pod \"c1bcac83-4d6b-46bc-83bc-3b5328c1c021\" (UID: \"c1bcac83-4d6b-46bc-83bc-3b5328c1c021\") " Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.196138 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1bcac83-4d6b-46bc-83bc-3b5328c1c021-inventory\") pod \"c1bcac83-4d6b-46bc-83bc-3b5328c1c021\" (UID: \"c1bcac83-4d6b-46bc-83bc-3b5328c1c021\") " Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.208769 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1bcac83-4d6b-46bc-83bc-3b5328c1c021-kube-api-access-n8ngk" (OuterVolumeSpecName: "kube-api-access-n8ngk") pod "c1bcac83-4d6b-46bc-83bc-3b5328c1c021" (UID: "c1bcac83-4d6b-46bc-83bc-3b5328c1c021"). InnerVolumeSpecName "kube-api-access-n8ngk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.228446 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1bcac83-4d6b-46bc-83bc-3b5328c1c021-inventory" (OuterVolumeSpecName: "inventory") pod "c1bcac83-4d6b-46bc-83bc-3b5328c1c021" (UID: "c1bcac83-4d6b-46bc-83bc-3b5328c1c021"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.228787 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1bcac83-4d6b-46bc-83bc-3b5328c1c021-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c1bcac83-4d6b-46bc-83bc-3b5328c1c021" (UID: "c1bcac83-4d6b-46bc-83bc-3b5328c1c021"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.300562 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1bcac83-4d6b-46bc-83bc-3b5328c1c021-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.300598 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1bcac83-4d6b-46bc-83bc-3b5328c1c021-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.300612 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8ngk\" (UniqueName: \"kubernetes.io/projected/c1bcac83-4d6b-46bc-83bc-3b5328c1c021-kube-api-access-n8ngk\") on node \"crc\" DevicePath \"\"" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.638015 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-ptllz" event={"ID":"c1bcac83-4d6b-46bc-83bc-3b5328c1c021","Type":"ContainerDied","Data":"a7525cac9ee14a4f3c05cbda89a835a96a5ca059d5c6b4e8910f498f4e4619cf"} Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.638511 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7525cac9ee14a4f3c05cbda89a835a96a5ca059d5c6b4e8910f498f4e4619cf" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.638138 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-ptllz" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.743101 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-kk9r8"] Dec 01 16:43:36 crc kubenswrapper[4810]: E1201 16:43:36.743560 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1bcac83-4d6b-46bc-83bc-3b5328c1c021" containerName="download-cache-openstack-openstack-cell1" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.743579 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1bcac83-4d6b-46bc-83bc-3b5328c1c021" containerName="download-cache-openstack-openstack-cell1" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.743811 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1bcac83-4d6b-46bc-83bc-3b5328c1c021" containerName="download-cache-openstack-openstack-cell1" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.745856 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-kk9r8" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.752712 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.752983 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.754677 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-kk9r8"] Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.756662 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.757012 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.915604 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e050d22d-6f4f-449a-99dc-8dbf9b5603a3-inventory\") pod \"configure-network-openstack-openstack-cell1-kk9r8\" (UID: \"e050d22d-6f4f-449a-99dc-8dbf9b5603a3\") " pod="openstack/configure-network-openstack-openstack-cell1-kk9r8" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.915815 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e050d22d-6f4f-449a-99dc-8dbf9b5603a3-ssh-key\") pod \"configure-network-openstack-openstack-cell1-kk9r8\" (UID: \"e050d22d-6f4f-449a-99dc-8dbf9b5603a3\") " pod="openstack/configure-network-openstack-openstack-cell1-kk9r8" Dec 01 16:43:36 crc kubenswrapper[4810]: I1201 16:43:36.915869 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5ns9\" (UniqueName: \"kubernetes.io/projected/e050d22d-6f4f-449a-99dc-8dbf9b5603a3-kube-api-access-c5ns9\") pod \"configure-network-openstack-openstack-cell1-kk9r8\" (UID: \"e050d22d-6f4f-449a-99dc-8dbf9b5603a3\") " pod="openstack/configure-network-openstack-openstack-cell1-kk9r8" Dec 01 16:43:37 crc kubenswrapper[4810]: I1201 16:43:37.017485 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e050d22d-6f4f-449a-99dc-8dbf9b5603a3-inventory\") pod \"configure-network-openstack-openstack-cell1-kk9r8\" (UID: \"e050d22d-6f4f-449a-99dc-8dbf9b5603a3\") " pod="openstack/configure-network-openstack-openstack-cell1-kk9r8" Dec 01 16:43:37 crc kubenswrapper[4810]: I1201 16:43:37.017612 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e050d22d-6f4f-449a-99dc-8dbf9b5603a3-ssh-key\") pod \"configure-network-openstack-openstack-cell1-kk9r8\" (UID: \"e050d22d-6f4f-449a-99dc-8dbf9b5603a3\") " pod="openstack/configure-network-openstack-openstack-cell1-kk9r8" Dec 01 16:43:37 crc kubenswrapper[4810]: I1201 16:43:37.017642 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5ns9\" (UniqueName: \"kubernetes.io/projected/e050d22d-6f4f-449a-99dc-8dbf9b5603a3-kube-api-access-c5ns9\") pod \"configure-network-openstack-openstack-cell1-kk9r8\" (UID: \"e050d22d-6f4f-449a-99dc-8dbf9b5603a3\") " pod="openstack/configure-network-openstack-openstack-cell1-kk9r8" Dec 01 16:43:37 crc kubenswrapper[4810]: I1201 16:43:37.023133 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e050d22d-6f4f-449a-99dc-8dbf9b5603a3-ssh-key\") pod \"configure-network-openstack-openstack-cell1-kk9r8\" (UID: \"e050d22d-6f4f-449a-99dc-8dbf9b5603a3\") " pod="openstack/configure-network-openstack-openstack-cell1-kk9r8" Dec 01 16:43:37 crc kubenswrapper[4810]: I1201 16:43:37.025707 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e050d22d-6f4f-449a-99dc-8dbf9b5603a3-inventory\") pod \"configure-network-openstack-openstack-cell1-kk9r8\" (UID: \"e050d22d-6f4f-449a-99dc-8dbf9b5603a3\") " pod="openstack/configure-network-openstack-openstack-cell1-kk9r8" Dec 01 16:43:37 crc kubenswrapper[4810]: I1201 16:43:37.048192 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5ns9\" (UniqueName: \"kubernetes.io/projected/e050d22d-6f4f-449a-99dc-8dbf9b5603a3-kube-api-access-c5ns9\") pod \"configure-network-openstack-openstack-cell1-kk9r8\" (UID: \"e050d22d-6f4f-449a-99dc-8dbf9b5603a3\") " pod="openstack/configure-network-openstack-openstack-cell1-kk9r8" Dec 01 16:43:37 crc kubenswrapper[4810]: I1201 16:43:37.066313 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-kk9r8" Dec 01 16:43:37 crc kubenswrapper[4810]: I1201 16:43:37.664814 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-kk9r8"] Dec 01 16:43:38 crc kubenswrapper[4810]: I1201 16:43:38.659549 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-kk9r8" event={"ID":"e050d22d-6f4f-449a-99dc-8dbf9b5603a3","Type":"ContainerStarted","Data":"6e4b776ea795a63b73766ab27057aeeec4f263efaefd83954fc87e5e06a845bd"} Dec 01 16:43:38 crc kubenswrapper[4810]: I1201 16:43:38.660115 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-kk9r8" event={"ID":"e050d22d-6f4f-449a-99dc-8dbf9b5603a3","Type":"ContainerStarted","Data":"733fba831881637e74e4b86eba0ef7668da25958a442ff76ffb1786ff9529d73"} Dec 01 16:43:38 crc kubenswrapper[4810]: I1201 16:43:38.685451 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-kk9r8" podStartSLOduration=2.003820392 podStartE2EDuration="2.68543019s" podCreationTimestamp="2025-12-01 16:43:36 +0000 UTC" firstStartedPulling="2025-12-01 16:43:37.663216055 +0000 UTC m=+7783.426725658" lastFinishedPulling="2025-12-01 16:43:38.344825853 +0000 UTC m=+7784.108335456" observedRunningTime="2025-12-01 16:43:38.679848419 +0000 UTC m=+7784.443358032" watchObservedRunningTime="2025-12-01 16:43:38.68543019 +0000 UTC m=+7784.448939813" Dec 01 16:44:58 crc kubenswrapper[4810]: I1201 16:44:58.452319 4810 generic.go:334] "Generic (PLEG): container finished" podID="e050d22d-6f4f-449a-99dc-8dbf9b5603a3" containerID="6e4b776ea795a63b73766ab27057aeeec4f263efaefd83954fc87e5e06a845bd" exitCode=0 Dec 01 16:44:58 crc kubenswrapper[4810]: I1201 16:44:58.452448 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-kk9r8" event={"ID":"e050d22d-6f4f-449a-99dc-8dbf9b5603a3","Type":"ContainerDied","Data":"6e4b776ea795a63b73766ab27057aeeec4f263efaefd83954fc87e5e06a845bd"} Dec 01 16:44:59 crc kubenswrapper[4810]: I1201 16:44:59.986323 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-kk9r8" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.161256 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf"] Dec 01 16:45:00 crc kubenswrapper[4810]: E1201 16:45:00.161894 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e050d22d-6f4f-449a-99dc-8dbf9b5603a3" containerName="configure-network-openstack-openstack-cell1" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.161914 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e050d22d-6f4f-449a-99dc-8dbf9b5603a3" containerName="configure-network-openstack-openstack-cell1" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.162234 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e050d22d-6f4f-449a-99dc-8dbf9b5603a3" containerName="configure-network-openstack-openstack-cell1" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.164209 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.167307 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.171133 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.177369 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e050d22d-6f4f-449a-99dc-8dbf9b5603a3-ssh-key\") pod \"e050d22d-6f4f-449a-99dc-8dbf9b5603a3\" (UID: \"e050d22d-6f4f-449a-99dc-8dbf9b5603a3\") " Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.177449 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf"] Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.177667 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5ns9\" (UniqueName: \"kubernetes.io/projected/e050d22d-6f4f-449a-99dc-8dbf9b5603a3-kube-api-access-c5ns9\") pod \"e050d22d-6f4f-449a-99dc-8dbf9b5603a3\" (UID: \"e050d22d-6f4f-449a-99dc-8dbf9b5603a3\") " Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.177971 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e050d22d-6f4f-449a-99dc-8dbf9b5603a3-inventory\") pod \"e050d22d-6f4f-449a-99dc-8dbf9b5603a3\" (UID: \"e050d22d-6f4f-449a-99dc-8dbf9b5603a3\") " Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.196420 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e050d22d-6f4f-449a-99dc-8dbf9b5603a3-kube-api-access-c5ns9" (OuterVolumeSpecName: "kube-api-access-c5ns9") pod "e050d22d-6f4f-449a-99dc-8dbf9b5603a3" (UID: "e050d22d-6f4f-449a-99dc-8dbf9b5603a3"). InnerVolumeSpecName "kube-api-access-c5ns9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.222529 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e050d22d-6f4f-449a-99dc-8dbf9b5603a3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e050d22d-6f4f-449a-99dc-8dbf9b5603a3" (UID: "e050d22d-6f4f-449a-99dc-8dbf9b5603a3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.232179 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e050d22d-6f4f-449a-99dc-8dbf9b5603a3-inventory" (OuterVolumeSpecName: "inventory") pod "e050d22d-6f4f-449a-99dc-8dbf9b5603a3" (UID: "e050d22d-6f4f-449a-99dc-8dbf9b5603a3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.282237 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/00c1105f-2db1-4336-8452-93682dd15d0e-config-volume\") pod \"collect-profiles-29410125-gwhhf\" (UID: \"00c1105f-2db1-4336-8452-93682dd15d0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.282374 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/00c1105f-2db1-4336-8452-93682dd15d0e-secret-volume\") pod \"collect-profiles-29410125-gwhhf\" (UID: \"00c1105f-2db1-4336-8452-93682dd15d0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.282455 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcnm6\" (UniqueName: \"kubernetes.io/projected/00c1105f-2db1-4336-8452-93682dd15d0e-kube-api-access-bcnm6\") pod \"collect-profiles-29410125-gwhhf\" (UID: \"00c1105f-2db1-4336-8452-93682dd15d0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.283232 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5ns9\" (UniqueName: \"kubernetes.io/projected/e050d22d-6f4f-449a-99dc-8dbf9b5603a3-kube-api-access-c5ns9\") on node \"crc\" DevicePath \"\"" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.283293 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e050d22d-6f4f-449a-99dc-8dbf9b5603a3-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.283307 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e050d22d-6f4f-449a-99dc-8dbf9b5603a3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.384934 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/00c1105f-2db1-4336-8452-93682dd15d0e-secret-volume\") pod \"collect-profiles-29410125-gwhhf\" (UID: \"00c1105f-2db1-4336-8452-93682dd15d0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.385053 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcnm6\" (UniqueName: \"kubernetes.io/projected/00c1105f-2db1-4336-8452-93682dd15d0e-kube-api-access-bcnm6\") pod \"collect-profiles-29410125-gwhhf\" (UID: \"00c1105f-2db1-4336-8452-93682dd15d0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.385109 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/00c1105f-2db1-4336-8452-93682dd15d0e-config-volume\") pod \"collect-profiles-29410125-gwhhf\" (UID: \"00c1105f-2db1-4336-8452-93682dd15d0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.386114 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/00c1105f-2db1-4336-8452-93682dd15d0e-config-volume\") pod \"collect-profiles-29410125-gwhhf\" (UID: \"00c1105f-2db1-4336-8452-93682dd15d0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.389356 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/00c1105f-2db1-4336-8452-93682dd15d0e-secret-volume\") pod \"collect-profiles-29410125-gwhhf\" (UID: \"00c1105f-2db1-4336-8452-93682dd15d0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.401337 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcnm6\" (UniqueName: \"kubernetes.io/projected/00c1105f-2db1-4336-8452-93682dd15d0e-kube-api-access-bcnm6\") pod \"collect-profiles-29410125-gwhhf\" (UID: \"00c1105f-2db1-4336-8452-93682dd15d0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.494146 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-kk9r8" event={"ID":"e050d22d-6f4f-449a-99dc-8dbf9b5603a3","Type":"ContainerDied","Data":"733fba831881637e74e4b86eba0ef7668da25958a442ff76ffb1786ff9529d73"} Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.494419 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="733fba831881637e74e4b86eba0ef7668da25958a442ff76ffb1786ff9529d73" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.494513 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-kk9r8" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.498353 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.556696 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-fr5h2"] Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.558229 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-fr5h2" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.560940 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.564448 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.564676 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.565113 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.566309 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-fr5h2"] Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.590606 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68a329d7-d28e-40ea-bf13-8b3209932027-ssh-key\") pod \"validate-network-openstack-openstack-cell1-fr5h2\" (UID: \"68a329d7-d28e-40ea-bf13-8b3209932027\") " pod="openstack/validate-network-openstack-openstack-cell1-fr5h2" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.590967 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pw9m\" (UniqueName: \"kubernetes.io/projected/68a329d7-d28e-40ea-bf13-8b3209932027-kube-api-access-4pw9m\") pod \"validate-network-openstack-openstack-cell1-fr5h2\" (UID: \"68a329d7-d28e-40ea-bf13-8b3209932027\") " pod="openstack/validate-network-openstack-openstack-cell1-fr5h2" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.591261 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68a329d7-d28e-40ea-bf13-8b3209932027-inventory\") pod \"validate-network-openstack-openstack-cell1-fr5h2\" (UID: \"68a329d7-d28e-40ea-bf13-8b3209932027\") " pod="openstack/validate-network-openstack-openstack-cell1-fr5h2" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.696428 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pw9m\" (UniqueName: \"kubernetes.io/projected/68a329d7-d28e-40ea-bf13-8b3209932027-kube-api-access-4pw9m\") pod \"validate-network-openstack-openstack-cell1-fr5h2\" (UID: \"68a329d7-d28e-40ea-bf13-8b3209932027\") " pod="openstack/validate-network-openstack-openstack-cell1-fr5h2" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.707880 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68a329d7-d28e-40ea-bf13-8b3209932027-inventory\") pod \"validate-network-openstack-openstack-cell1-fr5h2\" (UID: \"68a329d7-d28e-40ea-bf13-8b3209932027\") " pod="openstack/validate-network-openstack-openstack-cell1-fr5h2" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.708143 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68a329d7-d28e-40ea-bf13-8b3209932027-ssh-key\") pod \"validate-network-openstack-openstack-cell1-fr5h2\" (UID: \"68a329d7-d28e-40ea-bf13-8b3209932027\") " pod="openstack/validate-network-openstack-openstack-cell1-fr5h2" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.725225 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68a329d7-d28e-40ea-bf13-8b3209932027-ssh-key\") pod \"validate-network-openstack-openstack-cell1-fr5h2\" (UID: \"68a329d7-d28e-40ea-bf13-8b3209932027\") " pod="openstack/validate-network-openstack-openstack-cell1-fr5h2" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.728586 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68a329d7-d28e-40ea-bf13-8b3209932027-inventory\") pod \"validate-network-openstack-openstack-cell1-fr5h2\" (UID: \"68a329d7-d28e-40ea-bf13-8b3209932027\") " pod="openstack/validate-network-openstack-openstack-cell1-fr5h2" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.734205 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pw9m\" (UniqueName: \"kubernetes.io/projected/68a329d7-d28e-40ea-bf13-8b3209932027-kube-api-access-4pw9m\") pod \"validate-network-openstack-openstack-cell1-fr5h2\" (UID: \"68a329d7-d28e-40ea-bf13-8b3209932027\") " pod="openstack/validate-network-openstack-openstack-cell1-fr5h2" Dec 01 16:45:00 crc kubenswrapper[4810]: I1201 16:45:00.937246 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-fr5h2" Dec 01 16:45:01 crc kubenswrapper[4810]: W1201 16:45:01.136561 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00c1105f_2db1_4336_8452_93682dd15d0e.slice/crio-1cd6b444632e02c2ba40f414b6601195119dc55bcd200bc8ac059f21024ab51e WatchSource:0}: Error finding container 1cd6b444632e02c2ba40f414b6601195119dc55bcd200bc8ac059f21024ab51e: Status 404 returned error can't find the container with id 1cd6b444632e02c2ba40f414b6601195119dc55bcd200bc8ac059f21024ab51e Dec 01 16:45:01 crc kubenswrapper[4810]: I1201 16:45:01.139405 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf"] Dec 01 16:45:01 crc kubenswrapper[4810]: I1201 16:45:01.535667 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf" event={"ID":"00c1105f-2db1-4336-8452-93682dd15d0e","Type":"ContainerStarted","Data":"7d28af347c9a678c5144d00b33c1dcd32ba3e0e5adf63dd1c92efbb4bc135c8f"} Dec 01 16:45:01 crc kubenswrapper[4810]: I1201 16:45:01.535939 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf" event={"ID":"00c1105f-2db1-4336-8452-93682dd15d0e","Type":"ContainerStarted","Data":"1cd6b444632e02c2ba40f414b6601195119dc55bcd200bc8ac059f21024ab51e"} Dec 01 16:45:01 crc kubenswrapper[4810]: I1201 16:45:01.544621 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-fr5h2"] Dec 01 16:45:01 crc kubenswrapper[4810]: I1201 16:45:01.557827 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf" podStartSLOduration=1.557807782 podStartE2EDuration="1.557807782s" podCreationTimestamp="2025-12-01 16:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 16:45:01.555130499 +0000 UTC m=+7867.318640102" watchObservedRunningTime="2025-12-01 16:45:01.557807782 +0000 UTC m=+7867.321317385" Dec 01 16:45:02 crc kubenswrapper[4810]: I1201 16:45:02.555236 4810 generic.go:334] "Generic (PLEG): container finished" podID="00c1105f-2db1-4336-8452-93682dd15d0e" containerID="7d28af347c9a678c5144d00b33c1dcd32ba3e0e5adf63dd1c92efbb4bc135c8f" exitCode=0 Dec 01 16:45:02 crc kubenswrapper[4810]: I1201 16:45:02.556327 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf" event={"ID":"00c1105f-2db1-4336-8452-93682dd15d0e","Type":"ContainerDied","Data":"7d28af347c9a678c5144d00b33c1dcd32ba3e0e5adf63dd1c92efbb4bc135c8f"} Dec 01 16:45:02 crc kubenswrapper[4810]: I1201 16:45:02.557984 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-fr5h2" event={"ID":"68a329d7-d28e-40ea-bf13-8b3209932027","Type":"ContainerStarted","Data":"163fd482a6fb32319f2e8b08bc9b200b1f00c034b81ba8073b6fdf11a59f9312"} Dec 01 16:45:03 crc kubenswrapper[4810]: I1201 16:45:03.576752 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-fr5h2" event={"ID":"68a329d7-d28e-40ea-bf13-8b3209932027","Type":"ContainerStarted","Data":"79b58d8ab2d9fc57ba5f02fea28d6012a3d695b5f0122c9925f704c40dc19944"} Dec 01 16:45:03 crc kubenswrapper[4810]: I1201 16:45:03.601989 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-fr5h2" podStartSLOduration=2.819181673 podStartE2EDuration="3.601937474s" podCreationTimestamp="2025-12-01 16:45:00 +0000 UTC" firstStartedPulling="2025-12-01 16:45:01.557971526 +0000 UTC m=+7867.321481129" lastFinishedPulling="2025-12-01 16:45:02.340727327 +0000 UTC m=+7868.104236930" observedRunningTime="2025-12-01 16:45:03.594972465 +0000 UTC m=+7869.358482068" watchObservedRunningTime="2025-12-01 16:45:03.601937474 +0000 UTC m=+7869.365447077" Dec 01 16:45:03 crc kubenswrapper[4810]: I1201 16:45:03.948068 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf" Dec 01 16:45:03 crc kubenswrapper[4810]: I1201 16:45:03.992071 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcnm6\" (UniqueName: \"kubernetes.io/projected/00c1105f-2db1-4336-8452-93682dd15d0e-kube-api-access-bcnm6\") pod \"00c1105f-2db1-4336-8452-93682dd15d0e\" (UID: \"00c1105f-2db1-4336-8452-93682dd15d0e\") " Dec 01 16:45:03 crc kubenswrapper[4810]: I1201 16:45:03.992129 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/00c1105f-2db1-4336-8452-93682dd15d0e-secret-volume\") pod \"00c1105f-2db1-4336-8452-93682dd15d0e\" (UID: \"00c1105f-2db1-4336-8452-93682dd15d0e\") " Dec 01 16:45:03 crc kubenswrapper[4810]: I1201 16:45:03.992212 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/00c1105f-2db1-4336-8452-93682dd15d0e-config-volume\") pod \"00c1105f-2db1-4336-8452-93682dd15d0e\" (UID: \"00c1105f-2db1-4336-8452-93682dd15d0e\") " Dec 01 16:45:03 crc kubenswrapper[4810]: I1201 16:45:03.993159 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00c1105f-2db1-4336-8452-93682dd15d0e-config-volume" (OuterVolumeSpecName: "config-volume") pod "00c1105f-2db1-4336-8452-93682dd15d0e" (UID: "00c1105f-2db1-4336-8452-93682dd15d0e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:45:04 crc kubenswrapper[4810]: I1201 16:45:04.010133 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00c1105f-2db1-4336-8452-93682dd15d0e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "00c1105f-2db1-4336-8452-93682dd15d0e" (UID: "00c1105f-2db1-4336-8452-93682dd15d0e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:45:04 crc kubenswrapper[4810]: I1201 16:45:04.010175 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00c1105f-2db1-4336-8452-93682dd15d0e-kube-api-access-bcnm6" (OuterVolumeSpecName: "kube-api-access-bcnm6") pod "00c1105f-2db1-4336-8452-93682dd15d0e" (UID: "00c1105f-2db1-4336-8452-93682dd15d0e"). InnerVolumeSpecName "kube-api-access-bcnm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:45:04 crc kubenswrapper[4810]: I1201 16:45:04.093633 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcnm6\" (UniqueName: \"kubernetes.io/projected/00c1105f-2db1-4336-8452-93682dd15d0e-kube-api-access-bcnm6\") on node \"crc\" DevicePath \"\"" Dec 01 16:45:04 crc kubenswrapper[4810]: I1201 16:45:04.093659 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/00c1105f-2db1-4336-8452-93682dd15d0e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 16:45:04 crc kubenswrapper[4810]: I1201 16:45:04.093669 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/00c1105f-2db1-4336-8452-93682dd15d0e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 16:45:04 crc kubenswrapper[4810]: I1201 16:45:04.591486 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf" event={"ID":"00c1105f-2db1-4336-8452-93682dd15d0e","Type":"ContainerDied","Data":"1cd6b444632e02c2ba40f414b6601195119dc55bcd200bc8ac059f21024ab51e"} Dec 01 16:45:04 crc kubenswrapper[4810]: I1201 16:45:04.591554 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cd6b444632e02c2ba40f414b6601195119dc55bcd200bc8ac059f21024ab51e" Dec 01 16:45:04 crc kubenswrapper[4810]: I1201 16:45:04.591509 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf" Dec 01 16:45:04 crc kubenswrapper[4810]: I1201 16:45:04.623185 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz"] Dec 01 16:45:04 crc kubenswrapper[4810]: I1201 16:45:04.636003 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410080-4nllz"] Dec 01 16:45:06 crc kubenswrapper[4810]: I1201 16:45:06.503555 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="767af1af-2227-4c98-9b5d-24f8e64a7d6d" path="/var/lib/kubelet/pods/767af1af-2227-4c98-9b5d-24f8e64a7d6d/volumes" Dec 01 16:45:07 crc kubenswrapper[4810]: I1201 16:45:07.626251 4810 generic.go:334] "Generic (PLEG): container finished" podID="68a329d7-d28e-40ea-bf13-8b3209932027" containerID="79b58d8ab2d9fc57ba5f02fea28d6012a3d695b5f0122c9925f704c40dc19944" exitCode=0 Dec 01 16:45:07 crc kubenswrapper[4810]: I1201 16:45:07.626330 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-fr5h2" event={"ID":"68a329d7-d28e-40ea-bf13-8b3209932027","Type":"ContainerDied","Data":"79b58d8ab2d9fc57ba5f02fea28d6012a3d695b5f0122c9925f704c40dc19944"} Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.153440 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-fr5h2" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.202166 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pw9m\" (UniqueName: \"kubernetes.io/projected/68a329d7-d28e-40ea-bf13-8b3209932027-kube-api-access-4pw9m\") pod \"68a329d7-d28e-40ea-bf13-8b3209932027\" (UID: \"68a329d7-d28e-40ea-bf13-8b3209932027\") " Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.202598 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68a329d7-d28e-40ea-bf13-8b3209932027-inventory\") pod \"68a329d7-d28e-40ea-bf13-8b3209932027\" (UID: \"68a329d7-d28e-40ea-bf13-8b3209932027\") " Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.202736 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68a329d7-d28e-40ea-bf13-8b3209932027-ssh-key\") pod \"68a329d7-d28e-40ea-bf13-8b3209932027\" (UID: \"68a329d7-d28e-40ea-bf13-8b3209932027\") " Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.211318 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68a329d7-d28e-40ea-bf13-8b3209932027-kube-api-access-4pw9m" (OuterVolumeSpecName: "kube-api-access-4pw9m") pod "68a329d7-d28e-40ea-bf13-8b3209932027" (UID: "68a329d7-d28e-40ea-bf13-8b3209932027"). InnerVolumeSpecName "kube-api-access-4pw9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.241701 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68a329d7-d28e-40ea-bf13-8b3209932027-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "68a329d7-d28e-40ea-bf13-8b3209932027" (UID: "68a329d7-d28e-40ea-bf13-8b3209932027"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.244233 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68a329d7-d28e-40ea-bf13-8b3209932027-inventory" (OuterVolumeSpecName: "inventory") pod "68a329d7-d28e-40ea-bf13-8b3209932027" (UID: "68a329d7-d28e-40ea-bf13-8b3209932027"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.305391 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68a329d7-d28e-40ea-bf13-8b3209932027-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.305428 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68a329d7-d28e-40ea-bf13-8b3209932027-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.305441 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pw9m\" (UniqueName: \"kubernetes.io/projected/68a329d7-d28e-40ea-bf13-8b3209932027-kube-api-access-4pw9m\") on node \"crc\" DevicePath \"\"" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.649525 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-fr5h2" event={"ID":"68a329d7-d28e-40ea-bf13-8b3209932027","Type":"ContainerDied","Data":"163fd482a6fb32319f2e8b08bc9b200b1f00c034b81ba8073b6fdf11a59f9312"} Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.649573 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="163fd482a6fb32319f2e8b08bc9b200b1f00c034b81ba8073b6fdf11a59f9312" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.649752 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-fr5h2" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.736711 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-mfkj5"] Dec 01 16:45:09 crc kubenswrapper[4810]: E1201 16:45:09.737311 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00c1105f-2db1-4336-8452-93682dd15d0e" containerName="collect-profiles" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.737333 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="00c1105f-2db1-4336-8452-93682dd15d0e" containerName="collect-profiles" Dec 01 16:45:09 crc kubenswrapper[4810]: E1201 16:45:09.737385 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68a329d7-d28e-40ea-bf13-8b3209932027" containerName="validate-network-openstack-openstack-cell1" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.737396 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="68a329d7-d28e-40ea-bf13-8b3209932027" containerName="validate-network-openstack-openstack-cell1" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.737672 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="68a329d7-d28e-40ea-bf13-8b3209932027" containerName="validate-network-openstack-openstack-cell1" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.737725 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="00c1105f-2db1-4336-8452-93682dd15d0e" containerName="collect-profiles" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.738741 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-mfkj5" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.746415 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.746739 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.746825 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.746825 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.747428 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-mfkj5"] Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.820863 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84f52a8b-4f35-4a57-93c9-05090d31a876-ssh-key\") pod \"install-os-openstack-openstack-cell1-mfkj5\" (UID: \"84f52a8b-4f35-4a57-93c9-05090d31a876\") " pod="openstack/install-os-openstack-openstack-cell1-mfkj5" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.821146 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhr55\" (UniqueName: \"kubernetes.io/projected/84f52a8b-4f35-4a57-93c9-05090d31a876-kube-api-access-nhr55\") pod \"install-os-openstack-openstack-cell1-mfkj5\" (UID: \"84f52a8b-4f35-4a57-93c9-05090d31a876\") " pod="openstack/install-os-openstack-openstack-cell1-mfkj5" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.821201 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84f52a8b-4f35-4a57-93c9-05090d31a876-inventory\") pod \"install-os-openstack-openstack-cell1-mfkj5\" (UID: \"84f52a8b-4f35-4a57-93c9-05090d31a876\") " pod="openstack/install-os-openstack-openstack-cell1-mfkj5" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.924273 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhr55\" (UniqueName: \"kubernetes.io/projected/84f52a8b-4f35-4a57-93c9-05090d31a876-kube-api-access-nhr55\") pod \"install-os-openstack-openstack-cell1-mfkj5\" (UID: \"84f52a8b-4f35-4a57-93c9-05090d31a876\") " pod="openstack/install-os-openstack-openstack-cell1-mfkj5" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.924417 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84f52a8b-4f35-4a57-93c9-05090d31a876-inventory\") pod \"install-os-openstack-openstack-cell1-mfkj5\" (UID: \"84f52a8b-4f35-4a57-93c9-05090d31a876\") " pod="openstack/install-os-openstack-openstack-cell1-mfkj5" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.924570 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84f52a8b-4f35-4a57-93c9-05090d31a876-ssh-key\") pod \"install-os-openstack-openstack-cell1-mfkj5\" (UID: \"84f52a8b-4f35-4a57-93c9-05090d31a876\") " pod="openstack/install-os-openstack-openstack-cell1-mfkj5" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.930163 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84f52a8b-4f35-4a57-93c9-05090d31a876-ssh-key\") pod \"install-os-openstack-openstack-cell1-mfkj5\" (UID: \"84f52a8b-4f35-4a57-93c9-05090d31a876\") " pod="openstack/install-os-openstack-openstack-cell1-mfkj5" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.932554 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84f52a8b-4f35-4a57-93c9-05090d31a876-inventory\") pod \"install-os-openstack-openstack-cell1-mfkj5\" (UID: \"84f52a8b-4f35-4a57-93c9-05090d31a876\") " pod="openstack/install-os-openstack-openstack-cell1-mfkj5" Dec 01 16:45:09 crc kubenswrapper[4810]: I1201 16:45:09.950131 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhr55\" (UniqueName: \"kubernetes.io/projected/84f52a8b-4f35-4a57-93c9-05090d31a876-kube-api-access-nhr55\") pod \"install-os-openstack-openstack-cell1-mfkj5\" (UID: \"84f52a8b-4f35-4a57-93c9-05090d31a876\") " pod="openstack/install-os-openstack-openstack-cell1-mfkj5" Dec 01 16:45:10 crc kubenswrapper[4810]: I1201 16:45:10.064012 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-mfkj5" Dec 01 16:45:10 crc kubenswrapper[4810]: W1201 16:45:10.656275 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84f52a8b_4f35_4a57_93c9_05090d31a876.slice/crio-a92c493fd776229c2e949fdd1aaeab3d44733ea5b2b6232ee2732e0c91f18b7a WatchSource:0}: Error finding container a92c493fd776229c2e949fdd1aaeab3d44733ea5b2b6232ee2732e0c91f18b7a: Status 404 returned error can't find the container with id a92c493fd776229c2e949fdd1aaeab3d44733ea5b2b6232ee2732e0c91f18b7a Dec 01 16:45:10 crc kubenswrapper[4810]: I1201 16:45:10.660904 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-mfkj5"] Dec 01 16:45:11 crc kubenswrapper[4810]: I1201 16:45:11.670135 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-mfkj5" event={"ID":"84f52a8b-4f35-4a57-93c9-05090d31a876","Type":"ContainerStarted","Data":"87b39a785864033fa4389356ebf387966f66b9a9a6f539d51ad1136eb8914880"} Dec 01 16:45:11 crc kubenswrapper[4810]: I1201 16:45:11.670444 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-mfkj5" event={"ID":"84f52a8b-4f35-4a57-93c9-05090d31a876","Type":"ContainerStarted","Data":"a92c493fd776229c2e949fdd1aaeab3d44733ea5b2b6232ee2732e0c91f18b7a"} Dec 01 16:45:11 crc kubenswrapper[4810]: I1201 16:45:11.695692 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-mfkj5" podStartSLOduration=2.2917430100000002 podStartE2EDuration="2.695672706s" podCreationTimestamp="2025-12-01 16:45:09 +0000 UTC" firstStartedPulling="2025-12-01 16:45:10.658593108 +0000 UTC m=+7876.422102711" lastFinishedPulling="2025-12-01 16:45:11.062522804 +0000 UTC m=+7876.826032407" observedRunningTime="2025-12-01 16:45:11.690915666 +0000 UTC m=+7877.454425259" watchObservedRunningTime="2025-12-01 16:45:11.695672706 +0000 UTC m=+7877.459182309" Dec 01 16:45:17 crc kubenswrapper[4810]: I1201 16:45:17.320132 4810 scope.go:117] "RemoveContainer" containerID="fabd7de3f73e504c542541093725e23523892d07d1879c6b7a4f6160f29bcc59" Dec 01 16:45:32 crc kubenswrapper[4810]: I1201 16:45:32.972464 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:45:32 crc kubenswrapper[4810]: I1201 16:45:32.973041 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:45:55 crc kubenswrapper[4810]: I1201 16:45:55.229407 4810 generic.go:334] "Generic (PLEG): container finished" podID="84f52a8b-4f35-4a57-93c9-05090d31a876" containerID="87b39a785864033fa4389356ebf387966f66b9a9a6f539d51ad1136eb8914880" exitCode=0 Dec 01 16:45:55 crc kubenswrapper[4810]: I1201 16:45:55.229508 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-mfkj5" event={"ID":"84f52a8b-4f35-4a57-93c9-05090d31a876","Type":"ContainerDied","Data":"87b39a785864033fa4389356ebf387966f66b9a9a6f539d51ad1136eb8914880"} Dec 01 16:45:56 crc kubenswrapper[4810]: I1201 16:45:56.659606 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-mfkj5" Dec 01 16:45:56 crc kubenswrapper[4810]: I1201 16:45:56.669068 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84f52a8b-4f35-4a57-93c9-05090d31a876-inventory\") pod \"84f52a8b-4f35-4a57-93c9-05090d31a876\" (UID: \"84f52a8b-4f35-4a57-93c9-05090d31a876\") " Dec 01 16:45:56 crc kubenswrapper[4810]: I1201 16:45:56.669252 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhr55\" (UniqueName: \"kubernetes.io/projected/84f52a8b-4f35-4a57-93c9-05090d31a876-kube-api-access-nhr55\") pod \"84f52a8b-4f35-4a57-93c9-05090d31a876\" (UID: \"84f52a8b-4f35-4a57-93c9-05090d31a876\") " Dec 01 16:45:56 crc kubenswrapper[4810]: I1201 16:45:56.669379 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84f52a8b-4f35-4a57-93c9-05090d31a876-ssh-key\") pod \"84f52a8b-4f35-4a57-93c9-05090d31a876\" (UID: \"84f52a8b-4f35-4a57-93c9-05090d31a876\") " Dec 01 16:45:56 crc kubenswrapper[4810]: I1201 16:45:56.682123 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84f52a8b-4f35-4a57-93c9-05090d31a876-kube-api-access-nhr55" (OuterVolumeSpecName: "kube-api-access-nhr55") pod "84f52a8b-4f35-4a57-93c9-05090d31a876" (UID: "84f52a8b-4f35-4a57-93c9-05090d31a876"). InnerVolumeSpecName "kube-api-access-nhr55". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:45:56 crc kubenswrapper[4810]: I1201 16:45:56.706544 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84f52a8b-4f35-4a57-93c9-05090d31a876-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "84f52a8b-4f35-4a57-93c9-05090d31a876" (UID: "84f52a8b-4f35-4a57-93c9-05090d31a876"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:45:56 crc kubenswrapper[4810]: I1201 16:45:56.717311 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84f52a8b-4f35-4a57-93c9-05090d31a876-inventory" (OuterVolumeSpecName: "inventory") pod "84f52a8b-4f35-4a57-93c9-05090d31a876" (UID: "84f52a8b-4f35-4a57-93c9-05090d31a876"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:45:56 crc kubenswrapper[4810]: I1201 16:45:56.773621 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84f52a8b-4f35-4a57-93c9-05090d31a876-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:45:56 crc kubenswrapper[4810]: I1201 16:45:56.773677 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhr55\" (UniqueName: \"kubernetes.io/projected/84f52a8b-4f35-4a57-93c9-05090d31a876-kube-api-access-nhr55\") on node \"crc\" DevicePath \"\"" Dec 01 16:45:56 crc kubenswrapper[4810]: I1201 16:45:56.773699 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84f52a8b-4f35-4a57-93c9-05090d31a876-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.249192 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-mfkj5" event={"ID":"84f52a8b-4f35-4a57-93c9-05090d31a876","Type":"ContainerDied","Data":"a92c493fd776229c2e949fdd1aaeab3d44733ea5b2b6232ee2732e0c91f18b7a"} Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.249242 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-mfkj5" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.249251 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a92c493fd776229c2e949fdd1aaeab3d44733ea5b2b6232ee2732e0c91f18b7a" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.325851 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-5sbhr"] Dec 01 16:45:57 crc kubenswrapper[4810]: E1201 16:45:57.326328 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84f52a8b-4f35-4a57-93c9-05090d31a876" containerName="install-os-openstack-openstack-cell1" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.326348 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="84f52a8b-4f35-4a57-93c9-05090d31a876" containerName="install-os-openstack-openstack-cell1" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.326566 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="84f52a8b-4f35-4a57-93c9-05090d31a876" containerName="install-os-openstack-openstack-cell1" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.327372 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-5sbhr" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.329494 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.330132 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.330365 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.330663 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.336880 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-5sbhr"] Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.390384 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5d6z\" (UniqueName: \"kubernetes.io/projected/ba759ad7-3f5c-4d82-a406-54b887ede26a-kube-api-access-g5d6z\") pod \"configure-os-openstack-openstack-cell1-5sbhr\" (UID: \"ba759ad7-3f5c-4d82-a406-54b887ede26a\") " pod="openstack/configure-os-openstack-openstack-cell1-5sbhr" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.390679 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba759ad7-3f5c-4d82-a406-54b887ede26a-ssh-key\") pod \"configure-os-openstack-openstack-cell1-5sbhr\" (UID: \"ba759ad7-3f5c-4d82-a406-54b887ede26a\") " pod="openstack/configure-os-openstack-openstack-cell1-5sbhr" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.390736 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba759ad7-3f5c-4d82-a406-54b887ede26a-inventory\") pod \"configure-os-openstack-openstack-cell1-5sbhr\" (UID: \"ba759ad7-3f5c-4d82-a406-54b887ede26a\") " pod="openstack/configure-os-openstack-openstack-cell1-5sbhr" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.494892 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba759ad7-3f5c-4d82-a406-54b887ede26a-ssh-key\") pod \"configure-os-openstack-openstack-cell1-5sbhr\" (UID: \"ba759ad7-3f5c-4d82-a406-54b887ede26a\") " pod="openstack/configure-os-openstack-openstack-cell1-5sbhr" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.495250 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba759ad7-3f5c-4d82-a406-54b887ede26a-inventory\") pod \"configure-os-openstack-openstack-cell1-5sbhr\" (UID: \"ba759ad7-3f5c-4d82-a406-54b887ede26a\") " pod="openstack/configure-os-openstack-openstack-cell1-5sbhr" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.495518 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5d6z\" (UniqueName: \"kubernetes.io/projected/ba759ad7-3f5c-4d82-a406-54b887ede26a-kube-api-access-g5d6z\") pod \"configure-os-openstack-openstack-cell1-5sbhr\" (UID: \"ba759ad7-3f5c-4d82-a406-54b887ede26a\") " pod="openstack/configure-os-openstack-openstack-cell1-5sbhr" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.498493 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba759ad7-3f5c-4d82-a406-54b887ede26a-ssh-key\") pod \"configure-os-openstack-openstack-cell1-5sbhr\" (UID: \"ba759ad7-3f5c-4d82-a406-54b887ede26a\") " pod="openstack/configure-os-openstack-openstack-cell1-5sbhr" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.498874 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba759ad7-3f5c-4d82-a406-54b887ede26a-inventory\") pod \"configure-os-openstack-openstack-cell1-5sbhr\" (UID: \"ba759ad7-3f5c-4d82-a406-54b887ede26a\") " pod="openstack/configure-os-openstack-openstack-cell1-5sbhr" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.511128 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5d6z\" (UniqueName: \"kubernetes.io/projected/ba759ad7-3f5c-4d82-a406-54b887ede26a-kube-api-access-g5d6z\") pod \"configure-os-openstack-openstack-cell1-5sbhr\" (UID: \"ba759ad7-3f5c-4d82-a406-54b887ede26a\") " pod="openstack/configure-os-openstack-openstack-cell1-5sbhr" Dec 01 16:45:57 crc kubenswrapper[4810]: I1201 16:45:57.686362 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-5sbhr" Dec 01 16:45:58 crc kubenswrapper[4810]: I1201 16:45:58.224411 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-5sbhr"] Dec 01 16:45:58 crc kubenswrapper[4810]: W1201 16:45:58.228227 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba759ad7_3f5c_4d82_a406_54b887ede26a.slice/crio-3b98eddcc15a4be2a1f9839c18ab2b84b4958e7f677b596def05a3e4060bd4ae WatchSource:0}: Error finding container 3b98eddcc15a4be2a1f9839c18ab2b84b4958e7f677b596def05a3e4060bd4ae: Status 404 returned error can't find the container with id 3b98eddcc15a4be2a1f9839c18ab2b84b4958e7f677b596def05a3e4060bd4ae Dec 01 16:45:58 crc kubenswrapper[4810]: I1201 16:45:58.259262 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-5sbhr" event={"ID":"ba759ad7-3f5c-4d82-a406-54b887ede26a","Type":"ContainerStarted","Data":"3b98eddcc15a4be2a1f9839c18ab2b84b4958e7f677b596def05a3e4060bd4ae"} Dec 01 16:46:00 crc kubenswrapper[4810]: I1201 16:46:00.280967 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-5sbhr" event={"ID":"ba759ad7-3f5c-4d82-a406-54b887ede26a","Type":"ContainerStarted","Data":"c796c9b6212ffc7e1a82db08ea5c0ca0ad43dd5e1f08484e25eca185361c946b"} Dec 01 16:46:02 crc kubenswrapper[4810]: I1201 16:46:02.972556 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:46:02 crc kubenswrapper[4810]: I1201 16:46:02.972851 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:46:20 crc kubenswrapper[4810]: I1201 16:46:20.591838 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-5sbhr" podStartSLOduration=22.857083661 podStartE2EDuration="23.591821539s" podCreationTimestamp="2025-12-01 16:45:57 +0000 UTC" firstStartedPulling="2025-12-01 16:45:58.231179792 +0000 UTC m=+7923.994689395" lastFinishedPulling="2025-12-01 16:45:58.96591767 +0000 UTC m=+7924.729427273" observedRunningTime="2025-12-01 16:46:00.297518697 +0000 UTC m=+7926.061028310" watchObservedRunningTime="2025-12-01 16:46:20.591821539 +0000 UTC m=+7946.355331142" Dec 01 16:46:20 crc kubenswrapper[4810]: I1201 16:46:20.597917 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p58nl"] Dec 01 16:46:20 crc kubenswrapper[4810]: I1201 16:46:20.600604 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p58nl" Dec 01 16:46:20 crc kubenswrapper[4810]: I1201 16:46:20.624868 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p58nl"] Dec 01 16:46:20 crc kubenswrapper[4810]: I1201 16:46:20.681905 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af7c8a95-dc78-457a-a955-4a729b1425a6-utilities\") pod \"certified-operators-p58nl\" (UID: \"af7c8a95-dc78-457a-a955-4a729b1425a6\") " pod="openshift-marketplace/certified-operators-p58nl" Dec 01 16:46:20 crc kubenswrapper[4810]: I1201 16:46:20.681973 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg4dq\" (UniqueName: \"kubernetes.io/projected/af7c8a95-dc78-457a-a955-4a729b1425a6-kube-api-access-vg4dq\") pod \"certified-operators-p58nl\" (UID: \"af7c8a95-dc78-457a-a955-4a729b1425a6\") " pod="openshift-marketplace/certified-operators-p58nl" Dec 01 16:46:20 crc kubenswrapper[4810]: I1201 16:46:20.682149 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af7c8a95-dc78-457a-a955-4a729b1425a6-catalog-content\") pod \"certified-operators-p58nl\" (UID: \"af7c8a95-dc78-457a-a955-4a729b1425a6\") " pod="openshift-marketplace/certified-operators-p58nl" Dec 01 16:46:20 crc kubenswrapper[4810]: I1201 16:46:20.784373 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af7c8a95-dc78-457a-a955-4a729b1425a6-catalog-content\") pod \"certified-operators-p58nl\" (UID: \"af7c8a95-dc78-457a-a955-4a729b1425a6\") " pod="openshift-marketplace/certified-operators-p58nl" Dec 01 16:46:20 crc kubenswrapper[4810]: I1201 16:46:20.784426 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af7c8a95-dc78-457a-a955-4a729b1425a6-utilities\") pod \"certified-operators-p58nl\" (UID: \"af7c8a95-dc78-457a-a955-4a729b1425a6\") " pod="openshift-marketplace/certified-operators-p58nl" Dec 01 16:46:20 crc kubenswrapper[4810]: I1201 16:46:20.784496 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg4dq\" (UniqueName: \"kubernetes.io/projected/af7c8a95-dc78-457a-a955-4a729b1425a6-kube-api-access-vg4dq\") pod \"certified-operators-p58nl\" (UID: \"af7c8a95-dc78-457a-a955-4a729b1425a6\") " pod="openshift-marketplace/certified-operators-p58nl" Dec 01 16:46:20 crc kubenswrapper[4810]: I1201 16:46:20.785026 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af7c8a95-dc78-457a-a955-4a729b1425a6-utilities\") pod \"certified-operators-p58nl\" (UID: \"af7c8a95-dc78-457a-a955-4a729b1425a6\") " pod="openshift-marketplace/certified-operators-p58nl" Dec 01 16:46:20 crc kubenswrapper[4810]: I1201 16:46:20.785302 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af7c8a95-dc78-457a-a955-4a729b1425a6-catalog-content\") pod \"certified-operators-p58nl\" (UID: \"af7c8a95-dc78-457a-a955-4a729b1425a6\") " pod="openshift-marketplace/certified-operators-p58nl" Dec 01 16:46:20 crc kubenswrapper[4810]: I1201 16:46:20.815943 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg4dq\" (UniqueName: \"kubernetes.io/projected/af7c8a95-dc78-457a-a955-4a729b1425a6-kube-api-access-vg4dq\") pod \"certified-operators-p58nl\" (UID: \"af7c8a95-dc78-457a-a955-4a729b1425a6\") " pod="openshift-marketplace/certified-operators-p58nl" Dec 01 16:46:20 crc kubenswrapper[4810]: I1201 16:46:20.941891 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p58nl" Dec 01 16:46:21 crc kubenswrapper[4810]: I1201 16:46:21.488841 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p58nl"] Dec 01 16:46:21 crc kubenswrapper[4810]: W1201 16:46:21.497911 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf7c8a95_dc78_457a_a955_4a729b1425a6.slice/crio-d514cd000ec6adec9157a0abb1a020cc50d8330d185b18cf89396e4f781c432f WatchSource:0}: Error finding container d514cd000ec6adec9157a0abb1a020cc50d8330d185b18cf89396e4f781c432f: Status 404 returned error can't find the container with id d514cd000ec6adec9157a0abb1a020cc50d8330d185b18cf89396e4f781c432f Dec 01 16:46:22 crc kubenswrapper[4810]: I1201 16:46:22.507944 4810 generic.go:334] "Generic (PLEG): container finished" podID="af7c8a95-dc78-457a-a955-4a729b1425a6" containerID="8ef93578c7718b3d77500f82602bae59ac60c940b1742f2f01817d7f9fd754a8" exitCode=0 Dec 01 16:46:22 crc kubenswrapper[4810]: I1201 16:46:22.508018 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p58nl" event={"ID":"af7c8a95-dc78-457a-a955-4a729b1425a6","Type":"ContainerDied","Data":"8ef93578c7718b3d77500f82602bae59ac60c940b1742f2f01817d7f9fd754a8"} Dec 01 16:46:22 crc kubenswrapper[4810]: I1201 16:46:22.508261 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p58nl" event={"ID":"af7c8a95-dc78-457a-a955-4a729b1425a6","Type":"ContainerStarted","Data":"d514cd000ec6adec9157a0abb1a020cc50d8330d185b18cf89396e4f781c432f"} Dec 01 16:46:24 crc kubenswrapper[4810]: I1201 16:46:24.535116 4810 generic.go:334] "Generic (PLEG): container finished" podID="af7c8a95-dc78-457a-a955-4a729b1425a6" containerID="245cbd4e8bdad650eff4679b8235a8d432bab6ddd50a34c09d7ef8c928cd8b6d" exitCode=0 Dec 01 16:46:24 crc kubenswrapper[4810]: I1201 16:46:24.535249 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p58nl" event={"ID":"af7c8a95-dc78-457a-a955-4a729b1425a6","Type":"ContainerDied","Data":"245cbd4e8bdad650eff4679b8235a8d432bab6ddd50a34c09d7ef8c928cd8b6d"} Dec 01 16:46:26 crc kubenswrapper[4810]: I1201 16:46:26.558048 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p58nl" event={"ID":"af7c8a95-dc78-457a-a955-4a729b1425a6","Type":"ContainerStarted","Data":"2167c7bc0040fe653677893aee31cfdc07a4359777ad10ca97c79cb34c16d463"} Dec 01 16:46:26 crc kubenswrapper[4810]: I1201 16:46:26.589701 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p58nl" podStartSLOduration=3.293581298 podStartE2EDuration="6.589681416s" podCreationTimestamp="2025-12-01 16:46:20 +0000 UTC" firstStartedPulling="2025-12-01 16:46:22.509852551 +0000 UTC m=+7948.273362154" lastFinishedPulling="2025-12-01 16:46:25.805952669 +0000 UTC m=+7951.569462272" observedRunningTime="2025-12-01 16:46:26.574508435 +0000 UTC m=+7952.338018068" watchObservedRunningTime="2025-12-01 16:46:26.589681416 +0000 UTC m=+7952.353191019" Dec 01 16:46:30 crc kubenswrapper[4810]: I1201 16:46:30.943130 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p58nl" Dec 01 16:46:30 crc kubenswrapper[4810]: I1201 16:46:30.943763 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p58nl" Dec 01 16:46:31 crc kubenswrapper[4810]: I1201 16:46:31.002192 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p58nl" Dec 01 16:46:31 crc kubenswrapper[4810]: I1201 16:46:31.650648 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p58nl" Dec 01 16:46:31 crc kubenswrapper[4810]: I1201 16:46:31.702052 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p58nl"] Dec 01 16:46:32 crc kubenswrapper[4810]: I1201 16:46:32.972900 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:46:32 crc kubenswrapper[4810]: I1201 16:46:32.972991 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:46:32 crc kubenswrapper[4810]: I1201 16:46:32.973063 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 16:46:32 crc kubenswrapper[4810]: I1201 16:46:32.974347 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f65292529d9bba3480ced8eb81902012c9eb05b1dba8694d8ec12fc39a3e4b89"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:46:32 crc kubenswrapper[4810]: I1201 16:46:32.974455 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://f65292529d9bba3480ced8eb81902012c9eb05b1dba8694d8ec12fc39a3e4b89" gracePeriod=600 Dec 01 16:46:33 crc kubenswrapper[4810]: I1201 16:46:33.625526 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="f65292529d9bba3480ced8eb81902012c9eb05b1dba8694d8ec12fc39a3e4b89" exitCode=0 Dec 01 16:46:33 crc kubenswrapper[4810]: I1201 16:46:33.625605 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"f65292529d9bba3480ced8eb81902012c9eb05b1dba8694d8ec12fc39a3e4b89"} Dec 01 16:46:33 crc kubenswrapper[4810]: I1201 16:46:33.626240 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3"} Dec 01 16:46:33 crc kubenswrapper[4810]: I1201 16:46:33.626265 4810 scope.go:117] "RemoveContainer" containerID="bde9da24046959a0ef2cefe62ef90c152965d00c4bfc047e7d8897c7609e3e5b" Dec 01 16:46:33 crc kubenswrapper[4810]: I1201 16:46:33.626405 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p58nl" podUID="af7c8a95-dc78-457a-a955-4a729b1425a6" containerName="registry-server" containerID="cri-o://2167c7bc0040fe653677893aee31cfdc07a4359777ad10ca97c79cb34c16d463" gracePeriod=2 Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.087203 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p58nl" Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.198980 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af7c8a95-dc78-457a-a955-4a729b1425a6-catalog-content\") pod \"af7c8a95-dc78-457a-a955-4a729b1425a6\" (UID: \"af7c8a95-dc78-457a-a955-4a729b1425a6\") " Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.199446 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af7c8a95-dc78-457a-a955-4a729b1425a6-utilities\") pod \"af7c8a95-dc78-457a-a955-4a729b1425a6\" (UID: \"af7c8a95-dc78-457a-a955-4a729b1425a6\") " Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.199977 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vg4dq\" (UniqueName: \"kubernetes.io/projected/af7c8a95-dc78-457a-a955-4a729b1425a6-kube-api-access-vg4dq\") pod \"af7c8a95-dc78-457a-a955-4a729b1425a6\" (UID: \"af7c8a95-dc78-457a-a955-4a729b1425a6\") " Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.200775 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af7c8a95-dc78-457a-a955-4a729b1425a6-utilities" (OuterVolumeSpecName: "utilities") pod "af7c8a95-dc78-457a-a955-4a729b1425a6" (UID: "af7c8a95-dc78-457a-a955-4a729b1425a6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.201553 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af7c8a95-dc78-457a-a955-4a729b1425a6-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.206454 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af7c8a95-dc78-457a-a955-4a729b1425a6-kube-api-access-vg4dq" (OuterVolumeSpecName: "kube-api-access-vg4dq") pod "af7c8a95-dc78-457a-a955-4a729b1425a6" (UID: "af7c8a95-dc78-457a-a955-4a729b1425a6"). InnerVolumeSpecName "kube-api-access-vg4dq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.304538 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vg4dq\" (UniqueName: \"kubernetes.io/projected/af7c8a95-dc78-457a-a955-4a729b1425a6-kube-api-access-vg4dq\") on node \"crc\" DevicePath \"\"" Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.511579 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af7c8a95-dc78-457a-a955-4a729b1425a6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af7c8a95-dc78-457a-a955-4a729b1425a6" (UID: "af7c8a95-dc78-457a-a955-4a729b1425a6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.611204 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af7c8a95-dc78-457a-a955-4a729b1425a6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.637056 4810 generic.go:334] "Generic (PLEG): container finished" podID="af7c8a95-dc78-457a-a955-4a729b1425a6" containerID="2167c7bc0040fe653677893aee31cfdc07a4359777ad10ca97c79cb34c16d463" exitCode=0 Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.637102 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p58nl" event={"ID":"af7c8a95-dc78-457a-a955-4a729b1425a6","Type":"ContainerDied","Data":"2167c7bc0040fe653677893aee31cfdc07a4359777ad10ca97c79cb34c16d463"} Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.637147 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p58nl" event={"ID":"af7c8a95-dc78-457a-a955-4a729b1425a6","Type":"ContainerDied","Data":"d514cd000ec6adec9157a0abb1a020cc50d8330d185b18cf89396e4f781c432f"} Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.637176 4810 scope.go:117] "RemoveContainer" containerID="2167c7bc0040fe653677893aee31cfdc07a4359777ad10ca97c79cb34c16d463" Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.637541 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p58nl" Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.660148 4810 scope.go:117] "RemoveContainer" containerID="245cbd4e8bdad650eff4679b8235a8d432bab6ddd50a34c09d7ef8c928cd8b6d" Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.691989 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p58nl"] Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.702838 4810 scope.go:117] "RemoveContainer" containerID="8ef93578c7718b3d77500f82602bae59ac60c940b1742f2f01817d7f9fd754a8" Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.704988 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p58nl"] Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.738971 4810 scope.go:117] "RemoveContainer" containerID="2167c7bc0040fe653677893aee31cfdc07a4359777ad10ca97c79cb34c16d463" Dec 01 16:46:34 crc kubenswrapper[4810]: E1201 16:46:34.739684 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2167c7bc0040fe653677893aee31cfdc07a4359777ad10ca97c79cb34c16d463\": container with ID starting with 2167c7bc0040fe653677893aee31cfdc07a4359777ad10ca97c79cb34c16d463 not found: ID does not exist" containerID="2167c7bc0040fe653677893aee31cfdc07a4359777ad10ca97c79cb34c16d463" Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.739732 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2167c7bc0040fe653677893aee31cfdc07a4359777ad10ca97c79cb34c16d463"} err="failed to get container status \"2167c7bc0040fe653677893aee31cfdc07a4359777ad10ca97c79cb34c16d463\": rpc error: code = NotFound desc = could not find container \"2167c7bc0040fe653677893aee31cfdc07a4359777ad10ca97c79cb34c16d463\": container with ID starting with 2167c7bc0040fe653677893aee31cfdc07a4359777ad10ca97c79cb34c16d463 not found: ID does not exist" Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.739760 4810 scope.go:117] "RemoveContainer" containerID="245cbd4e8bdad650eff4679b8235a8d432bab6ddd50a34c09d7ef8c928cd8b6d" Dec 01 16:46:34 crc kubenswrapper[4810]: E1201 16:46:34.740097 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"245cbd4e8bdad650eff4679b8235a8d432bab6ddd50a34c09d7ef8c928cd8b6d\": container with ID starting with 245cbd4e8bdad650eff4679b8235a8d432bab6ddd50a34c09d7ef8c928cd8b6d not found: ID does not exist" containerID="245cbd4e8bdad650eff4679b8235a8d432bab6ddd50a34c09d7ef8c928cd8b6d" Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.740125 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"245cbd4e8bdad650eff4679b8235a8d432bab6ddd50a34c09d7ef8c928cd8b6d"} err="failed to get container status \"245cbd4e8bdad650eff4679b8235a8d432bab6ddd50a34c09d7ef8c928cd8b6d\": rpc error: code = NotFound desc = could not find container \"245cbd4e8bdad650eff4679b8235a8d432bab6ddd50a34c09d7ef8c928cd8b6d\": container with ID starting with 245cbd4e8bdad650eff4679b8235a8d432bab6ddd50a34c09d7ef8c928cd8b6d not found: ID does not exist" Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.740140 4810 scope.go:117] "RemoveContainer" containerID="8ef93578c7718b3d77500f82602bae59ac60c940b1742f2f01817d7f9fd754a8" Dec 01 16:46:34 crc kubenswrapper[4810]: E1201 16:46:34.740344 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ef93578c7718b3d77500f82602bae59ac60c940b1742f2f01817d7f9fd754a8\": container with ID starting with 8ef93578c7718b3d77500f82602bae59ac60c940b1742f2f01817d7f9fd754a8 not found: ID does not exist" containerID="8ef93578c7718b3d77500f82602bae59ac60c940b1742f2f01817d7f9fd754a8" Dec 01 16:46:34 crc kubenswrapper[4810]: I1201 16:46:34.740378 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ef93578c7718b3d77500f82602bae59ac60c940b1742f2f01817d7f9fd754a8"} err="failed to get container status \"8ef93578c7718b3d77500f82602bae59ac60c940b1742f2f01817d7f9fd754a8\": rpc error: code = NotFound desc = could not find container \"8ef93578c7718b3d77500f82602bae59ac60c940b1742f2f01817d7f9fd754a8\": container with ID starting with 8ef93578c7718b3d77500f82602bae59ac60c940b1742f2f01817d7f9fd754a8 not found: ID does not exist" Dec 01 16:46:36 crc kubenswrapper[4810]: I1201 16:46:36.507556 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af7c8a95-dc78-457a-a955-4a729b1425a6" path="/var/lib/kubelet/pods/af7c8a95-dc78-457a-a955-4a729b1425a6/volumes" Dec 01 16:46:41 crc kubenswrapper[4810]: I1201 16:46:41.732493 4810 generic.go:334] "Generic (PLEG): container finished" podID="ba759ad7-3f5c-4d82-a406-54b887ede26a" containerID="c796c9b6212ffc7e1a82db08ea5c0ca0ad43dd5e1f08484e25eca185361c946b" exitCode=0 Dec 01 16:46:41 crc kubenswrapper[4810]: I1201 16:46:41.732597 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-5sbhr" event={"ID":"ba759ad7-3f5c-4d82-a406-54b887ede26a","Type":"ContainerDied","Data":"c796c9b6212ffc7e1a82db08ea5c0ca0ad43dd5e1f08484e25eca185361c946b"} Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.303834 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-5sbhr" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.409667 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5d6z\" (UniqueName: \"kubernetes.io/projected/ba759ad7-3f5c-4d82-a406-54b887ede26a-kube-api-access-g5d6z\") pod \"ba759ad7-3f5c-4d82-a406-54b887ede26a\" (UID: \"ba759ad7-3f5c-4d82-a406-54b887ede26a\") " Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.409722 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba759ad7-3f5c-4d82-a406-54b887ede26a-ssh-key\") pod \"ba759ad7-3f5c-4d82-a406-54b887ede26a\" (UID: \"ba759ad7-3f5c-4d82-a406-54b887ede26a\") " Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.409795 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba759ad7-3f5c-4d82-a406-54b887ede26a-inventory\") pod \"ba759ad7-3f5c-4d82-a406-54b887ede26a\" (UID: \"ba759ad7-3f5c-4d82-a406-54b887ede26a\") " Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.426781 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba759ad7-3f5c-4d82-a406-54b887ede26a-kube-api-access-g5d6z" (OuterVolumeSpecName: "kube-api-access-g5d6z") pod "ba759ad7-3f5c-4d82-a406-54b887ede26a" (UID: "ba759ad7-3f5c-4d82-a406-54b887ede26a"). InnerVolumeSpecName "kube-api-access-g5d6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.441747 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba759ad7-3f5c-4d82-a406-54b887ede26a-inventory" (OuterVolumeSpecName: "inventory") pod "ba759ad7-3f5c-4d82-a406-54b887ede26a" (UID: "ba759ad7-3f5c-4d82-a406-54b887ede26a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.460546 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba759ad7-3f5c-4d82-a406-54b887ede26a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ba759ad7-3f5c-4d82-a406-54b887ede26a" (UID: "ba759ad7-3f5c-4d82-a406-54b887ede26a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.511764 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5d6z\" (UniqueName: \"kubernetes.io/projected/ba759ad7-3f5c-4d82-a406-54b887ede26a-kube-api-access-g5d6z\") on node \"crc\" DevicePath \"\"" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.511817 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba759ad7-3f5c-4d82-a406-54b887ede26a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.511846 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba759ad7-3f5c-4d82-a406-54b887ede26a-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.755312 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-5sbhr" event={"ID":"ba759ad7-3f5c-4d82-a406-54b887ede26a","Type":"ContainerDied","Data":"3b98eddcc15a4be2a1f9839c18ab2b84b4958e7f677b596def05a3e4060bd4ae"} Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.755351 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b98eddcc15a4be2a1f9839c18ab2b84b4958e7f677b596def05a3e4060bd4ae" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.755385 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-5sbhr" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.848917 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-ksq6w"] Dec 01 16:46:43 crc kubenswrapper[4810]: E1201 16:46:43.849705 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba759ad7-3f5c-4d82-a406-54b887ede26a" containerName="configure-os-openstack-openstack-cell1" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.849825 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba759ad7-3f5c-4d82-a406-54b887ede26a" containerName="configure-os-openstack-openstack-cell1" Dec 01 16:46:43 crc kubenswrapper[4810]: E1201 16:46:43.850024 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af7c8a95-dc78-457a-a955-4a729b1425a6" containerName="extract-utilities" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.850123 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="af7c8a95-dc78-457a-a955-4a729b1425a6" containerName="extract-utilities" Dec 01 16:46:43 crc kubenswrapper[4810]: E1201 16:46:43.850188 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af7c8a95-dc78-457a-a955-4a729b1425a6" containerName="extract-content" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.850241 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="af7c8a95-dc78-457a-a955-4a729b1425a6" containerName="extract-content" Dec 01 16:46:43 crc kubenswrapper[4810]: E1201 16:46:43.850339 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af7c8a95-dc78-457a-a955-4a729b1425a6" containerName="registry-server" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.850392 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="af7c8a95-dc78-457a-a955-4a729b1425a6" containerName="registry-server" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.850705 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="af7c8a95-dc78-457a-a955-4a729b1425a6" containerName="registry-server" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.850787 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba759ad7-3f5c-4d82-a406-54b887ede26a" containerName="configure-os-openstack-openstack-cell1" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.851646 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-ksq6w" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.855552 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.856141 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.856834 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.857074 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.862588 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-ksq6w"] Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.927026 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6f47201d-aeaa-430b-bbc6-86b19519f233-inventory-0\") pod \"ssh-known-hosts-openstack-ksq6w\" (UID: \"6f47201d-aeaa-430b-bbc6-86b19519f233\") " pod="openstack/ssh-known-hosts-openstack-ksq6w" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.927210 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwjjr\" (UniqueName: \"kubernetes.io/projected/6f47201d-aeaa-430b-bbc6-86b19519f233-kube-api-access-cwjjr\") pod \"ssh-known-hosts-openstack-ksq6w\" (UID: \"6f47201d-aeaa-430b-bbc6-86b19519f233\") " pod="openstack/ssh-known-hosts-openstack-ksq6w" Dec 01 16:46:43 crc kubenswrapper[4810]: I1201 16:46:43.927448 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/6f47201d-aeaa-430b-bbc6-86b19519f233-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-ksq6w\" (UID: \"6f47201d-aeaa-430b-bbc6-86b19519f233\") " pod="openstack/ssh-known-hosts-openstack-ksq6w" Dec 01 16:46:44 crc kubenswrapper[4810]: I1201 16:46:44.029894 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/6f47201d-aeaa-430b-bbc6-86b19519f233-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-ksq6w\" (UID: \"6f47201d-aeaa-430b-bbc6-86b19519f233\") " pod="openstack/ssh-known-hosts-openstack-ksq6w" Dec 01 16:46:44 crc kubenswrapper[4810]: I1201 16:46:44.030024 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6f47201d-aeaa-430b-bbc6-86b19519f233-inventory-0\") pod \"ssh-known-hosts-openstack-ksq6w\" (UID: \"6f47201d-aeaa-430b-bbc6-86b19519f233\") " pod="openstack/ssh-known-hosts-openstack-ksq6w" Dec 01 16:46:44 crc kubenswrapper[4810]: I1201 16:46:44.030089 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwjjr\" (UniqueName: \"kubernetes.io/projected/6f47201d-aeaa-430b-bbc6-86b19519f233-kube-api-access-cwjjr\") pod \"ssh-known-hosts-openstack-ksq6w\" (UID: \"6f47201d-aeaa-430b-bbc6-86b19519f233\") " pod="openstack/ssh-known-hosts-openstack-ksq6w" Dec 01 16:46:44 crc kubenswrapper[4810]: I1201 16:46:44.035861 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6f47201d-aeaa-430b-bbc6-86b19519f233-inventory-0\") pod \"ssh-known-hosts-openstack-ksq6w\" (UID: \"6f47201d-aeaa-430b-bbc6-86b19519f233\") " pod="openstack/ssh-known-hosts-openstack-ksq6w" Dec 01 16:46:44 crc kubenswrapper[4810]: I1201 16:46:44.035972 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/6f47201d-aeaa-430b-bbc6-86b19519f233-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-ksq6w\" (UID: \"6f47201d-aeaa-430b-bbc6-86b19519f233\") " pod="openstack/ssh-known-hosts-openstack-ksq6w" Dec 01 16:46:44 crc kubenswrapper[4810]: I1201 16:46:44.057079 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwjjr\" (UniqueName: \"kubernetes.io/projected/6f47201d-aeaa-430b-bbc6-86b19519f233-kube-api-access-cwjjr\") pod \"ssh-known-hosts-openstack-ksq6w\" (UID: \"6f47201d-aeaa-430b-bbc6-86b19519f233\") " pod="openstack/ssh-known-hosts-openstack-ksq6w" Dec 01 16:46:44 crc kubenswrapper[4810]: I1201 16:46:44.171514 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-ksq6w" Dec 01 16:46:44 crc kubenswrapper[4810]: I1201 16:46:44.682415 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-ksq6w"] Dec 01 16:46:44 crc kubenswrapper[4810]: I1201 16:46:44.765283 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-ksq6w" event={"ID":"6f47201d-aeaa-430b-bbc6-86b19519f233","Type":"ContainerStarted","Data":"f66a19c9a0e3a4b08497d540ddde22c5fc1fc66f42cd1949f2240dad7b78ab52"} Dec 01 16:46:45 crc kubenswrapper[4810]: I1201 16:46:45.775593 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-ksq6w" event={"ID":"6f47201d-aeaa-430b-bbc6-86b19519f233","Type":"ContainerStarted","Data":"eb2a233358e7ecd8c60db991d75f0bff2055e1240e52a3e6e50926d7887fefca"} Dec 01 16:46:45 crc kubenswrapper[4810]: I1201 16:46:45.802940 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-ksq6w" podStartSLOduration=2.247112181 podStartE2EDuration="2.802923736s" podCreationTimestamp="2025-12-01 16:46:43 +0000 UTC" firstStartedPulling="2025-12-01 16:46:44.687399271 +0000 UTC m=+7970.450908874" lastFinishedPulling="2025-12-01 16:46:45.243210826 +0000 UTC m=+7971.006720429" observedRunningTime="2025-12-01 16:46:45.793931993 +0000 UTC m=+7971.557441626" watchObservedRunningTime="2025-12-01 16:46:45.802923736 +0000 UTC m=+7971.566433339" Dec 01 16:46:53 crc kubenswrapper[4810]: I1201 16:46:53.853322 4810 generic.go:334] "Generic (PLEG): container finished" podID="6f47201d-aeaa-430b-bbc6-86b19519f233" containerID="eb2a233358e7ecd8c60db991d75f0bff2055e1240e52a3e6e50926d7887fefca" exitCode=0 Dec 01 16:46:53 crc kubenswrapper[4810]: I1201 16:46:53.853388 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-ksq6w" event={"ID":"6f47201d-aeaa-430b-bbc6-86b19519f233","Type":"ContainerDied","Data":"eb2a233358e7ecd8c60db991d75f0bff2055e1240e52a3e6e50926d7887fefca"} Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.305138 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-ksq6w" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.374847 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/6f47201d-aeaa-430b-bbc6-86b19519f233-ssh-key-openstack-cell1\") pod \"6f47201d-aeaa-430b-bbc6-86b19519f233\" (UID: \"6f47201d-aeaa-430b-bbc6-86b19519f233\") " Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.374936 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwjjr\" (UniqueName: \"kubernetes.io/projected/6f47201d-aeaa-430b-bbc6-86b19519f233-kube-api-access-cwjjr\") pod \"6f47201d-aeaa-430b-bbc6-86b19519f233\" (UID: \"6f47201d-aeaa-430b-bbc6-86b19519f233\") " Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.375058 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6f47201d-aeaa-430b-bbc6-86b19519f233-inventory-0\") pod \"6f47201d-aeaa-430b-bbc6-86b19519f233\" (UID: \"6f47201d-aeaa-430b-bbc6-86b19519f233\") " Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.381705 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f47201d-aeaa-430b-bbc6-86b19519f233-kube-api-access-cwjjr" (OuterVolumeSpecName: "kube-api-access-cwjjr") pod "6f47201d-aeaa-430b-bbc6-86b19519f233" (UID: "6f47201d-aeaa-430b-bbc6-86b19519f233"). InnerVolumeSpecName "kube-api-access-cwjjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.403747 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f47201d-aeaa-430b-bbc6-86b19519f233-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "6f47201d-aeaa-430b-bbc6-86b19519f233" (UID: "6f47201d-aeaa-430b-bbc6-86b19519f233"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.405560 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f47201d-aeaa-430b-bbc6-86b19519f233-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "6f47201d-aeaa-430b-bbc6-86b19519f233" (UID: "6f47201d-aeaa-430b-bbc6-86b19519f233"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.478354 4810 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6f47201d-aeaa-430b-bbc6-86b19519f233-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.478409 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/6f47201d-aeaa-430b-bbc6-86b19519f233-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.478432 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwjjr\" (UniqueName: \"kubernetes.io/projected/6f47201d-aeaa-430b-bbc6-86b19519f233-kube-api-access-cwjjr\") on node \"crc\" DevicePath \"\"" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.872806 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-ksq6w" event={"ID":"6f47201d-aeaa-430b-bbc6-86b19519f233","Type":"ContainerDied","Data":"f66a19c9a0e3a4b08497d540ddde22c5fc1fc66f42cd1949f2240dad7b78ab52"} Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.873132 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f66a19c9a0e3a4b08497d540ddde22c5fc1fc66f42cd1949f2240dad7b78ab52" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.873064 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-ksq6w" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.945975 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-h4xb9"] Dec 01 16:46:55 crc kubenswrapper[4810]: E1201 16:46:55.946557 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f47201d-aeaa-430b-bbc6-86b19519f233" containerName="ssh-known-hosts-openstack" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.946576 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f47201d-aeaa-430b-bbc6-86b19519f233" containerName="ssh-known-hosts-openstack" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.946907 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f47201d-aeaa-430b-bbc6-86b19519f233" containerName="ssh-known-hosts-openstack" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.947710 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-h4xb9" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.952674 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.952763 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.952913 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.952971 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.966372 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-h4xb9"] Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.995691 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z9pg\" (UniqueName: \"kubernetes.io/projected/c2ad2c28-93e4-4521-96ab-40c186aade64-kube-api-access-9z9pg\") pod \"run-os-openstack-openstack-cell1-h4xb9\" (UID: \"c2ad2c28-93e4-4521-96ab-40c186aade64\") " pod="openstack/run-os-openstack-openstack-cell1-h4xb9" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.996316 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2ad2c28-93e4-4521-96ab-40c186aade64-ssh-key\") pod \"run-os-openstack-openstack-cell1-h4xb9\" (UID: \"c2ad2c28-93e4-4521-96ab-40c186aade64\") " pod="openstack/run-os-openstack-openstack-cell1-h4xb9" Dec 01 16:46:55 crc kubenswrapper[4810]: I1201 16:46:55.996406 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2ad2c28-93e4-4521-96ab-40c186aade64-inventory\") pod \"run-os-openstack-openstack-cell1-h4xb9\" (UID: \"c2ad2c28-93e4-4521-96ab-40c186aade64\") " pod="openstack/run-os-openstack-openstack-cell1-h4xb9" Dec 01 16:46:56 crc kubenswrapper[4810]: I1201 16:46:56.097890 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2ad2c28-93e4-4521-96ab-40c186aade64-ssh-key\") pod \"run-os-openstack-openstack-cell1-h4xb9\" (UID: \"c2ad2c28-93e4-4521-96ab-40c186aade64\") " pod="openstack/run-os-openstack-openstack-cell1-h4xb9" Dec 01 16:46:56 crc kubenswrapper[4810]: I1201 16:46:56.097951 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2ad2c28-93e4-4521-96ab-40c186aade64-inventory\") pod \"run-os-openstack-openstack-cell1-h4xb9\" (UID: \"c2ad2c28-93e4-4521-96ab-40c186aade64\") " pod="openstack/run-os-openstack-openstack-cell1-h4xb9" Dec 01 16:46:56 crc kubenswrapper[4810]: I1201 16:46:56.098010 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z9pg\" (UniqueName: \"kubernetes.io/projected/c2ad2c28-93e4-4521-96ab-40c186aade64-kube-api-access-9z9pg\") pod \"run-os-openstack-openstack-cell1-h4xb9\" (UID: \"c2ad2c28-93e4-4521-96ab-40c186aade64\") " pod="openstack/run-os-openstack-openstack-cell1-h4xb9" Dec 01 16:46:56 crc kubenswrapper[4810]: I1201 16:46:56.109102 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2ad2c28-93e4-4521-96ab-40c186aade64-ssh-key\") pod \"run-os-openstack-openstack-cell1-h4xb9\" (UID: \"c2ad2c28-93e4-4521-96ab-40c186aade64\") " pod="openstack/run-os-openstack-openstack-cell1-h4xb9" Dec 01 16:46:56 crc kubenswrapper[4810]: I1201 16:46:56.109946 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2ad2c28-93e4-4521-96ab-40c186aade64-inventory\") pod \"run-os-openstack-openstack-cell1-h4xb9\" (UID: \"c2ad2c28-93e4-4521-96ab-40c186aade64\") " pod="openstack/run-os-openstack-openstack-cell1-h4xb9" Dec 01 16:46:56 crc kubenswrapper[4810]: I1201 16:46:56.114542 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z9pg\" (UniqueName: \"kubernetes.io/projected/c2ad2c28-93e4-4521-96ab-40c186aade64-kube-api-access-9z9pg\") pod \"run-os-openstack-openstack-cell1-h4xb9\" (UID: \"c2ad2c28-93e4-4521-96ab-40c186aade64\") " pod="openstack/run-os-openstack-openstack-cell1-h4xb9" Dec 01 16:46:56 crc kubenswrapper[4810]: I1201 16:46:56.272725 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-h4xb9" Dec 01 16:46:56 crc kubenswrapper[4810]: W1201 16:46:56.811703 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2ad2c28_93e4_4521_96ab_40c186aade64.slice/crio-e3845b67f71d19aaaa94386f1c17ded1d7168771d1da9b23d64e9eddb913014f WatchSource:0}: Error finding container e3845b67f71d19aaaa94386f1c17ded1d7168771d1da9b23d64e9eddb913014f: Status 404 returned error can't find the container with id e3845b67f71d19aaaa94386f1c17ded1d7168771d1da9b23d64e9eddb913014f Dec 01 16:46:56 crc kubenswrapper[4810]: I1201 16:46:56.812161 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-h4xb9"] Dec 01 16:46:56 crc kubenswrapper[4810]: I1201 16:46:56.883730 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-h4xb9" event={"ID":"c2ad2c28-93e4-4521-96ab-40c186aade64","Type":"ContainerStarted","Data":"e3845b67f71d19aaaa94386f1c17ded1d7168771d1da9b23d64e9eddb913014f"} Dec 01 16:46:57 crc kubenswrapper[4810]: I1201 16:46:57.908459 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-h4xb9" event={"ID":"c2ad2c28-93e4-4521-96ab-40c186aade64","Type":"ContainerStarted","Data":"81d846f9a3c800abbc1ea2fd5332d60ac6454e73bc90536de748a658f62b784f"} Dec 01 16:46:57 crc kubenswrapper[4810]: I1201 16:46:57.929793 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-h4xb9" podStartSLOduration=2.302023517 podStartE2EDuration="2.929764264s" podCreationTimestamp="2025-12-01 16:46:55 +0000 UTC" firstStartedPulling="2025-12-01 16:46:56.814826063 +0000 UTC m=+7982.578335666" lastFinishedPulling="2025-12-01 16:46:57.44256681 +0000 UTC m=+7983.206076413" observedRunningTime="2025-12-01 16:46:57.925821647 +0000 UTC m=+7983.689331260" watchObservedRunningTime="2025-12-01 16:46:57.929764264 +0000 UTC m=+7983.693273887" Dec 01 16:47:07 crc kubenswrapper[4810]: I1201 16:47:07.028075 4810 generic.go:334] "Generic (PLEG): container finished" podID="c2ad2c28-93e4-4521-96ab-40c186aade64" containerID="81d846f9a3c800abbc1ea2fd5332d60ac6454e73bc90536de748a658f62b784f" exitCode=0 Dec 01 16:47:07 crc kubenswrapper[4810]: I1201 16:47:07.028176 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-h4xb9" event={"ID":"c2ad2c28-93e4-4521-96ab-40c186aade64","Type":"ContainerDied","Data":"81d846f9a3c800abbc1ea2fd5332d60ac6454e73bc90536de748a658f62b784f"} Dec 01 16:47:08 crc kubenswrapper[4810]: I1201 16:47:08.469663 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-h4xb9" Dec 01 16:47:08 crc kubenswrapper[4810]: I1201 16:47:08.590342 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2ad2c28-93e4-4521-96ab-40c186aade64-inventory\") pod \"c2ad2c28-93e4-4521-96ab-40c186aade64\" (UID: \"c2ad2c28-93e4-4521-96ab-40c186aade64\") " Dec 01 16:47:08 crc kubenswrapper[4810]: I1201 16:47:08.590479 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2ad2c28-93e4-4521-96ab-40c186aade64-ssh-key\") pod \"c2ad2c28-93e4-4521-96ab-40c186aade64\" (UID: \"c2ad2c28-93e4-4521-96ab-40c186aade64\") " Dec 01 16:47:08 crc kubenswrapper[4810]: I1201 16:47:08.590503 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9z9pg\" (UniqueName: \"kubernetes.io/projected/c2ad2c28-93e4-4521-96ab-40c186aade64-kube-api-access-9z9pg\") pod \"c2ad2c28-93e4-4521-96ab-40c186aade64\" (UID: \"c2ad2c28-93e4-4521-96ab-40c186aade64\") " Dec 01 16:47:08 crc kubenswrapper[4810]: I1201 16:47:08.605048 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2ad2c28-93e4-4521-96ab-40c186aade64-kube-api-access-9z9pg" (OuterVolumeSpecName: "kube-api-access-9z9pg") pod "c2ad2c28-93e4-4521-96ab-40c186aade64" (UID: "c2ad2c28-93e4-4521-96ab-40c186aade64"). InnerVolumeSpecName "kube-api-access-9z9pg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:47:08 crc kubenswrapper[4810]: I1201 16:47:08.623691 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2ad2c28-93e4-4521-96ab-40c186aade64-inventory" (OuterVolumeSpecName: "inventory") pod "c2ad2c28-93e4-4521-96ab-40c186aade64" (UID: "c2ad2c28-93e4-4521-96ab-40c186aade64"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:47:08 crc kubenswrapper[4810]: I1201 16:47:08.625492 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2ad2c28-93e4-4521-96ab-40c186aade64-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c2ad2c28-93e4-4521-96ab-40c186aade64" (UID: "c2ad2c28-93e4-4521-96ab-40c186aade64"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:47:08 crc kubenswrapper[4810]: I1201 16:47:08.694228 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2ad2c28-93e4-4521-96ab-40c186aade64-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:47:08 crc kubenswrapper[4810]: I1201 16:47:08.694257 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2ad2c28-93e4-4521-96ab-40c186aade64-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:47:08 crc kubenswrapper[4810]: I1201 16:47:08.694266 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9z9pg\" (UniqueName: \"kubernetes.io/projected/c2ad2c28-93e4-4521-96ab-40c186aade64-kube-api-access-9z9pg\") on node \"crc\" DevicePath \"\"" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.047580 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-h4xb9" event={"ID":"c2ad2c28-93e4-4521-96ab-40c186aade64","Type":"ContainerDied","Data":"e3845b67f71d19aaaa94386f1c17ded1d7168771d1da9b23d64e9eddb913014f"} Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.047620 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3845b67f71d19aaaa94386f1c17ded1d7168771d1da9b23d64e9eddb913014f" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.047641 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-h4xb9" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.152538 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-j84h9"] Dec 01 16:47:09 crc kubenswrapper[4810]: E1201 16:47:09.153325 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2ad2c28-93e4-4521-96ab-40c186aade64" containerName="run-os-openstack-openstack-cell1" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.153351 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2ad2c28-93e4-4521-96ab-40c186aade64" containerName="run-os-openstack-openstack-cell1" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.153690 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2ad2c28-93e4-4521-96ab-40c186aade64" containerName="run-os-openstack-openstack-cell1" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.154912 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-j84h9" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.160917 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.163771 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.164053 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.165616 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.172922 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-j84h9"] Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.206132 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/60cc48f6-a14e-4bb3-b979-60e2a240313c-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-j84h9\" (UID: \"60cc48f6-a14e-4bb3-b979-60e2a240313c\") " pod="openstack/reboot-os-openstack-openstack-cell1-j84h9" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.206184 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-794dz\" (UniqueName: \"kubernetes.io/projected/60cc48f6-a14e-4bb3-b979-60e2a240313c-kube-api-access-794dz\") pod \"reboot-os-openstack-openstack-cell1-j84h9\" (UID: \"60cc48f6-a14e-4bb3-b979-60e2a240313c\") " pod="openstack/reboot-os-openstack-openstack-cell1-j84h9" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.206496 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60cc48f6-a14e-4bb3-b979-60e2a240313c-inventory\") pod \"reboot-os-openstack-openstack-cell1-j84h9\" (UID: \"60cc48f6-a14e-4bb3-b979-60e2a240313c\") " pod="openstack/reboot-os-openstack-openstack-cell1-j84h9" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.308690 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60cc48f6-a14e-4bb3-b979-60e2a240313c-inventory\") pod \"reboot-os-openstack-openstack-cell1-j84h9\" (UID: \"60cc48f6-a14e-4bb3-b979-60e2a240313c\") " pod="openstack/reboot-os-openstack-openstack-cell1-j84h9" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.308855 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/60cc48f6-a14e-4bb3-b979-60e2a240313c-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-j84h9\" (UID: \"60cc48f6-a14e-4bb3-b979-60e2a240313c\") " pod="openstack/reboot-os-openstack-openstack-cell1-j84h9" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.308892 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-794dz\" (UniqueName: \"kubernetes.io/projected/60cc48f6-a14e-4bb3-b979-60e2a240313c-kube-api-access-794dz\") pod \"reboot-os-openstack-openstack-cell1-j84h9\" (UID: \"60cc48f6-a14e-4bb3-b979-60e2a240313c\") " pod="openstack/reboot-os-openstack-openstack-cell1-j84h9" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.320010 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60cc48f6-a14e-4bb3-b979-60e2a240313c-inventory\") pod \"reboot-os-openstack-openstack-cell1-j84h9\" (UID: \"60cc48f6-a14e-4bb3-b979-60e2a240313c\") " pod="openstack/reboot-os-openstack-openstack-cell1-j84h9" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.320009 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/60cc48f6-a14e-4bb3-b979-60e2a240313c-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-j84h9\" (UID: \"60cc48f6-a14e-4bb3-b979-60e2a240313c\") " pod="openstack/reboot-os-openstack-openstack-cell1-j84h9" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.333037 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-794dz\" (UniqueName: \"kubernetes.io/projected/60cc48f6-a14e-4bb3-b979-60e2a240313c-kube-api-access-794dz\") pod \"reboot-os-openstack-openstack-cell1-j84h9\" (UID: \"60cc48f6-a14e-4bb3-b979-60e2a240313c\") " pod="openstack/reboot-os-openstack-openstack-cell1-j84h9" Dec 01 16:47:09 crc kubenswrapper[4810]: I1201 16:47:09.482930 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-j84h9" Dec 01 16:47:10 crc kubenswrapper[4810]: I1201 16:47:10.047330 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-j84h9"] Dec 01 16:47:10 crc kubenswrapper[4810]: I1201 16:47:10.050670 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:47:10 crc kubenswrapper[4810]: I1201 16:47:10.059792 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-j84h9" event={"ID":"60cc48f6-a14e-4bb3-b979-60e2a240313c","Type":"ContainerStarted","Data":"1370381f32ca17be968d330ef72130db5ce3e6fe5a56b1e798509766bcb65648"} Dec 01 16:47:11 crc kubenswrapper[4810]: I1201 16:47:11.070189 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-j84h9" event={"ID":"60cc48f6-a14e-4bb3-b979-60e2a240313c","Type":"ContainerStarted","Data":"57e148950badb68a5412578c29067cede5ae1c350ae0c757847b3e3278b37821"} Dec 01 16:47:11 crc kubenswrapper[4810]: I1201 16:47:11.095649 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-j84h9" podStartSLOduration=1.652924568 podStartE2EDuration="2.095626315s" podCreationTimestamp="2025-12-01 16:47:09 +0000 UTC" firstStartedPulling="2025-12-01 16:47:10.050453167 +0000 UTC m=+7995.813962770" lastFinishedPulling="2025-12-01 16:47:10.493154914 +0000 UTC m=+7996.256664517" observedRunningTime="2025-12-01 16:47:11.092386977 +0000 UTC m=+7996.855896620" watchObservedRunningTime="2025-12-01 16:47:11.095626315 +0000 UTC m=+7996.859135918" Dec 01 16:47:27 crc kubenswrapper[4810]: I1201 16:47:27.211705 4810 generic.go:334] "Generic (PLEG): container finished" podID="60cc48f6-a14e-4bb3-b979-60e2a240313c" containerID="57e148950badb68a5412578c29067cede5ae1c350ae0c757847b3e3278b37821" exitCode=0 Dec 01 16:47:27 crc kubenswrapper[4810]: I1201 16:47:27.211795 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-j84h9" event={"ID":"60cc48f6-a14e-4bb3-b979-60e2a240313c","Type":"ContainerDied","Data":"57e148950badb68a5412578c29067cede5ae1c350ae0c757847b3e3278b37821"} Dec 01 16:47:28 crc kubenswrapper[4810]: I1201 16:47:28.700585 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-j84h9" Dec 01 16:47:28 crc kubenswrapper[4810]: I1201 16:47:28.737880 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60cc48f6-a14e-4bb3-b979-60e2a240313c-inventory\") pod \"60cc48f6-a14e-4bb3-b979-60e2a240313c\" (UID: \"60cc48f6-a14e-4bb3-b979-60e2a240313c\") " Dec 01 16:47:28 crc kubenswrapper[4810]: I1201 16:47:28.737936 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/60cc48f6-a14e-4bb3-b979-60e2a240313c-ssh-key\") pod \"60cc48f6-a14e-4bb3-b979-60e2a240313c\" (UID: \"60cc48f6-a14e-4bb3-b979-60e2a240313c\") " Dec 01 16:47:28 crc kubenswrapper[4810]: I1201 16:47:28.737978 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-794dz\" (UniqueName: \"kubernetes.io/projected/60cc48f6-a14e-4bb3-b979-60e2a240313c-kube-api-access-794dz\") pod \"60cc48f6-a14e-4bb3-b979-60e2a240313c\" (UID: \"60cc48f6-a14e-4bb3-b979-60e2a240313c\") " Dec 01 16:47:28 crc kubenswrapper[4810]: I1201 16:47:28.745018 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60cc48f6-a14e-4bb3-b979-60e2a240313c-kube-api-access-794dz" (OuterVolumeSpecName: "kube-api-access-794dz") pod "60cc48f6-a14e-4bb3-b979-60e2a240313c" (UID: "60cc48f6-a14e-4bb3-b979-60e2a240313c"). InnerVolumeSpecName "kube-api-access-794dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:47:28 crc kubenswrapper[4810]: I1201 16:47:28.774755 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60cc48f6-a14e-4bb3-b979-60e2a240313c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "60cc48f6-a14e-4bb3-b979-60e2a240313c" (UID: "60cc48f6-a14e-4bb3-b979-60e2a240313c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:47:28 crc kubenswrapper[4810]: I1201 16:47:28.793202 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60cc48f6-a14e-4bb3-b979-60e2a240313c-inventory" (OuterVolumeSpecName: "inventory") pod "60cc48f6-a14e-4bb3-b979-60e2a240313c" (UID: "60cc48f6-a14e-4bb3-b979-60e2a240313c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:47:28 crc kubenswrapper[4810]: I1201 16:47:28.839567 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60cc48f6-a14e-4bb3-b979-60e2a240313c-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:47:28 crc kubenswrapper[4810]: I1201 16:47:28.839604 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/60cc48f6-a14e-4bb3-b979-60e2a240313c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:47:28 crc kubenswrapper[4810]: I1201 16:47:28.839616 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-794dz\" (UniqueName: \"kubernetes.io/projected/60cc48f6-a14e-4bb3-b979-60e2a240313c-kube-api-access-794dz\") on node \"crc\" DevicePath \"\"" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.233258 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-j84h9" event={"ID":"60cc48f6-a14e-4bb3-b979-60e2a240313c","Type":"ContainerDied","Data":"1370381f32ca17be968d330ef72130db5ce3e6fe5a56b1e798509766bcb65648"} Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.233308 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1370381f32ca17be968d330ef72130db5ce3e6fe5a56b1e798509766bcb65648" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.233384 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-j84h9" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.340766 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-sr2t8"] Dec 01 16:47:29 crc kubenswrapper[4810]: E1201 16:47:29.341693 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60cc48f6-a14e-4bb3-b979-60e2a240313c" containerName="reboot-os-openstack-openstack-cell1" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.341715 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="60cc48f6-a14e-4bb3-b979-60e2a240313c" containerName="reboot-os-openstack-openstack-cell1" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.341964 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="60cc48f6-a14e-4bb3-b979-60e2a240313c" containerName="reboot-os-openstack-openstack-cell1" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.342966 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.346646 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.347265 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-ovn-default-certs-0" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.347394 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-telemetry-default-certs-0" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.347537 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.347711 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-neutron-metadata-default-certs-0" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.347884 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.347991 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.348119 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-libvirt-default-certs-0" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.352708 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-sr2t8"] Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.451738 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.452365 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.452497 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.452582 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-inventory\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.452782 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.452836 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.452902 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.452918 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.453140 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.453211 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.453248 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.453289 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5jws\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-kube-api-access-l5jws\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.453325 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.453509 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-ssh-key\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.453549 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.555987 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-ssh-key\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.556048 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.556126 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.556223 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.556255 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.557152 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-inventory\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.557221 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.557243 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.557270 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.557787 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.557991 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.558034 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.558119 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.558170 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5jws\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-kube-api-access-l5jws\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.558191 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.561952 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.562671 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.563287 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.563556 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-ssh-key\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.563586 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.563713 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-inventory\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.564117 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.564391 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.564790 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.564969 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.566200 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.566921 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.575268 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.575410 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.582253 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5jws\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-kube-api-access-l5jws\") pod \"install-certs-openstack-openstack-cell1-sr2t8\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:29 crc kubenswrapper[4810]: I1201 16:47:29.662986 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:47:30 crc kubenswrapper[4810]: I1201 16:47:30.253707 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-sr2t8"] Dec 01 16:47:31 crc kubenswrapper[4810]: I1201 16:47:31.252048 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" event={"ID":"17bffe79-728c-4500-a27e-a2b246f82517","Type":"ContainerStarted","Data":"73d803a0d0bc679294517a01ec0fd01a0631dfc10547865863889465df9287ca"} Dec 01 16:47:32 crc kubenswrapper[4810]: I1201 16:47:32.262948 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" event={"ID":"17bffe79-728c-4500-a27e-a2b246f82517","Type":"ContainerStarted","Data":"8346b581e46a6b78db0a34224ac6e0ef6caa4c783bf115c905ce6075c8213994"} Dec 01 16:47:32 crc kubenswrapper[4810]: I1201 16:47:32.286949 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" podStartSLOduration=1.6087568399999999 podStartE2EDuration="3.286929329s" podCreationTimestamp="2025-12-01 16:47:29 +0000 UTC" firstStartedPulling="2025-12-01 16:47:30.263586439 +0000 UTC m=+8016.027096042" lastFinishedPulling="2025-12-01 16:47:31.941758928 +0000 UTC m=+8017.705268531" observedRunningTime="2025-12-01 16:47:32.279988881 +0000 UTC m=+8018.043498504" watchObservedRunningTime="2025-12-01 16:47:32.286929329 +0000 UTC m=+8018.050438932" Dec 01 16:47:41 crc kubenswrapper[4810]: I1201 16:47:41.116989 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vqd54"] Dec 01 16:47:41 crc kubenswrapper[4810]: I1201 16:47:41.121674 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqd54" Dec 01 16:47:41 crc kubenswrapper[4810]: I1201 16:47:41.127363 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vqd54"] Dec 01 16:47:41 crc kubenswrapper[4810]: I1201 16:47:41.235963 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsxhd\" (UniqueName: \"kubernetes.io/projected/71af81c3-413f-4a2c-9c5f-220f412b2998-kube-api-access-dsxhd\") pod \"community-operators-vqd54\" (UID: \"71af81c3-413f-4a2c-9c5f-220f412b2998\") " pod="openshift-marketplace/community-operators-vqd54" Dec 01 16:47:41 crc kubenswrapper[4810]: I1201 16:47:41.236040 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71af81c3-413f-4a2c-9c5f-220f412b2998-utilities\") pod \"community-operators-vqd54\" (UID: \"71af81c3-413f-4a2c-9c5f-220f412b2998\") " pod="openshift-marketplace/community-operators-vqd54" Dec 01 16:47:41 crc kubenswrapper[4810]: I1201 16:47:41.236367 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71af81c3-413f-4a2c-9c5f-220f412b2998-catalog-content\") pod \"community-operators-vqd54\" (UID: \"71af81c3-413f-4a2c-9c5f-220f412b2998\") " pod="openshift-marketplace/community-operators-vqd54" Dec 01 16:47:41 crc kubenswrapper[4810]: I1201 16:47:41.338096 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71af81c3-413f-4a2c-9c5f-220f412b2998-catalog-content\") pod \"community-operators-vqd54\" (UID: \"71af81c3-413f-4a2c-9c5f-220f412b2998\") " pod="openshift-marketplace/community-operators-vqd54" Dec 01 16:47:41 crc kubenswrapper[4810]: I1201 16:47:41.338210 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsxhd\" (UniqueName: \"kubernetes.io/projected/71af81c3-413f-4a2c-9c5f-220f412b2998-kube-api-access-dsxhd\") pod \"community-operators-vqd54\" (UID: \"71af81c3-413f-4a2c-9c5f-220f412b2998\") " pod="openshift-marketplace/community-operators-vqd54" Dec 01 16:47:41 crc kubenswrapper[4810]: I1201 16:47:41.338253 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71af81c3-413f-4a2c-9c5f-220f412b2998-utilities\") pod \"community-operators-vqd54\" (UID: \"71af81c3-413f-4a2c-9c5f-220f412b2998\") " pod="openshift-marketplace/community-operators-vqd54" Dec 01 16:47:41 crc kubenswrapper[4810]: I1201 16:47:41.338727 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71af81c3-413f-4a2c-9c5f-220f412b2998-catalog-content\") pod \"community-operators-vqd54\" (UID: \"71af81c3-413f-4a2c-9c5f-220f412b2998\") " pod="openshift-marketplace/community-operators-vqd54" Dec 01 16:47:41 crc kubenswrapper[4810]: I1201 16:47:41.338792 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71af81c3-413f-4a2c-9c5f-220f412b2998-utilities\") pod \"community-operators-vqd54\" (UID: \"71af81c3-413f-4a2c-9c5f-220f412b2998\") " pod="openshift-marketplace/community-operators-vqd54" Dec 01 16:47:41 crc kubenswrapper[4810]: I1201 16:47:41.357405 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsxhd\" (UniqueName: \"kubernetes.io/projected/71af81c3-413f-4a2c-9c5f-220f412b2998-kube-api-access-dsxhd\") pod \"community-operators-vqd54\" (UID: \"71af81c3-413f-4a2c-9c5f-220f412b2998\") " pod="openshift-marketplace/community-operators-vqd54" Dec 01 16:47:41 crc kubenswrapper[4810]: I1201 16:47:41.457066 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqd54" Dec 01 16:47:42 crc kubenswrapper[4810]: I1201 16:47:42.055216 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vqd54"] Dec 01 16:47:42 crc kubenswrapper[4810]: I1201 16:47:42.345131 4810 generic.go:334] "Generic (PLEG): container finished" podID="71af81c3-413f-4a2c-9c5f-220f412b2998" containerID="d70b13e2f828295eb256f5605dc37c39afa73b9477b46c8d304e44c2257e4f97" exitCode=0 Dec 01 16:47:42 crc kubenswrapper[4810]: I1201 16:47:42.345184 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqd54" event={"ID":"71af81c3-413f-4a2c-9c5f-220f412b2998","Type":"ContainerDied","Data":"d70b13e2f828295eb256f5605dc37c39afa73b9477b46c8d304e44c2257e4f97"} Dec 01 16:47:42 crc kubenswrapper[4810]: I1201 16:47:42.345211 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqd54" event={"ID":"71af81c3-413f-4a2c-9c5f-220f412b2998","Type":"ContainerStarted","Data":"414dedd59061a8ab97aedce9562710ec34d90be360bc56ded41b7a61b90f9c25"} Dec 01 16:47:43 crc kubenswrapper[4810]: I1201 16:47:43.359562 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqd54" event={"ID":"71af81c3-413f-4a2c-9c5f-220f412b2998","Type":"ContainerStarted","Data":"5048f7c6b2aaa906ec16f2b423f33337907dbf6de3e427fc84c509c9471afdac"} Dec 01 16:47:44 crc kubenswrapper[4810]: I1201 16:47:44.371990 4810 generic.go:334] "Generic (PLEG): container finished" podID="71af81c3-413f-4a2c-9c5f-220f412b2998" containerID="5048f7c6b2aaa906ec16f2b423f33337907dbf6de3e427fc84c509c9471afdac" exitCode=0 Dec 01 16:47:44 crc kubenswrapper[4810]: I1201 16:47:44.372075 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqd54" event={"ID":"71af81c3-413f-4a2c-9c5f-220f412b2998","Type":"ContainerDied","Data":"5048f7c6b2aaa906ec16f2b423f33337907dbf6de3e427fc84c509c9471afdac"} Dec 01 16:47:45 crc kubenswrapper[4810]: I1201 16:47:45.381644 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqd54" event={"ID":"71af81c3-413f-4a2c-9c5f-220f412b2998","Type":"ContainerStarted","Data":"a2e3718e8a3ff5bb29bf2c9ebff1f76b6c219c99d1cedc5f77dd90ef0de55b28"} Dec 01 16:47:45 crc kubenswrapper[4810]: I1201 16:47:45.404248 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vqd54" podStartSLOduration=1.6622101009999999 podStartE2EDuration="4.404229869s" podCreationTimestamp="2025-12-01 16:47:41 +0000 UTC" firstStartedPulling="2025-12-01 16:47:42.348544545 +0000 UTC m=+8028.112054148" lastFinishedPulling="2025-12-01 16:47:45.090564313 +0000 UTC m=+8030.854073916" observedRunningTime="2025-12-01 16:47:45.395952255 +0000 UTC m=+8031.159461858" watchObservedRunningTime="2025-12-01 16:47:45.404229869 +0000 UTC m=+8031.167739472" Dec 01 16:47:51 crc kubenswrapper[4810]: I1201 16:47:51.461622 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vqd54" Dec 01 16:47:51 crc kubenswrapper[4810]: I1201 16:47:51.462598 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vqd54" Dec 01 16:47:51 crc kubenswrapper[4810]: I1201 16:47:51.521086 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vqd54" Dec 01 16:47:52 crc kubenswrapper[4810]: I1201 16:47:52.509795 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vqd54" Dec 01 16:47:52 crc kubenswrapper[4810]: I1201 16:47:52.573269 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vqd54"] Dec 01 16:47:54 crc kubenswrapper[4810]: I1201 16:47:54.479121 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vqd54" podUID="71af81c3-413f-4a2c-9c5f-220f412b2998" containerName="registry-server" containerID="cri-o://a2e3718e8a3ff5bb29bf2c9ebff1f76b6c219c99d1cedc5f77dd90ef0de55b28" gracePeriod=2 Dec 01 16:47:54 crc kubenswrapper[4810]: I1201 16:47:54.943846 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqd54" Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.047894 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsxhd\" (UniqueName: \"kubernetes.io/projected/71af81c3-413f-4a2c-9c5f-220f412b2998-kube-api-access-dsxhd\") pod \"71af81c3-413f-4a2c-9c5f-220f412b2998\" (UID: \"71af81c3-413f-4a2c-9c5f-220f412b2998\") " Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.048105 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71af81c3-413f-4a2c-9c5f-220f412b2998-catalog-content\") pod \"71af81c3-413f-4a2c-9c5f-220f412b2998\" (UID: \"71af81c3-413f-4a2c-9c5f-220f412b2998\") " Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.048177 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71af81c3-413f-4a2c-9c5f-220f412b2998-utilities\") pod \"71af81c3-413f-4a2c-9c5f-220f412b2998\" (UID: \"71af81c3-413f-4a2c-9c5f-220f412b2998\") " Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.049064 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71af81c3-413f-4a2c-9c5f-220f412b2998-utilities" (OuterVolumeSpecName: "utilities") pod "71af81c3-413f-4a2c-9c5f-220f412b2998" (UID: "71af81c3-413f-4a2c-9c5f-220f412b2998"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.062202 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71af81c3-413f-4a2c-9c5f-220f412b2998-kube-api-access-dsxhd" (OuterVolumeSpecName: "kube-api-access-dsxhd") pod "71af81c3-413f-4a2c-9c5f-220f412b2998" (UID: "71af81c3-413f-4a2c-9c5f-220f412b2998"). InnerVolumeSpecName "kube-api-access-dsxhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.099742 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71af81c3-413f-4a2c-9c5f-220f412b2998-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "71af81c3-413f-4a2c-9c5f-220f412b2998" (UID: "71af81c3-413f-4a2c-9c5f-220f412b2998"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.151610 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsxhd\" (UniqueName: \"kubernetes.io/projected/71af81c3-413f-4a2c-9c5f-220f412b2998-kube-api-access-dsxhd\") on node \"crc\" DevicePath \"\"" Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.151653 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71af81c3-413f-4a2c-9c5f-220f412b2998-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.151667 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71af81c3-413f-4a2c-9c5f-220f412b2998-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.491072 4810 generic.go:334] "Generic (PLEG): container finished" podID="71af81c3-413f-4a2c-9c5f-220f412b2998" containerID="a2e3718e8a3ff5bb29bf2c9ebff1f76b6c219c99d1cedc5f77dd90ef0de55b28" exitCode=0 Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.491116 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vqd54" Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.491120 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqd54" event={"ID":"71af81c3-413f-4a2c-9c5f-220f412b2998","Type":"ContainerDied","Data":"a2e3718e8a3ff5bb29bf2c9ebff1f76b6c219c99d1cedc5f77dd90ef0de55b28"} Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.491160 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vqd54" event={"ID":"71af81c3-413f-4a2c-9c5f-220f412b2998","Type":"ContainerDied","Data":"414dedd59061a8ab97aedce9562710ec34d90be360bc56ded41b7a61b90f9c25"} Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.491200 4810 scope.go:117] "RemoveContainer" containerID="a2e3718e8a3ff5bb29bf2c9ebff1f76b6c219c99d1cedc5f77dd90ef0de55b28" Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.525281 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vqd54"] Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.528807 4810 scope.go:117] "RemoveContainer" containerID="5048f7c6b2aaa906ec16f2b423f33337907dbf6de3e427fc84c509c9471afdac" Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.535275 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vqd54"] Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.561946 4810 scope.go:117] "RemoveContainer" containerID="d70b13e2f828295eb256f5605dc37c39afa73b9477b46c8d304e44c2257e4f97" Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.622177 4810 scope.go:117] "RemoveContainer" containerID="a2e3718e8a3ff5bb29bf2c9ebff1f76b6c219c99d1cedc5f77dd90ef0de55b28" Dec 01 16:47:55 crc kubenswrapper[4810]: E1201 16:47:55.622827 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2e3718e8a3ff5bb29bf2c9ebff1f76b6c219c99d1cedc5f77dd90ef0de55b28\": container with ID starting with a2e3718e8a3ff5bb29bf2c9ebff1f76b6c219c99d1cedc5f77dd90ef0de55b28 not found: ID does not exist" containerID="a2e3718e8a3ff5bb29bf2c9ebff1f76b6c219c99d1cedc5f77dd90ef0de55b28" Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.622868 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2e3718e8a3ff5bb29bf2c9ebff1f76b6c219c99d1cedc5f77dd90ef0de55b28"} err="failed to get container status \"a2e3718e8a3ff5bb29bf2c9ebff1f76b6c219c99d1cedc5f77dd90ef0de55b28\": rpc error: code = NotFound desc = could not find container \"a2e3718e8a3ff5bb29bf2c9ebff1f76b6c219c99d1cedc5f77dd90ef0de55b28\": container with ID starting with a2e3718e8a3ff5bb29bf2c9ebff1f76b6c219c99d1cedc5f77dd90ef0de55b28 not found: ID does not exist" Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.622901 4810 scope.go:117] "RemoveContainer" containerID="5048f7c6b2aaa906ec16f2b423f33337907dbf6de3e427fc84c509c9471afdac" Dec 01 16:47:55 crc kubenswrapper[4810]: E1201 16:47:55.623252 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5048f7c6b2aaa906ec16f2b423f33337907dbf6de3e427fc84c509c9471afdac\": container with ID starting with 5048f7c6b2aaa906ec16f2b423f33337907dbf6de3e427fc84c509c9471afdac not found: ID does not exist" containerID="5048f7c6b2aaa906ec16f2b423f33337907dbf6de3e427fc84c509c9471afdac" Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.623284 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5048f7c6b2aaa906ec16f2b423f33337907dbf6de3e427fc84c509c9471afdac"} err="failed to get container status \"5048f7c6b2aaa906ec16f2b423f33337907dbf6de3e427fc84c509c9471afdac\": rpc error: code = NotFound desc = could not find container \"5048f7c6b2aaa906ec16f2b423f33337907dbf6de3e427fc84c509c9471afdac\": container with ID starting with 5048f7c6b2aaa906ec16f2b423f33337907dbf6de3e427fc84c509c9471afdac not found: ID does not exist" Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.623302 4810 scope.go:117] "RemoveContainer" containerID="d70b13e2f828295eb256f5605dc37c39afa73b9477b46c8d304e44c2257e4f97" Dec 01 16:47:55 crc kubenswrapper[4810]: E1201 16:47:55.623762 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d70b13e2f828295eb256f5605dc37c39afa73b9477b46c8d304e44c2257e4f97\": container with ID starting with d70b13e2f828295eb256f5605dc37c39afa73b9477b46c8d304e44c2257e4f97 not found: ID does not exist" containerID="d70b13e2f828295eb256f5605dc37c39afa73b9477b46c8d304e44c2257e4f97" Dec 01 16:47:55 crc kubenswrapper[4810]: I1201 16:47:55.623792 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d70b13e2f828295eb256f5605dc37c39afa73b9477b46c8d304e44c2257e4f97"} err="failed to get container status \"d70b13e2f828295eb256f5605dc37c39afa73b9477b46c8d304e44c2257e4f97\": rpc error: code = NotFound desc = could not find container \"d70b13e2f828295eb256f5605dc37c39afa73b9477b46c8d304e44c2257e4f97\": container with ID starting with d70b13e2f828295eb256f5605dc37c39afa73b9477b46c8d304e44c2257e4f97 not found: ID does not exist" Dec 01 16:47:56 crc kubenswrapper[4810]: I1201 16:47:56.522771 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71af81c3-413f-4a2c-9c5f-220f412b2998" path="/var/lib/kubelet/pods/71af81c3-413f-4a2c-9c5f-220f412b2998/volumes" Dec 01 16:48:08 crc kubenswrapper[4810]: I1201 16:48:08.652210 4810 generic.go:334] "Generic (PLEG): container finished" podID="17bffe79-728c-4500-a27e-a2b246f82517" containerID="8346b581e46a6b78db0a34224ac6e0ef6caa4c783bf115c905ce6075c8213994" exitCode=0 Dec 01 16:48:08 crc kubenswrapper[4810]: I1201 16:48:08.652297 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" event={"ID":"17bffe79-728c-4500-a27e-a2b246f82517","Type":"ContainerDied","Data":"8346b581e46a6b78db0a34224ac6e0ef6caa4c783bf115c905ce6075c8213994"} Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.130418 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.231113 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-libvirt-combined-ca-bundle\") pod \"17bffe79-728c-4500-a27e-a2b246f82517\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.231168 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-neutron-dhcp-combined-ca-bundle\") pod \"17bffe79-728c-4500-a27e-a2b246f82517\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.231236 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-libvirt-default-certs-0\") pod \"17bffe79-728c-4500-a27e-a2b246f82517\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.231271 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-neutron-metadata-combined-ca-bundle\") pod \"17bffe79-728c-4500-a27e-a2b246f82517\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.231350 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-nova-combined-ca-bundle\") pod \"17bffe79-728c-4500-a27e-a2b246f82517\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.231368 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-inventory\") pod \"17bffe79-728c-4500-a27e-a2b246f82517\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.231388 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-neutron-sriov-combined-ca-bundle\") pod \"17bffe79-728c-4500-a27e-a2b246f82517\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.231404 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-ovn-default-certs-0\") pod \"17bffe79-728c-4500-a27e-a2b246f82517\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.231422 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-ssh-key\") pod \"17bffe79-728c-4500-a27e-a2b246f82517\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.231437 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-neutron-metadata-default-certs-0\") pod \"17bffe79-728c-4500-a27e-a2b246f82517\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.231460 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-telemetry-default-certs-0\") pod \"17bffe79-728c-4500-a27e-a2b246f82517\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.231523 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-telemetry-combined-ca-bundle\") pod \"17bffe79-728c-4500-a27e-a2b246f82517\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.231547 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-bootstrap-combined-ca-bundle\") pod \"17bffe79-728c-4500-a27e-a2b246f82517\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.231607 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-ovn-combined-ca-bundle\") pod \"17bffe79-728c-4500-a27e-a2b246f82517\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.231697 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5jws\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-kube-api-access-l5jws\") pod \"17bffe79-728c-4500-a27e-a2b246f82517\" (UID: \"17bffe79-728c-4500-a27e-a2b246f82517\") " Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.237461 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-kube-api-access-l5jws" (OuterVolumeSpecName: "kube-api-access-l5jws") pod "17bffe79-728c-4500-a27e-a2b246f82517" (UID: "17bffe79-728c-4500-a27e-a2b246f82517"). InnerVolumeSpecName "kube-api-access-l5jws". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.239043 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-libvirt-default-certs-0") pod "17bffe79-728c-4500-a27e-a2b246f82517" (UID: "17bffe79-728c-4500-a27e-a2b246f82517"). InnerVolumeSpecName "openstack-cell1-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.239638 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "17bffe79-728c-4500-a27e-a2b246f82517" (UID: "17bffe79-728c-4500-a27e-a2b246f82517"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.239793 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "17bffe79-728c-4500-a27e-a2b246f82517" (UID: "17bffe79-728c-4500-a27e-a2b246f82517"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.239829 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "17bffe79-728c-4500-a27e-a2b246f82517" (UID: "17bffe79-728c-4500-a27e-a2b246f82517"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.239885 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "17bffe79-728c-4500-a27e-a2b246f82517" (UID: "17bffe79-728c-4500-a27e-a2b246f82517"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.240219 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "17bffe79-728c-4500-a27e-a2b246f82517" (UID: "17bffe79-728c-4500-a27e-a2b246f82517"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.240891 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "17bffe79-728c-4500-a27e-a2b246f82517" (UID: "17bffe79-728c-4500-a27e-a2b246f82517"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.241692 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "17bffe79-728c-4500-a27e-a2b246f82517" (UID: "17bffe79-728c-4500-a27e-a2b246f82517"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.242069 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "17bffe79-728c-4500-a27e-a2b246f82517" (UID: "17bffe79-728c-4500-a27e-a2b246f82517"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.242163 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-neutron-metadata-default-certs-0") pod "17bffe79-728c-4500-a27e-a2b246f82517" (UID: "17bffe79-728c-4500-a27e-a2b246f82517"). InnerVolumeSpecName "openstack-cell1-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.245673 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-ovn-default-certs-0") pod "17bffe79-728c-4500-a27e-a2b246f82517" (UID: "17bffe79-728c-4500-a27e-a2b246f82517"). InnerVolumeSpecName "openstack-cell1-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.252856 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-telemetry-default-certs-0") pod "17bffe79-728c-4500-a27e-a2b246f82517" (UID: "17bffe79-728c-4500-a27e-a2b246f82517"). InnerVolumeSpecName "openstack-cell1-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.266613 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-inventory" (OuterVolumeSpecName: "inventory") pod "17bffe79-728c-4500-a27e-a2b246f82517" (UID: "17bffe79-728c-4500-a27e-a2b246f82517"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.267016 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "17bffe79-728c-4500-a27e-a2b246f82517" (UID: "17bffe79-728c-4500-a27e-a2b246f82517"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.335619 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5jws\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-kube-api-access-l5jws\") on node \"crc\" DevicePath \"\"" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.335673 4810 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.335690 4810 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.335725 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.335741 4810 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.335753 4810 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.335767 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.335779 4810 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.335810 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.335823 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.335840 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.335852 4810 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/17bffe79-728c-4500-a27e-a2b246f82517-openstack-cell1-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.335885 4810 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.335900 4810 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.335912 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17bffe79-728c-4500-a27e-a2b246f82517-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.673399 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" event={"ID":"17bffe79-728c-4500-a27e-a2b246f82517","Type":"ContainerDied","Data":"73d803a0d0bc679294517a01ec0fd01a0631dfc10547865863889465df9287ca"} Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.673447 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73d803a0d0bc679294517a01ec0fd01a0631dfc10547865863889465df9287ca" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.673540 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-sr2t8" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.776540 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-zsqk6"] Dec 01 16:48:10 crc kubenswrapper[4810]: E1201 16:48:10.777185 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71af81c3-413f-4a2c-9c5f-220f412b2998" containerName="extract-utilities" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.777209 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="71af81c3-413f-4a2c-9c5f-220f412b2998" containerName="extract-utilities" Dec 01 16:48:10 crc kubenswrapper[4810]: E1201 16:48:10.777242 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17bffe79-728c-4500-a27e-a2b246f82517" containerName="install-certs-openstack-openstack-cell1" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.777251 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="17bffe79-728c-4500-a27e-a2b246f82517" containerName="install-certs-openstack-openstack-cell1" Dec 01 16:48:10 crc kubenswrapper[4810]: E1201 16:48:10.777296 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71af81c3-413f-4a2c-9c5f-220f412b2998" containerName="registry-server" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.777304 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="71af81c3-413f-4a2c-9c5f-220f412b2998" containerName="registry-server" Dec 01 16:48:10 crc kubenswrapper[4810]: E1201 16:48:10.777324 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71af81c3-413f-4a2c-9c5f-220f412b2998" containerName="extract-content" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.777333 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="71af81c3-413f-4a2c-9c5f-220f412b2998" containerName="extract-content" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.777601 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="71af81c3-413f-4a2c-9c5f-220f412b2998" containerName="registry-server" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.777649 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="17bffe79-728c-4500-a27e-a2b246f82517" containerName="install-certs-openstack-openstack-cell1" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.778672 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.781193 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.781324 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.781588 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.781745 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.781989 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.787696 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-zsqk6"] Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.950222 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1252894-fad3-4959-99b1-5564dff922ec-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-zsqk6\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.950492 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x7n2\" (UniqueName: \"kubernetes.io/projected/e1252894-fad3-4959-99b1-5564dff922ec-kube-api-access-8x7n2\") pod \"ovn-openstack-openstack-cell1-zsqk6\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.950518 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e1252894-fad3-4959-99b1-5564dff922ec-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-zsqk6\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.950539 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1252894-fad3-4959-99b1-5564dff922ec-ssh-key\") pod \"ovn-openstack-openstack-cell1-zsqk6\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:48:10 crc kubenswrapper[4810]: I1201 16:48:10.950566 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1252894-fad3-4959-99b1-5564dff922ec-inventory\") pod \"ovn-openstack-openstack-cell1-zsqk6\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:48:11 crc kubenswrapper[4810]: I1201 16:48:11.053350 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1252894-fad3-4959-99b1-5564dff922ec-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-zsqk6\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:48:11 crc kubenswrapper[4810]: I1201 16:48:11.053893 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x7n2\" (UniqueName: \"kubernetes.io/projected/e1252894-fad3-4959-99b1-5564dff922ec-kube-api-access-8x7n2\") pod \"ovn-openstack-openstack-cell1-zsqk6\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:48:11 crc kubenswrapper[4810]: I1201 16:48:11.054029 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e1252894-fad3-4959-99b1-5564dff922ec-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-zsqk6\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:48:11 crc kubenswrapper[4810]: I1201 16:48:11.054147 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1252894-fad3-4959-99b1-5564dff922ec-ssh-key\") pod \"ovn-openstack-openstack-cell1-zsqk6\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:48:11 crc kubenswrapper[4810]: I1201 16:48:11.054300 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1252894-fad3-4959-99b1-5564dff922ec-inventory\") pod \"ovn-openstack-openstack-cell1-zsqk6\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:48:11 crc kubenswrapper[4810]: I1201 16:48:11.055319 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e1252894-fad3-4959-99b1-5564dff922ec-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-zsqk6\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:48:11 crc kubenswrapper[4810]: I1201 16:48:11.057347 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1252894-fad3-4959-99b1-5564dff922ec-ssh-key\") pod \"ovn-openstack-openstack-cell1-zsqk6\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:48:11 crc kubenswrapper[4810]: I1201 16:48:11.058203 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1252894-fad3-4959-99b1-5564dff922ec-inventory\") pod \"ovn-openstack-openstack-cell1-zsqk6\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:48:11 crc kubenswrapper[4810]: I1201 16:48:11.058664 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1252894-fad3-4959-99b1-5564dff922ec-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-zsqk6\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:48:11 crc kubenswrapper[4810]: I1201 16:48:11.071248 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x7n2\" (UniqueName: \"kubernetes.io/projected/e1252894-fad3-4959-99b1-5564dff922ec-kube-api-access-8x7n2\") pod \"ovn-openstack-openstack-cell1-zsqk6\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:48:11 crc kubenswrapper[4810]: I1201 16:48:11.103490 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:48:11 crc kubenswrapper[4810]: I1201 16:48:11.613329 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-zsqk6"] Dec 01 16:48:11 crc kubenswrapper[4810]: I1201 16:48:11.683479 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-zsqk6" event={"ID":"e1252894-fad3-4959-99b1-5564dff922ec","Type":"ContainerStarted","Data":"bd555cb3c26cd5c2f2465a96451af883611f8687aef69f5afd8588001ca4a52a"} Dec 01 16:48:12 crc kubenswrapper[4810]: I1201 16:48:12.696045 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-zsqk6" event={"ID":"e1252894-fad3-4959-99b1-5564dff922ec","Type":"ContainerStarted","Data":"33983162c20a36ea54c6ef14fe05e0ff46fa56e79e9168b61cb1214a85cf9a4b"} Dec 01 16:48:12 crc kubenswrapper[4810]: I1201 16:48:12.717343 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-zsqk6" podStartSLOduration=2.234470568 podStartE2EDuration="2.717325608s" podCreationTimestamp="2025-12-01 16:48:10 +0000 UTC" firstStartedPulling="2025-12-01 16:48:11.622995494 +0000 UTC m=+8057.386505097" lastFinishedPulling="2025-12-01 16:48:12.105850524 +0000 UTC m=+8057.869360137" observedRunningTime="2025-12-01 16:48:12.716813523 +0000 UTC m=+8058.480323126" watchObservedRunningTime="2025-12-01 16:48:12.717325608 +0000 UTC m=+8058.480835211" Dec 01 16:49:02 crc kubenswrapper[4810]: I1201 16:49:02.971983 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:49:02 crc kubenswrapper[4810]: I1201 16:49:02.972552 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:49:19 crc kubenswrapper[4810]: I1201 16:49:19.381306 4810 generic.go:334] "Generic (PLEG): container finished" podID="e1252894-fad3-4959-99b1-5564dff922ec" containerID="33983162c20a36ea54c6ef14fe05e0ff46fa56e79e9168b61cb1214a85cf9a4b" exitCode=0 Dec 01 16:49:19 crc kubenswrapper[4810]: I1201 16:49:19.381404 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-zsqk6" event={"ID":"e1252894-fad3-4959-99b1-5564dff922ec","Type":"ContainerDied","Data":"33983162c20a36ea54c6ef14fe05e0ff46fa56e79e9168b61cb1214a85cf9a4b"} Dec 01 16:49:20 crc kubenswrapper[4810]: I1201 16:49:20.840502 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:49:20 crc kubenswrapper[4810]: I1201 16:49:20.883219 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1252894-fad3-4959-99b1-5564dff922ec-ovn-combined-ca-bundle\") pod \"e1252894-fad3-4959-99b1-5564dff922ec\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " Dec 01 16:49:20 crc kubenswrapper[4810]: I1201 16:49:20.883348 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1252894-fad3-4959-99b1-5564dff922ec-ssh-key\") pod \"e1252894-fad3-4959-99b1-5564dff922ec\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " Dec 01 16:49:20 crc kubenswrapper[4810]: I1201 16:49:20.883397 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e1252894-fad3-4959-99b1-5564dff922ec-ovncontroller-config-0\") pod \"e1252894-fad3-4959-99b1-5564dff922ec\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " Dec 01 16:49:20 crc kubenswrapper[4810]: I1201 16:49:20.883495 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1252894-fad3-4959-99b1-5564dff922ec-inventory\") pod \"e1252894-fad3-4959-99b1-5564dff922ec\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " Dec 01 16:49:20 crc kubenswrapper[4810]: I1201 16:49:20.883637 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8x7n2\" (UniqueName: \"kubernetes.io/projected/e1252894-fad3-4959-99b1-5564dff922ec-kube-api-access-8x7n2\") pod \"e1252894-fad3-4959-99b1-5564dff922ec\" (UID: \"e1252894-fad3-4959-99b1-5564dff922ec\") " Dec 01 16:49:20 crc kubenswrapper[4810]: I1201 16:49:20.899579 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1252894-fad3-4959-99b1-5564dff922ec-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "e1252894-fad3-4959-99b1-5564dff922ec" (UID: "e1252894-fad3-4959-99b1-5564dff922ec"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:49:20 crc kubenswrapper[4810]: I1201 16:49:20.902750 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1252894-fad3-4959-99b1-5564dff922ec-kube-api-access-8x7n2" (OuterVolumeSpecName: "kube-api-access-8x7n2") pod "e1252894-fad3-4959-99b1-5564dff922ec" (UID: "e1252894-fad3-4959-99b1-5564dff922ec"). InnerVolumeSpecName "kube-api-access-8x7n2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:49:20 crc kubenswrapper[4810]: I1201 16:49:20.963344 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1252894-fad3-4959-99b1-5564dff922ec-inventory" (OuterVolumeSpecName: "inventory") pod "e1252894-fad3-4959-99b1-5564dff922ec" (UID: "e1252894-fad3-4959-99b1-5564dff922ec"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:49:20 crc kubenswrapper[4810]: I1201 16:49:20.983388 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1252894-fad3-4959-99b1-5564dff922ec-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "e1252894-fad3-4959-99b1-5564dff922ec" (UID: "e1252894-fad3-4959-99b1-5564dff922ec"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:49:20 crc kubenswrapper[4810]: I1201 16:49:20.985822 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1252894-fad3-4959-99b1-5564dff922ec-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e1252894-fad3-4959-99b1-5564dff922ec" (UID: "e1252894-fad3-4959-99b1-5564dff922ec"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:49:20 crc kubenswrapper[4810]: I1201 16:49:20.986397 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8x7n2\" (UniqueName: \"kubernetes.io/projected/e1252894-fad3-4959-99b1-5564dff922ec-kube-api-access-8x7n2\") on node \"crc\" DevicePath \"\"" Dec 01 16:49:20 crc kubenswrapper[4810]: I1201 16:49:20.986444 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1252894-fad3-4959-99b1-5564dff922ec-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:49:20 crc kubenswrapper[4810]: I1201 16:49:20.986457 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1252894-fad3-4959-99b1-5564dff922ec-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:49:20 crc kubenswrapper[4810]: I1201 16:49:20.986493 4810 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e1252894-fad3-4959-99b1-5564dff922ec-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:49:20 crc kubenswrapper[4810]: I1201 16:49:20.986506 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1252894-fad3-4959-99b1-5564dff922ec-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.402743 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-zsqk6" event={"ID":"e1252894-fad3-4959-99b1-5564dff922ec","Type":"ContainerDied","Data":"bd555cb3c26cd5c2f2465a96451af883611f8687aef69f5afd8588001ca4a52a"} Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.402791 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd555cb3c26cd5c2f2465a96451af883611f8687aef69f5afd8588001ca4a52a" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.402833 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-zsqk6" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.513214 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-xpffb"] Dec 01 16:49:21 crc kubenswrapper[4810]: E1201 16:49:21.513732 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1252894-fad3-4959-99b1-5564dff922ec" containerName="ovn-openstack-openstack-cell1" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.513750 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1252894-fad3-4959-99b1-5564dff922ec" containerName="ovn-openstack-openstack-cell1" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.513933 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1252894-fad3-4959-99b1-5564dff922ec" containerName="ovn-openstack-openstack-cell1" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.514693 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.516922 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.517124 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.517243 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.518909 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.519059 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.519289 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.525019 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-xpffb"] Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.597334 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld88r\" (UniqueName: \"kubernetes.io/projected/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-kube-api-access-ld88r\") pod \"neutron-metadata-openstack-openstack-cell1-xpffb\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.597490 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-xpffb\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.597976 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xpffb\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.598048 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xpffb\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.598087 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-xpffb\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.598117 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-xpffb\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.699663 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-xpffb\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.700086 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xpffb\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.700171 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xpffb\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.700215 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-xpffb\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.700279 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-xpffb\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.700425 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld88r\" (UniqueName: \"kubernetes.io/projected/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-kube-api-access-ld88r\") pod \"neutron-metadata-openstack-openstack-cell1-xpffb\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.706770 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-xpffb\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.707525 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xpffb\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.707898 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-xpffb\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.708291 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xpffb\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.710234 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-xpffb\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.718396 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld88r\" (UniqueName: \"kubernetes.io/projected/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-kube-api-access-ld88r\") pod \"neutron-metadata-openstack-openstack-cell1-xpffb\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:21 crc kubenswrapper[4810]: I1201 16:49:21.836880 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:49:22 crc kubenswrapper[4810]: I1201 16:49:22.371080 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-xpffb"] Dec 01 16:49:22 crc kubenswrapper[4810]: I1201 16:49:22.415352 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" event={"ID":"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a","Type":"ContainerStarted","Data":"3bd9ed4270384baecea41cdd373913112d4197dda5654cbdcf097cce1b648ac3"} Dec 01 16:49:23 crc kubenswrapper[4810]: I1201 16:49:23.426872 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" event={"ID":"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a","Type":"ContainerStarted","Data":"5f9ff8a7e9563c85d245c2fdd47cd54a087d14efa3e536a7caa31eb153f6fab1"} Dec 01 16:49:23 crc kubenswrapper[4810]: I1201 16:49:23.458238 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" podStartSLOduration=1.943774135 podStartE2EDuration="2.45821514s" podCreationTimestamp="2025-12-01 16:49:21 +0000 UTC" firstStartedPulling="2025-12-01 16:49:22.37287029 +0000 UTC m=+8128.136379893" lastFinishedPulling="2025-12-01 16:49:22.887311295 +0000 UTC m=+8128.650820898" observedRunningTime="2025-12-01 16:49:23.446149493 +0000 UTC m=+8129.209659096" watchObservedRunningTime="2025-12-01 16:49:23.45821514 +0000 UTC m=+8129.221724743" Dec 01 16:49:26 crc kubenswrapper[4810]: I1201 16:49:26.521426 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gb9g8"] Dec 01 16:49:26 crc kubenswrapper[4810]: I1201 16:49:26.525160 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gb9g8" Dec 01 16:49:26 crc kubenswrapper[4810]: I1201 16:49:26.541692 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gb9g8"] Dec 01 16:49:26 crc kubenswrapper[4810]: I1201 16:49:26.719607 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/897a5bed-90d1-409a-b986-ad6fb9587df6-utilities\") pod \"redhat-operators-gb9g8\" (UID: \"897a5bed-90d1-409a-b986-ad6fb9587df6\") " pod="openshift-marketplace/redhat-operators-gb9g8" Dec 01 16:49:26 crc kubenswrapper[4810]: I1201 16:49:26.720106 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/897a5bed-90d1-409a-b986-ad6fb9587df6-catalog-content\") pod \"redhat-operators-gb9g8\" (UID: \"897a5bed-90d1-409a-b986-ad6fb9587df6\") " pod="openshift-marketplace/redhat-operators-gb9g8" Dec 01 16:49:26 crc kubenswrapper[4810]: I1201 16:49:26.720270 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8l2s\" (UniqueName: \"kubernetes.io/projected/897a5bed-90d1-409a-b986-ad6fb9587df6-kube-api-access-m8l2s\") pod \"redhat-operators-gb9g8\" (UID: \"897a5bed-90d1-409a-b986-ad6fb9587df6\") " pod="openshift-marketplace/redhat-operators-gb9g8" Dec 01 16:49:26 crc kubenswrapper[4810]: I1201 16:49:26.823139 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8l2s\" (UniqueName: \"kubernetes.io/projected/897a5bed-90d1-409a-b986-ad6fb9587df6-kube-api-access-m8l2s\") pod \"redhat-operators-gb9g8\" (UID: \"897a5bed-90d1-409a-b986-ad6fb9587df6\") " pod="openshift-marketplace/redhat-operators-gb9g8" Dec 01 16:49:26 crc kubenswrapper[4810]: I1201 16:49:26.823315 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/897a5bed-90d1-409a-b986-ad6fb9587df6-utilities\") pod \"redhat-operators-gb9g8\" (UID: \"897a5bed-90d1-409a-b986-ad6fb9587df6\") " pod="openshift-marketplace/redhat-operators-gb9g8" Dec 01 16:49:26 crc kubenswrapper[4810]: I1201 16:49:26.823363 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/897a5bed-90d1-409a-b986-ad6fb9587df6-catalog-content\") pod \"redhat-operators-gb9g8\" (UID: \"897a5bed-90d1-409a-b986-ad6fb9587df6\") " pod="openshift-marketplace/redhat-operators-gb9g8" Dec 01 16:49:26 crc kubenswrapper[4810]: I1201 16:49:26.823824 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/897a5bed-90d1-409a-b986-ad6fb9587df6-utilities\") pod \"redhat-operators-gb9g8\" (UID: \"897a5bed-90d1-409a-b986-ad6fb9587df6\") " pod="openshift-marketplace/redhat-operators-gb9g8" Dec 01 16:49:26 crc kubenswrapper[4810]: I1201 16:49:26.823856 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/897a5bed-90d1-409a-b986-ad6fb9587df6-catalog-content\") pod \"redhat-operators-gb9g8\" (UID: \"897a5bed-90d1-409a-b986-ad6fb9587df6\") " pod="openshift-marketplace/redhat-operators-gb9g8" Dec 01 16:49:26 crc kubenswrapper[4810]: I1201 16:49:26.846821 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8l2s\" (UniqueName: \"kubernetes.io/projected/897a5bed-90d1-409a-b986-ad6fb9587df6-kube-api-access-m8l2s\") pod \"redhat-operators-gb9g8\" (UID: \"897a5bed-90d1-409a-b986-ad6fb9587df6\") " pod="openshift-marketplace/redhat-operators-gb9g8" Dec 01 16:49:26 crc kubenswrapper[4810]: I1201 16:49:26.860889 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gb9g8" Dec 01 16:49:27 crc kubenswrapper[4810]: I1201 16:49:27.334968 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gb9g8"] Dec 01 16:49:27 crc kubenswrapper[4810]: I1201 16:49:27.465176 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gb9g8" event={"ID":"897a5bed-90d1-409a-b986-ad6fb9587df6","Type":"ContainerStarted","Data":"037ff318093c8da228b74cc43abb2b45ad4274dd9c323dd578946b8d375884df"} Dec 01 16:49:28 crc kubenswrapper[4810]: I1201 16:49:28.478725 4810 generic.go:334] "Generic (PLEG): container finished" podID="897a5bed-90d1-409a-b986-ad6fb9587df6" containerID="234a40f8d218315aceffd700cc2c1870e9744fc4c39c67afab77663313aa4679" exitCode=0 Dec 01 16:49:28 crc kubenswrapper[4810]: I1201 16:49:28.479052 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gb9g8" event={"ID":"897a5bed-90d1-409a-b986-ad6fb9587df6","Type":"ContainerDied","Data":"234a40f8d218315aceffd700cc2c1870e9744fc4c39c67afab77663313aa4679"} Dec 01 16:49:30 crc kubenswrapper[4810]: I1201 16:49:30.503954 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gb9g8" event={"ID":"897a5bed-90d1-409a-b986-ad6fb9587df6","Type":"ContainerStarted","Data":"3c056a52fa3030cf7f657ff63c8b982a6e8783c03a3072694c3a7dc74a4e02bb"} Dec 01 16:49:32 crc kubenswrapper[4810]: I1201 16:49:32.527737 4810 generic.go:334] "Generic (PLEG): container finished" podID="897a5bed-90d1-409a-b986-ad6fb9587df6" containerID="3c056a52fa3030cf7f657ff63c8b982a6e8783c03a3072694c3a7dc74a4e02bb" exitCode=0 Dec 01 16:49:32 crc kubenswrapper[4810]: I1201 16:49:32.527828 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gb9g8" event={"ID":"897a5bed-90d1-409a-b986-ad6fb9587df6","Type":"ContainerDied","Data":"3c056a52fa3030cf7f657ff63c8b982a6e8783c03a3072694c3a7dc74a4e02bb"} Dec 01 16:49:32 crc kubenswrapper[4810]: I1201 16:49:32.971902 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:49:32 crc kubenswrapper[4810]: I1201 16:49:32.971977 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:49:34 crc kubenswrapper[4810]: I1201 16:49:34.553001 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gb9g8" event={"ID":"897a5bed-90d1-409a-b986-ad6fb9587df6","Type":"ContainerStarted","Data":"a37c20bc64237e4c58a842cf25466c2d7a85bc0481f837271dec30e66c26ca26"} Dec 01 16:49:34 crc kubenswrapper[4810]: I1201 16:49:34.582130 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gb9g8" podStartSLOduration=3.638568067 podStartE2EDuration="8.582104301s" podCreationTimestamp="2025-12-01 16:49:26 +0000 UTC" firstStartedPulling="2025-12-01 16:49:28.484974938 +0000 UTC m=+8134.248484541" lastFinishedPulling="2025-12-01 16:49:33.428511162 +0000 UTC m=+8139.192020775" observedRunningTime="2025-12-01 16:49:34.571947876 +0000 UTC m=+8140.335457479" watchObservedRunningTime="2025-12-01 16:49:34.582104301 +0000 UTC m=+8140.345613934" Dec 01 16:49:36 crc kubenswrapper[4810]: I1201 16:49:36.861261 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gb9g8" Dec 01 16:49:36 crc kubenswrapper[4810]: I1201 16:49:36.862016 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gb9g8" Dec 01 16:49:37 crc kubenswrapper[4810]: I1201 16:49:37.940912 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gb9g8" podUID="897a5bed-90d1-409a-b986-ad6fb9587df6" containerName="registry-server" probeResult="failure" output=< Dec 01 16:49:37 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Dec 01 16:49:37 crc kubenswrapper[4810]: > Dec 01 16:49:46 crc kubenswrapper[4810]: I1201 16:49:46.922772 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gb9g8" Dec 01 16:49:46 crc kubenswrapper[4810]: I1201 16:49:46.992400 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gb9g8" Dec 01 16:49:47 crc kubenswrapper[4810]: I1201 16:49:47.169450 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gb9g8"] Dec 01 16:49:48 crc kubenswrapper[4810]: I1201 16:49:48.718316 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gb9g8" podUID="897a5bed-90d1-409a-b986-ad6fb9587df6" containerName="registry-server" containerID="cri-o://a37c20bc64237e4c58a842cf25466c2d7a85bc0481f837271dec30e66c26ca26" gracePeriod=2 Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.268146 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gb9g8" Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.380000 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8l2s\" (UniqueName: \"kubernetes.io/projected/897a5bed-90d1-409a-b986-ad6fb9587df6-kube-api-access-m8l2s\") pod \"897a5bed-90d1-409a-b986-ad6fb9587df6\" (UID: \"897a5bed-90d1-409a-b986-ad6fb9587df6\") " Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.380797 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/897a5bed-90d1-409a-b986-ad6fb9587df6-utilities\") pod \"897a5bed-90d1-409a-b986-ad6fb9587df6\" (UID: \"897a5bed-90d1-409a-b986-ad6fb9587df6\") " Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.381002 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/897a5bed-90d1-409a-b986-ad6fb9587df6-catalog-content\") pod \"897a5bed-90d1-409a-b986-ad6fb9587df6\" (UID: \"897a5bed-90d1-409a-b986-ad6fb9587df6\") " Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.381707 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/897a5bed-90d1-409a-b986-ad6fb9587df6-utilities" (OuterVolumeSpecName: "utilities") pod "897a5bed-90d1-409a-b986-ad6fb9587df6" (UID: "897a5bed-90d1-409a-b986-ad6fb9587df6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.391154 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/897a5bed-90d1-409a-b986-ad6fb9587df6-kube-api-access-m8l2s" (OuterVolumeSpecName: "kube-api-access-m8l2s") pod "897a5bed-90d1-409a-b986-ad6fb9587df6" (UID: "897a5bed-90d1-409a-b986-ad6fb9587df6"). InnerVolumeSpecName "kube-api-access-m8l2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.476561 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/897a5bed-90d1-409a-b986-ad6fb9587df6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "897a5bed-90d1-409a-b986-ad6fb9587df6" (UID: "897a5bed-90d1-409a-b986-ad6fb9587df6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.483635 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8l2s\" (UniqueName: \"kubernetes.io/projected/897a5bed-90d1-409a-b986-ad6fb9587df6-kube-api-access-m8l2s\") on node \"crc\" DevicePath \"\"" Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.483669 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/897a5bed-90d1-409a-b986-ad6fb9587df6-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.483681 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/897a5bed-90d1-409a-b986-ad6fb9587df6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.727520 4810 generic.go:334] "Generic (PLEG): container finished" podID="897a5bed-90d1-409a-b986-ad6fb9587df6" containerID="a37c20bc64237e4c58a842cf25466c2d7a85bc0481f837271dec30e66c26ca26" exitCode=0 Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.727571 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gb9g8" event={"ID":"897a5bed-90d1-409a-b986-ad6fb9587df6","Type":"ContainerDied","Data":"a37c20bc64237e4c58a842cf25466c2d7a85bc0481f837271dec30e66c26ca26"} Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.727611 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gb9g8" event={"ID":"897a5bed-90d1-409a-b986-ad6fb9587df6","Type":"ContainerDied","Data":"037ff318093c8da228b74cc43abb2b45ad4274dd9c323dd578946b8d375884df"} Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.727632 4810 scope.go:117] "RemoveContainer" containerID="a37c20bc64237e4c58a842cf25466c2d7a85bc0481f837271dec30e66c26ca26" Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.727629 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gb9g8" Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.751106 4810 scope.go:117] "RemoveContainer" containerID="3c056a52fa3030cf7f657ff63c8b982a6e8783c03a3072694c3a7dc74a4e02bb" Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.772627 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gb9g8"] Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.783840 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gb9g8"] Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.788033 4810 scope.go:117] "RemoveContainer" containerID="234a40f8d218315aceffd700cc2c1870e9744fc4c39c67afab77663313aa4679" Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.865521 4810 scope.go:117] "RemoveContainer" containerID="a37c20bc64237e4c58a842cf25466c2d7a85bc0481f837271dec30e66c26ca26" Dec 01 16:49:49 crc kubenswrapper[4810]: E1201 16:49:49.866037 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a37c20bc64237e4c58a842cf25466c2d7a85bc0481f837271dec30e66c26ca26\": container with ID starting with a37c20bc64237e4c58a842cf25466c2d7a85bc0481f837271dec30e66c26ca26 not found: ID does not exist" containerID="a37c20bc64237e4c58a842cf25466c2d7a85bc0481f837271dec30e66c26ca26" Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.866095 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a37c20bc64237e4c58a842cf25466c2d7a85bc0481f837271dec30e66c26ca26"} err="failed to get container status \"a37c20bc64237e4c58a842cf25466c2d7a85bc0481f837271dec30e66c26ca26\": rpc error: code = NotFound desc = could not find container \"a37c20bc64237e4c58a842cf25466c2d7a85bc0481f837271dec30e66c26ca26\": container with ID starting with a37c20bc64237e4c58a842cf25466c2d7a85bc0481f837271dec30e66c26ca26 not found: ID does not exist" Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.866123 4810 scope.go:117] "RemoveContainer" containerID="3c056a52fa3030cf7f657ff63c8b982a6e8783c03a3072694c3a7dc74a4e02bb" Dec 01 16:49:49 crc kubenswrapper[4810]: E1201 16:49:49.866456 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c056a52fa3030cf7f657ff63c8b982a6e8783c03a3072694c3a7dc74a4e02bb\": container with ID starting with 3c056a52fa3030cf7f657ff63c8b982a6e8783c03a3072694c3a7dc74a4e02bb not found: ID does not exist" containerID="3c056a52fa3030cf7f657ff63c8b982a6e8783c03a3072694c3a7dc74a4e02bb" Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.866524 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c056a52fa3030cf7f657ff63c8b982a6e8783c03a3072694c3a7dc74a4e02bb"} err="failed to get container status \"3c056a52fa3030cf7f657ff63c8b982a6e8783c03a3072694c3a7dc74a4e02bb\": rpc error: code = NotFound desc = could not find container \"3c056a52fa3030cf7f657ff63c8b982a6e8783c03a3072694c3a7dc74a4e02bb\": container with ID starting with 3c056a52fa3030cf7f657ff63c8b982a6e8783c03a3072694c3a7dc74a4e02bb not found: ID does not exist" Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.866577 4810 scope.go:117] "RemoveContainer" containerID="234a40f8d218315aceffd700cc2c1870e9744fc4c39c67afab77663313aa4679" Dec 01 16:49:49 crc kubenswrapper[4810]: E1201 16:49:49.866832 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"234a40f8d218315aceffd700cc2c1870e9744fc4c39c67afab77663313aa4679\": container with ID starting with 234a40f8d218315aceffd700cc2c1870e9744fc4c39c67afab77663313aa4679 not found: ID does not exist" containerID="234a40f8d218315aceffd700cc2c1870e9744fc4c39c67afab77663313aa4679" Dec 01 16:49:49 crc kubenswrapper[4810]: I1201 16:49:49.866857 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"234a40f8d218315aceffd700cc2c1870e9744fc4c39c67afab77663313aa4679"} err="failed to get container status \"234a40f8d218315aceffd700cc2c1870e9744fc4c39c67afab77663313aa4679\": rpc error: code = NotFound desc = could not find container \"234a40f8d218315aceffd700cc2c1870e9744fc4c39c67afab77663313aa4679\": container with ID starting with 234a40f8d218315aceffd700cc2c1870e9744fc4c39c67afab77663313aa4679 not found: ID does not exist" Dec 01 16:49:50 crc kubenswrapper[4810]: I1201 16:49:50.505589 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="897a5bed-90d1-409a-b986-ad6fb9587df6" path="/var/lib/kubelet/pods/897a5bed-90d1-409a-b986-ad6fb9587df6/volumes" Dec 01 16:49:52 crc kubenswrapper[4810]: I1201 16:49:52.572893 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wskxk"] Dec 01 16:49:52 crc kubenswrapper[4810]: E1201 16:49:52.573724 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="897a5bed-90d1-409a-b986-ad6fb9587df6" containerName="extract-utilities" Dec 01 16:49:52 crc kubenswrapper[4810]: I1201 16:49:52.573737 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="897a5bed-90d1-409a-b986-ad6fb9587df6" containerName="extract-utilities" Dec 01 16:49:52 crc kubenswrapper[4810]: E1201 16:49:52.573769 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="897a5bed-90d1-409a-b986-ad6fb9587df6" containerName="extract-content" Dec 01 16:49:52 crc kubenswrapper[4810]: I1201 16:49:52.573775 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="897a5bed-90d1-409a-b986-ad6fb9587df6" containerName="extract-content" Dec 01 16:49:52 crc kubenswrapper[4810]: E1201 16:49:52.573815 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="897a5bed-90d1-409a-b986-ad6fb9587df6" containerName="registry-server" Dec 01 16:49:52 crc kubenswrapper[4810]: I1201 16:49:52.573821 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="897a5bed-90d1-409a-b986-ad6fb9587df6" containerName="registry-server" Dec 01 16:49:52 crc kubenswrapper[4810]: I1201 16:49:52.574040 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="897a5bed-90d1-409a-b986-ad6fb9587df6" containerName="registry-server" Dec 01 16:49:52 crc kubenswrapper[4810]: I1201 16:49:52.575597 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wskxk" Dec 01 16:49:52 crc kubenswrapper[4810]: I1201 16:49:52.610110 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wskxk"] Dec 01 16:49:52 crc kubenswrapper[4810]: I1201 16:49:52.649320 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e574877-f6d8-42df-801f-c37977e216f3-catalog-content\") pod \"redhat-marketplace-wskxk\" (UID: \"2e574877-f6d8-42df-801f-c37977e216f3\") " pod="openshift-marketplace/redhat-marketplace-wskxk" Dec 01 16:49:52 crc kubenswrapper[4810]: I1201 16:49:52.649381 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tswn\" (UniqueName: \"kubernetes.io/projected/2e574877-f6d8-42df-801f-c37977e216f3-kube-api-access-6tswn\") pod \"redhat-marketplace-wskxk\" (UID: \"2e574877-f6d8-42df-801f-c37977e216f3\") " pod="openshift-marketplace/redhat-marketplace-wskxk" Dec 01 16:49:52 crc kubenswrapper[4810]: I1201 16:49:52.649775 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e574877-f6d8-42df-801f-c37977e216f3-utilities\") pod \"redhat-marketplace-wskxk\" (UID: \"2e574877-f6d8-42df-801f-c37977e216f3\") " pod="openshift-marketplace/redhat-marketplace-wskxk" Dec 01 16:49:52 crc kubenswrapper[4810]: I1201 16:49:52.752278 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tswn\" (UniqueName: \"kubernetes.io/projected/2e574877-f6d8-42df-801f-c37977e216f3-kube-api-access-6tswn\") pod \"redhat-marketplace-wskxk\" (UID: \"2e574877-f6d8-42df-801f-c37977e216f3\") " pod="openshift-marketplace/redhat-marketplace-wskxk" Dec 01 16:49:52 crc kubenswrapper[4810]: I1201 16:49:52.752328 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e574877-f6d8-42df-801f-c37977e216f3-catalog-content\") pod \"redhat-marketplace-wskxk\" (UID: \"2e574877-f6d8-42df-801f-c37977e216f3\") " pod="openshift-marketplace/redhat-marketplace-wskxk" Dec 01 16:49:52 crc kubenswrapper[4810]: I1201 16:49:52.752435 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e574877-f6d8-42df-801f-c37977e216f3-utilities\") pod \"redhat-marketplace-wskxk\" (UID: \"2e574877-f6d8-42df-801f-c37977e216f3\") " pod="openshift-marketplace/redhat-marketplace-wskxk" Dec 01 16:49:52 crc kubenswrapper[4810]: I1201 16:49:52.753085 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e574877-f6d8-42df-801f-c37977e216f3-utilities\") pod \"redhat-marketplace-wskxk\" (UID: \"2e574877-f6d8-42df-801f-c37977e216f3\") " pod="openshift-marketplace/redhat-marketplace-wskxk" Dec 01 16:49:52 crc kubenswrapper[4810]: I1201 16:49:52.753283 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e574877-f6d8-42df-801f-c37977e216f3-catalog-content\") pod \"redhat-marketplace-wskxk\" (UID: \"2e574877-f6d8-42df-801f-c37977e216f3\") " pod="openshift-marketplace/redhat-marketplace-wskxk" Dec 01 16:49:52 crc kubenswrapper[4810]: I1201 16:49:52.780034 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tswn\" (UniqueName: \"kubernetes.io/projected/2e574877-f6d8-42df-801f-c37977e216f3-kube-api-access-6tswn\") pod \"redhat-marketplace-wskxk\" (UID: \"2e574877-f6d8-42df-801f-c37977e216f3\") " pod="openshift-marketplace/redhat-marketplace-wskxk" Dec 01 16:49:52 crc kubenswrapper[4810]: I1201 16:49:52.917118 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wskxk" Dec 01 16:49:53 crc kubenswrapper[4810]: I1201 16:49:53.465326 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wskxk"] Dec 01 16:49:53 crc kubenswrapper[4810]: W1201 16:49:53.465731 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e574877_f6d8_42df_801f_c37977e216f3.slice/crio-4564839261d67b9f5a4db2514a9687881c1133544855c95958a44fa827d74342 WatchSource:0}: Error finding container 4564839261d67b9f5a4db2514a9687881c1133544855c95958a44fa827d74342: Status 404 returned error can't find the container with id 4564839261d67b9f5a4db2514a9687881c1133544855c95958a44fa827d74342 Dec 01 16:49:53 crc kubenswrapper[4810]: I1201 16:49:53.777374 4810 generic.go:334] "Generic (PLEG): container finished" podID="2e574877-f6d8-42df-801f-c37977e216f3" containerID="6732788318a1f8ed42b03f531721c90c1c1d37ac41000fd65ea78dc3d7af5cf4" exitCode=0 Dec 01 16:49:53 crc kubenswrapper[4810]: I1201 16:49:53.777438 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wskxk" event={"ID":"2e574877-f6d8-42df-801f-c37977e216f3","Type":"ContainerDied","Data":"6732788318a1f8ed42b03f531721c90c1c1d37ac41000fd65ea78dc3d7af5cf4"} Dec 01 16:49:53 crc kubenswrapper[4810]: I1201 16:49:53.777504 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wskxk" event={"ID":"2e574877-f6d8-42df-801f-c37977e216f3","Type":"ContainerStarted","Data":"4564839261d67b9f5a4db2514a9687881c1133544855c95958a44fa827d74342"} Dec 01 16:49:55 crc kubenswrapper[4810]: I1201 16:49:55.799297 4810 generic.go:334] "Generic (PLEG): container finished" podID="2e574877-f6d8-42df-801f-c37977e216f3" containerID="a67f7b292c34c8a3c1953868034d11bd6d8d0d38c32a44a636909c9476515c52" exitCode=0 Dec 01 16:49:55 crc kubenswrapper[4810]: I1201 16:49:55.799373 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wskxk" event={"ID":"2e574877-f6d8-42df-801f-c37977e216f3","Type":"ContainerDied","Data":"a67f7b292c34c8a3c1953868034d11bd6d8d0d38c32a44a636909c9476515c52"} Dec 01 16:49:56 crc kubenswrapper[4810]: I1201 16:49:56.817390 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wskxk" event={"ID":"2e574877-f6d8-42df-801f-c37977e216f3","Type":"ContainerStarted","Data":"3e20587a3aa9a354cfcbd8b1b7760a75bef724ffffa36e050ce8f6ed9c2c3174"} Dec 01 16:49:56 crc kubenswrapper[4810]: I1201 16:49:56.850965 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wskxk" podStartSLOduration=2.184164275 podStartE2EDuration="4.850939365s" podCreationTimestamp="2025-12-01 16:49:52 +0000 UTC" firstStartedPulling="2025-12-01 16:49:53.77936906 +0000 UTC m=+8159.542878653" lastFinishedPulling="2025-12-01 16:49:56.44614414 +0000 UTC m=+8162.209653743" observedRunningTime="2025-12-01 16:49:56.838667183 +0000 UTC m=+8162.602176846" watchObservedRunningTime="2025-12-01 16:49:56.850939365 +0000 UTC m=+8162.614449008" Dec 01 16:50:02 crc kubenswrapper[4810]: I1201 16:50:02.917591 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wskxk" Dec 01 16:50:02 crc kubenswrapper[4810]: I1201 16:50:02.918022 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wskxk" Dec 01 16:50:02 crc kubenswrapper[4810]: I1201 16:50:02.971938 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:50:02 crc kubenswrapper[4810]: I1201 16:50:02.972548 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:50:02 crc kubenswrapper[4810]: I1201 16:50:02.972776 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 16:50:02 crc kubenswrapper[4810]: I1201 16:50:02.974087 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:50:02 crc kubenswrapper[4810]: I1201 16:50:02.974361 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" gracePeriod=600 Dec 01 16:50:02 crc kubenswrapper[4810]: I1201 16:50:02.978292 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wskxk" Dec 01 16:50:03 crc kubenswrapper[4810]: E1201 16:50:03.615197 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:50:03 crc kubenswrapper[4810]: I1201 16:50:03.887828 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" exitCode=0 Dec 01 16:50:03 crc kubenswrapper[4810]: I1201 16:50:03.887913 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3"} Dec 01 16:50:03 crc kubenswrapper[4810]: I1201 16:50:03.887971 4810 scope.go:117] "RemoveContainer" containerID="f65292529d9bba3480ced8eb81902012c9eb05b1dba8694d8ec12fc39a3e4b89" Dec 01 16:50:03 crc kubenswrapper[4810]: I1201 16:50:03.889263 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:50:03 crc kubenswrapper[4810]: E1201 16:50:03.889865 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:50:03 crc kubenswrapper[4810]: I1201 16:50:03.983457 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wskxk" Dec 01 16:50:04 crc kubenswrapper[4810]: I1201 16:50:04.039485 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wskxk"] Dec 01 16:50:05 crc kubenswrapper[4810]: I1201 16:50:05.909520 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wskxk" podUID="2e574877-f6d8-42df-801f-c37977e216f3" containerName="registry-server" containerID="cri-o://3e20587a3aa9a354cfcbd8b1b7760a75bef724ffffa36e050ce8f6ed9c2c3174" gracePeriod=2 Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.368163 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wskxk" Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.484069 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e574877-f6d8-42df-801f-c37977e216f3-utilities\") pod \"2e574877-f6d8-42df-801f-c37977e216f3\" (UID: \"2e574877-f6d8-42df-801f-c37977e216f3\") " Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.484307 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e574877-f6d8-42df-801f-c37977e216f3-catalog-content\") pod \"2e574877-f6d8-42df-801f-c37977e216f3\" (UID: \"2e574877-f6d8-42df-801f-c37977e216f3\") " Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.484419 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tswn\" (UniqueName: \"kubernetes.io/projected/2e574877-f6d8-42df-801f-c37977e216f3-kube-api-access-6tswn\") pod \"2e574877-f6d8-42df-801f-c37977e216f3\" (UID: \"2e574877-f6d8-42df-801f-c37977e216f3\") " Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.485112 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e574877-f6d8-42df-801f-c37977e216f3-utilities" (OuterVolumeSpecName: "utilities") pod "2e574877-f6d8-42df-801f-c37977e216f3" (UID: "2e574877-f6d8-42df-801f-c37977e216f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.489360 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e574877-f6d8-42df-801f-c37977e216f3-kube-api-access-6tswn" (OuterVolumeSpecName: "kube-api-access-6tswn") pod "2e574877-f6d8-42df-801f-c37977e216f3" (UID: "2e574877-f6d8-42df-801f-c37977e216f3"). InnerVolumeSpecName "kube-api-access-6tswn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.506933 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e574877-f6d8-42df-801f-c37977e216f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e574877-f6d8-42df-801f-c37977e216f3" (UID: "2e574877-f6d8-42df-801f-c37977e216f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.588831 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e574877-f6d8-42df-801f-c37977e216f3-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.588890 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e574877-f6d8-42df-801f-c37977e216f3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.588911 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tswn\" (UniqueName: \"kubernetes.io/projected/2e574877-f6d8-42df-801f-c37977e216f3-kube-api-access-6tswn\") on node \"crc\" DevicePath \"\"" Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.920068 4810 generic.go:334] "Generic (PLEG): container finished" podID="2e574877-f6d8-42df-801f-c37977e216f3" containerID="3e20587a3aa9a354cfcbd8b1b7760a75bef724ffffa36e050ce8f6ed9c2c3174" exitCode=0 Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.920122 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wskxk" event={"ID":"2e574877-f6d8-42df-801f-c37977e216f3","Type":"ContainerDied","Data":"3e20587a3aa9a354cfcbd8b1b7760a75bef724ffffa36e050ce8f6ed9c2c3174"} Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.920185 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wskxk" event={"ID":"2e574877-f6d8-42df-801f-c37977e216f3","Type":"ContainerDied","Data":"4564839261d67b9f5a4db2514a9687881c1133544855c95958a44fa827d74342"} Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.920217 4810 scope.go:117] "RemoveContainer" containerID="3e20587a3aa9a354cfcbd8b1b7760a75bef724ffffa36e050ce8f6ed9c2c3174" Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.920145 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wskxk" Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.949173 4810 scope.go:117] "RemoveContainer" containerID="a67f7b292c34c8a3c1953868034d11bd6d8d0d38c32a44a636909c9476515c52" Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.966394 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wskxk"] Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.980545 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wskxk"] Dec 01 16:50:06 crc kubenswrapper[4810]: I1201 16:50:06.984386 4810 scope.go:117] "RemoveContainer" containerID="6732788318a1f8ed42b03f531721c90c1c1d37ac41000fd65ea78dc3d7af5cf4" Dec 01 16:50:07 crc kubenswrapper[4810]: I1201 16:50:07.037647 4810 scope.go:117] "RemoveContainer" containerID="3e20587a3aa9a354cfcbd8b1b7760a75bef724ffffa36e050ce8f6ed9c2c3174" Dec 01 16:50:07 crc kubenswrapper[4810]: E1201 16:50:07.038225 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e20587a3aa9a354cfcbd8b1b7760a75bef724ffffa36e050ce8f6ed9c2c3174\": container with ID starting with 3e20587a3aa9a354cfcbd8b1b7760a75bef724ffffa36e050ce8f6ed9c2c3174 not found: ID does not exist" containerID="3e20587a3aa9a354cfcbd8b1b7760a75bef724ffffa36e050ce8f6ed9c2c3174" Dec 01 16:50:07 crc kubenswrapper[4810]: I1201 16:50:07.038253 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e20587a3aa9a354cfcbd8b1b7760a75bef724ffffa36e050ce8f6ed9c2c3174"} err="failed to get container status \"3e20587a3aa9a354cfcbd8b1b7760a75bef724ffffa36e050ce8f6ed9c2c3174\": rpc error: code = NotFound desc = could not find container \"3e20587a3aa9a354cfcbd8b1b7760a75bef724ffffa36e050ce8f6ed9c2c3174\": container with ID starting with 3e20587a3aa9a354cfcbd8b1b7760a75bef724ffffa36e050ce8f6ed9c2c3174 not found: ID does not exist" Dec 01 16:50:07 crc kubenswrapper[4810]: I1201 16:50:07.038271 4810 scope.go:117] "RemoveContainer" containerID="a67f7b292c34c8a3c1953868034d11bd6d8d0d38c32a44a636909c9476515c52" Dec 01 16:50:07 crc kubenswrapper[4810]: E1201 16:50:07.038690 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a67f7b292c34c8a3c1953868034d11bd6d8d0d38c32a44a636909c9476515c52\": container with ID starting with a67f7b292c34c8a3c1953868034d11bd6d8d0d38c32a44a636909c9476515c52 not found: ID does not exist" containerID="a67f7b292c34c8a3c1953868034d11bd6d8d0d38c32a44a636909c9476515c52" Dec 01 16:50:07 crc kubenswrapper[4810]: I1201 16:50:07.038714 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a67f7b292c34c8a3c1953868034d11bd6d8d0d38c32a44a636909c9476515c52"} err="failed to get container status \"a67f7b292c34c8a3c1953868034d11bd6d8d0d38c32a44a636909c9476515c52\": rpc error: code = NotFound desc = could not find container \"a67f7b292c34c8a3c1953868034d11bd6d8d0d38c32a44a636909c9476515c52\": container with ID starting with a67f7b292c34c8a3c1953868034d11bd6d8d0d38c32a44a636909c9476515c52 not found: ID does not exist" Dec 01 16:50:07 crc kubenswrapper[4810]: I1201 16:50:07.038771 4810 scope.go:117] "RemoveContainer" containerID="6732788318a1f8ed42b03f531721c90c1c1d37ac41000fd65ea78dc3d7af5cf4" Dec 01 16:50:07 crc kubenswrapper[4810]: E1201 16:50:07.039196 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6732788318a1f8ed42b03f531721c90c1c1d37ac41000fd65ea78dc3d7af5cf4\": container with ID starting with 6732788318a1f8ed42b03f531721c90c1c1d37ac41000fd65ea78dc3d7af5cf4 not found: ID does not exist" containerID="6732788318a1f8ed42b03f531721c90c1c1d37ac41000fd65ea78dc3d7af5cf4" Dec 01 16:50:07 crc kubenswrapper[4810]: I1201 16:50:07.039235 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6732788318a1f8ed42b03f531721c90c1c1d37ac41000fd65ea78dc3d7af5cf4"} err="failed to get container status \"6732788318a1f8ed42b03f531721c90c1c1d37ac41000fd65ea78dc3d7af5cf4\": rpc error: code = NotFound desc = could not find container \"6732788318a1f8ed42b03f531721c90c1c1d37ac41000fd65ea78dc3d7af5cf4\": container with ID starting with 6732788318a1f8ed42b03f531721c90c1c1d37ac41000fd65ea78dc3d7af5cf4 not found: ID does not exist" Dec 01 16:50:08 crc kubenswrapper[4810]: I1201 16:50:08.504157 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e574877-f6d8-42df-801f-c37977e216f3" path="/var/lib/kubelet/pods/2e574877-f6d8-42df-801f-c37977e216f3/volumes" Dec 01 16:50:14 crc kubenswrapper[4810]: I1201 16:50:14.497977 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:50:14 crc kubenswrapper[4810]: E1201 16:50:14.500016 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:50:18 crc kubenswrapper[4810]: I1201 16:50:18.039240 4810 generic.go:334] "Generic (PLEG): container finished" podID="69ca686f-3e3b-42eb-a5b5-f1a28a950c6a" containerID="5f9ff8a7e9563c85d245c2fdd47cd54a087d14efa3e536a7caa31eb153f6fab1" exitCode=0 Dec 01 16:50:18 crc kubenswrapper[4810]: I1201 16:50:18.039328 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" event={"ID":"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a","Type":"ContainerDied","Data":"5f9ff8a7e9563c85d245c2fdd47cd54a087d14efa3e536a7caa31eb153f6fab1"} Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.552466 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.643533 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-inventory\") pod \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.643624 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-ssh-key\") pod \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.643670 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld88r\" (UniqueName: \"kubernetes.io/projected/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-kube-api-access-ld88r\") pod \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.643744 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-neutron-metadata-combined-ca-bundle\") pod \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.643788 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.644002 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-nova-metadata-neutron-config-0\") pod \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\" (UID: \"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a\") " Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.649037 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "69ca686f-3e3b-42eb-a5b5-f1a28a950c6a" (UID: "69ca686f-3e3b-42eb-a5b5-f1a28a950c6a"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.649844 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-kube-api-access-ld88r" (OuterVolumeSpecName: "kube-api-access-ld88r") pod "69ca686f-3e3b-42eb-a5b5-f1a28a950c6a" (UID: "69ca686f-3e3b-42eb-a5b5-f1a28a950c6a"). InnerVolumeSpecName "kube-api-access-ld88r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.679248 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "69ca686f-3e3b-42eb-a5b5-f1a28a950c6a" (UID: "69ca686f-3e3b-42eb-a5b5-f1a28a950c6a"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.680120 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-inventory" (OuterVolumeSpecName: "inventory") pod "69ca686f-3e3b-42eb-a5b5-f1a28a950c6a" (UID: "69ca686f-3e3b-42eb-a5b5-f1a28a950c6a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.680814 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "69ca686f-3e3b-42eb-a5b5-f1a28a950c6a" (UID: "69ca686f-3e3b-42eb-a5b5-f1a28a950c6a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.716023 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "69ca686f-3e3b-42eb-a5b5-f1a28a950c6a" (UID: "69ca686f-3e3b-42eb-a5b5-f1a28a950c6a"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.746469 4810 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.746533 4810 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.746546 4810 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.746558 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.746568 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:50:19 crc kubenswrapper[4810]: I1201 16:50:19.746580 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld88r\" (UniqueName: \"kubernetes.io/projected/69ca686f-3e3b-42eb-a5b5-f1a28a950c6a-kube-api-access-ld88r\") on node \"crc\" DevicePath \"\"" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.066072 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" event={"ID":"69ca686f-3e3b-42eb-a5b5-f1a28a950c6a","Type":"ContainerDied","Data":"3bd9ed4270384baecea41cdd373913112d4197dda5654cbdcf097cce1b648ac3"} Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.066116 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bd9ed4270384baecea41cdd373913112d4197dda5654cbdcf097cce1b648ac3" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.066175 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xpffb" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.211684 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-xjz8d"] Dec 01 16:50:20 crc kubenswrapper[4810]: E1201 16:50:20.212260 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e574877-f6d8-42df-801f-c37977e216f3" containerName="extract-utilities" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.212275 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e574877-f6d8-42df-801f-c37977e216f3" containerName="extract-utilities" Dec 01 16:50:20 crc kubenswrapper[4810]: E1201 16:50:20.212305 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e574877-f6d8-42df-801f-c37977e216f3" containerName="registry-server" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.212313 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e574877-f6d8-42df-801f-c37977e216f3" containerName="registry-server" Dec 01 16:50:20 crc kubenswrapper[4810]: E1201 16:50:20.212324 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e574877-f6d8-42df-801f-c37977e216f3" containerName="extract-content" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.212331 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e574877-f6d8-42df-801f-c37977e216f3" containerName="extract-content" Dec 01 16:50:20 crc kubenswrapper[4810]: E1201 16:50:20.212373 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69ca686f-3e3b-42eb-a5b5-f1a28a950c6a" containerName="neutron-metadata-openstack-openstack-cell1" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.212382 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="69ca686f-3e3b-42eb-a5b5-f1a28a950c6a" containerName="neutron-metadata-openstack-openstack-cell1" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.212619 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e574877-f6d8-42df-801f-c37977e216f3" containerName="registry-server" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.212632 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="69ca686f-3e3b-42eb-a5b5-f1a28a950c6a" containerName="neutron-metadata-openstack-openstack-cell1" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.213607 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.217724 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.217888 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.217921 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.218243 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.218432 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.225407 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-xjz8d"] Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.359300 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-xjz8d\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.359347 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hc2t\" (UniqueName: \"kubernetes.io/projected/051ab66c-513f-4f4f-98a9-a20426bea652-kube-api-access-6hc2t\") pod \"libvirt-openstack-openstack-cell1-xjz8d\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.359401 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-inventory\") pod \"libvirt-openstack-openstack-cell1-xjz8d\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.359672 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-xjz8d\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.359736 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-ssh-key\") pod \"libvirt-openstack-openstack-cell1-xjz8d\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.461878 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-xjz8d\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.461933 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hc2t\" (UniqueName: \"kubernetes.io/projected/051ab66c-513f-4f4f-98a9-a20426bea652-kube-api-access-6hc2t\") pod \"libvirt-openstack-openstack-cell1-xjz8d\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.461988 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-inventory\") pod \"libvirt-openstack-openstack-cell1-xjz8d\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.462062 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-xjz8d\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.462084 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-ssh-key\") pod \"libvirt-openstack-openstack-cell1-xjz8d\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.466513 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-ssh-key\") pod \"libvirt-openstack-openstack-cell1-xjz8d\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.467297 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-xjz8d\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.467437 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-inventory\") pod \"libvirt-openstack-openstack-cell1-xjz8d\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.467721 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-xjz8d\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.480616 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hc2t\" (UniqueName: \"kubernetes.io/projected/051ab66c-513f-4f4f-98a9-a20426bea652-kube-api-access-6hc2t\") pod \"libvirt-openstack-openstack-cell1-xjz8d\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:50:20 crc kubenswrapper[4810]: I1201 16:50:20.576742 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:50:21 crc kubenswrapper[4810]: I1201 16:50:21.109139 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-xjz8d"] Dec 01 16:50:22 crc kubenswrapper[4810]: I1201 16:50:22.088335 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" event={"ID":"051ab66c-513f-4f4f-98a9-a20426bea652","Type":"ContainerStarted","Data":"8f2bc33f1dd40b42327fde44100ae06a504d4c45e9f98e86c2e59b4b0fe24f89"} Dec 01 16:50:23 crc kubenswrapper[4810]: I1201 16:50:23.102334 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" event={"ID":"051ab66c-513f-4f4f-98a9-a20426bea652","Type":"ContainerStarted","Data":"f6a8ea34826aecdacae2296b102454bae4fe5f596efff7fbf205f3a4a1b00942"} Dec 01 16:50:23 crc kubenswrapper[4810]: I1201 16:50:23.123337 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" podStartSLOduration=2.327180247 podStartE2EDuration="3.123319423s" podCreationTimestamp="2025-12-01 16:50:20 +0000 UTC" firstStartedPulling="2025-12-01 16:50:21.120270663 +0000 UTC m=+8186.883780266" lastFinishedPulling="2025-12-01 16:50:21.916409819 +0000 UTC m=+8187.679919442" observedRunningTime="2025-12-01 16:50:23.118308117 +0000 UTC m=+8188.881817760" watchObservedRunningTime="2025-12-01 16:50:23.123319423 +0000 UTC m=+8188.886829046" Dec 01 16:50:25 crc kubenswrapper[4810]: I1201 16:50:25.492660 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:50:25 crc kubenswrapper[4810]: E1201 16:50:25.494042 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:50:40 crc kubenswrapper[4810]: I1201 16:50:40.493181 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:50:40 crc kubenswrapper[4810]: E1201 16:50:40.494772 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:50:55 crc kubenswrapper[4810]: I1201 16:50:55.492530 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:50:55 crc kubenswrapper[4810]: E1201 16:50:55.494081 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:51:10 crc kubenswrapper[4810]: I1201 16:51:10.492035 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:51:10 crc kubenswrapper[4810]: E1201 16:51:10.492885 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:51:22 crc kubenswrapper[4810]: I1201 16:51:22.491256 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:51:22 crc kubenswrapper[4810]: E1201 16:51:22.492136 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:51:35 crc kubenswrapper[4810]: I1201 16:51:35.493010 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:51:35 crc kubenswrapper[4810]: E1201 16:51:35.494519 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:51:47 crc kubenswrapper[4810]: I1201 16:51:47.491094 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:51:47 crc kubenswrapper[4810]: E1201 16:51:47.493832 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:52:01 crc kubenswrapper[4810]: I1201 16:52:01.490982 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:52:01 crc kubenswrapper[4810]: E1201 16:52:01.491908 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:52:14 crc kubenswrapper[4810]: I1201 16:52:14.507260 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:52:14 crc kubenswrapper[4810]: E1201 16:52:14.508594 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:52:29 crc kubenswrapper[4810]: I1201 16:52:29.492185 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:52:29 crc kubenswrapper[4810]: E1201 16:52:29.493415 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:52:44 crc kubenswrapper[4810]: I1201 16:52:44.501028 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:52:44 crc kubenswrapper[4810]: E1201 16:52:44.503490 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:52:55 crc kubenswrapper[4810]: I1201 16:52:55.491665 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:52:55 crc kubenswrapper[4810]: E1201 16:52:55.492865 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:53:08 crc kubenswrapper[4810]: I1201 16:53:08.491780 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:53:08 crc kubenswrapper[4810]: E1201 16:53:08.492616 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:53:22 crc kubenswrapper[4810]: I1201 16:53:22.491246 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:53:22 crc kubenswrapper[4810]: E1201 16:53:22.492061 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:53:35 crc kubenswrapper[4810]: I1201 16:53:35.490959 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:53:35 crc kubenswrapper[4810]: E1201 16:53:35.491651 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:53:49 crc kubenswrapper[4810]: I1201 16:53:49.492004 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:53:49 crc kubenswrapper[4810]: E1201 16:53:49.493012 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:54:04 crc kubenswrapper[4810]: I1201 16:54:04.514914 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:54:04 crc kubenswrapper[4810]: E1201 16:54:04.516140 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:54:19 crc kubenswrapper[4810]: I1201 16:54:19.491626 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:54:19 crc kubenswrapper[4810]: E1201 16:54:19.492894 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:54:34 crc kubenswrapper[4810]: I1201 16:54:34.504786 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:54:34 crc kubenswrapper[4810]: E1201 16:54:34.505635 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:54:48 crc kubenswrapper[4810]: I1201 16:54:48.491808 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:54:48 crc kubenswrapper[4810]: E1201 16:54:48.492814 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:55:02 crc kubenswrapper[4810]: I1201 16:55:02.492433 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:55:02 crc kubenswrapper[4810]: E1201 16:55:02.493522 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 16:55:13 crc kubenswrapper[4810]: I1201 16:55:13.089629 4810 generic.go:334] "Generic (PLEG): container finished" podID="051ab66c-513f-4f4f-98a9-a20426bea652" containerID="f6a8ea34826aecdacae2296b102454bae4fe5f596efff7fbf205f3a4a1b00942" exitCode=0 Dec 01 16:55:13 crc kubenswrapper[4810]: I1201 16:55:13.089725 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" event={"ID":"051ab66c-513f-4f4f-98a9-a20426bea652","Type":"ContainerDied","Data":"f6a8ea34826aecdacae2296b102454bae4fe5f596efff7fbf205f3a4a1b00942"} Dec 01 16:55:14 crc kubenswrapper[4810]: I1201 16:55:14.694539 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:55:14 crc kubenswrapper[4810]: I1201 16:55:14.771435 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hc2t\" (UniqueName: \"kubernetes.io/projected/051ab66c-513f-4f4f-98a9-a20426bea652-kube-api-access-6hc2t\") pod \"051ab66c-513f-4f4f-98a9-a20426bea652\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " Dec 01 16:55:14 crc kubenswrapper[4810]: I1201 16:55:14.771609 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-libvirt-secret-0\") pod \"051ab66c-513f-4f4f-98a9-a20426bea652\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " Dec 01 16:55:14 crc kubenswrapper[4810]: I1201 16:55:14.771678 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-inventory\") pod \"051ab66c-513f-4f4f-98a9-a20426bea652\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " Dec 01 16:55:14 crc kubenswrapper[4810]: I1201 16:55:14.771826 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-ssh-key\") pod \"051ab66c-513f-4f4f-98a9-a20426bea652\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " Dec 01 16:55:14 crc kubenswrapper[4810]: I1201 16:55:14.771881 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-libvirt-combined-ca-bundle\") pod \"051ab66c-513f-4f4f-98a9-a20426bea652\" (UID: \"051ab66c-513f-4f4f-98a9-a20426bea652\") " Dec 01 16:55:14 crc kubenswrapper[4810]: I1201 16:55:14.777755 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "051ab66c-513f-4f4f-98a9-a20426bea652" (UID: "051ab66c-513f-4f4f-98a9-a20426bea652"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:55:14 crc kubenswrapper[4810]: I1201 16:55:14.778969 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/051ab66c-513f-4f4f-98a9-a20426bea652-kube-api-access-6hc2t" (OuterVolumeSpecName: "kube-api-access-6hc2t") pod "051ab66c-513f-4f4f-98a9-a20426bea652" (UID: "051ab66c-513f-4f4f-98a9-a20426bea652"). InnerVolumeSpecName "kube-api-access-6hc2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:55:14 crc kubenswrapper[4810]: I1201 16:55:14.802187 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-inventory" (OuterVolumeSpecName: "inventory") pod "051ab66c-513f-4f4f-98a9-a20426bea652" (UID: "051ab66c-513f-4f4f-98a9-a20426bea652"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:55:14 crc kubenswrapper[4810]: I1201 16:55:14.807670 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "051ab66c-513f-4f4f-98a9-a20426bea652" (UID: "051ab66c-513f-4f4f-98a9-a20426bea652"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:55:14 crc kubenswrapper[4810]: I1201 16:55:14.809623 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "051ab66c-513f-4f4f-98a9-a20426bea652" (UID: "051ab66c-513f-4f4f-98a9-a20426bea652"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:55:14 crc kubenswrapper[4810]: I1201 16:55:14.875581 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hc2t\" (UniqueName: \"kubernetes.io/projected/051ab66c-513f-4f4f-98a9-a20426bea652-kube-api-access-6hc2t\") on node \"crc\" DevicePath \"\"" Dec 01 16:55:14 crc kubenswrapper[4810]: I1201 16:55:14.875617 4810 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:55:14 crc kubenswrapper[4810]: I1201 16:55:14.875634 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:55:14 crc kubenswrapper[4810]: I1201 16:55:14.875647 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:55:14 crc kubenswrapper[4810]: I1201 16:55:14.875660 4810 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/051ab66c-513f-4f4f-98a9-a20426bea652-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.111191 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" event={"ID":"051ab66c-513f-4f4f-98a9-a20426bea652","Type":"ContainerDied","Data":"8f2bc33f1dd40b42327fde44100ae06a504d4c45e9f98e86c2e59b4b0fe24f89"} Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.111232 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f2bc33f1dd40b42327fde44100ae06a504d4c45e9f98e86c2e59b4b0fe24f89" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.111252 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-xjz8d" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.237908 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-8cbjx"] Dec 01 16:55:15 crc kubenswrapper[4810]: E1201 16:55:15.238560 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="051ab66c-513f-4f4f-98a9-a20426bea652" containerName="libvirt-openstack-openstack-cell1" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.238590 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="051ab66c-513f-4f4f-98a9-a20426bea652" containerName="libvirt-openstack-openstack-cell1" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.238912 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="051ab66c-513f-4f4f-98a9-a20426bea652" containerName="libvirt-openstack-openstack-cell1" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.239964 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.244266 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.244298 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.244545 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.244645 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.244917 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.245260 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.247828 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.259352 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-8cbjx"] Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.386119 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.386507 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.386603 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.386647 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.386688 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.386818 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-inventory\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.387125 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztcr7\" (UniqueName: \"kubernetes.io/projected/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-kube-api-access-ztcr7\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.387271 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.387318 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.490203 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.490332 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.490438 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.490562 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.490694 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.490761 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-inventory\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.490954 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztcr7\" (UniqueName: \"kubernetes.io/projected/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-kube-api-access-ztcr7\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.491092 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.491166 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.491413 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.494786 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-inventory\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.495521 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.498534 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.500322 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.513274 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.516187 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.517039 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.520067 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztcr7\" (UniqueName: \"kubernetes.io/projected/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-kube-api-access-ztcr7\") pod \"nova-cell1-openstack-openstack-cell1-8cbjx\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:15 crc kubenswrapper[4810]: I1201 16:55:15.609841 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:55:16 crc kubenswrapper[4810]: I1201 16:55:16.139161 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 16:55:16 crc kubenswrapper[4810]: I1201 16:55:16.142541 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-8cbjx"] Dec 01 16:55:17 crc kubenswrapper[4810]: I1201 16:55:17.138241 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" event={"ID":"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f","Type":"ContainerStarted","Data":"388eec5629f089a426f76fd3ec4a0eddb8479c73c4049047201254a643a878ee"} Dec 01 16:55:17 crc kubenswrapper[4810]: I1201 16:55:17.491503 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:55:18 crc kubenswrapper[4810]: I1201 16:55:18.150123 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"58ed6efdf6a9eee8fc2c0ff2afbf931ed73ba5c194f897a984932d52cef19b8d"} Dec 01 16:55:18 crc kubenswrapper[4810]: I1201 16:55:18.153520 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" event={"ID":"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f","Type":"ContainerStarted","Data":"6e0ce25193ab2d6d0fb6e1a77c303fe57b40d03cdd01125b950c81dbfc9f6ad9"} Dec 01 16:56:31 crc kubenswrapper[4810]: I1201 16:56:31.956075 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" podStartSLOduration=75.607432635 podStartE2EDuration="1m16.956043448s" podCreationTimestamp="2025-12-01 16:55:15 +0000 UTC" firstStartedPulling="2025-12-01 16:55:16.138964493 +0000 UTC m=+8481.902474096" lastFinishedPulling="2025-12-01 16:55:17.487575306 +0000 UTC m=+8483.251084909" observedRunningTime="2025-12-01 16:55:18.192924622 +0000 UTC m=+8483.956434225" watchObservedRunningTime="2025-12-01 16:56:31.956043448 +0000 UTC m=+8557.719553091" Dec 01 16:56:31 crc kubenswrapper[4810]: I1201 16:56:31.965739 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dtmhc"] Dec 01 16:56:31 crc kubenswrapper[4810]: I1201 16:56:31.969201 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtmhc" Dec 01 16:56:31 crc kubenswrapper[4810]: I1201 16:56:31.980460 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dtmhc"] Dec 01 16:56:32 crc kubenswrapper[4810]: I1201 16:56:32.168169 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b55d01e7-b6ef-498e-a55e-3dbcfb19945c-utilities\") pod \"certified-operators-dtmhc\" (UID: \"b55d01e7-b6ef-498e-a55e-3dbcfb19945c\") " pod="openshift-marketplace/certified-operators-dtmhc" Dec 01 16:56:32 crc kubenswrapper[4810]: I1201 16:56:32.168303 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b55d01e7-b6ef-498e-a55e-3dbcfb19945c-catalog-content\") pod \"certified-operators-dtmhc\" (UID: \"b55d01e7-b6ef-498e-a55e-3dbcfb19945c\") " pod="openshift-marketplace/certified-operators-dtmhc" Dec 01 16:56:32 crc kubenswrapper[4810]: I1201 16:56:32.168399 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt5kp\" (UniqueName: \"kubernetes.io/projected/b55d01e7-b6ef-498e-a55e-3dbcfb19945c-kube-api-access-bt5kp\") pod \"certified-operators-dtmhc\" (UID: \"b55d01e7-b6ef-498e-a55e-3dbcfb19945c\") " pod="openshift-marketplace/certified-operators-dtmhc" Dec 01 16:56:32 crc kubenswrapper[4810]: I1201 16:56:32.270756 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b55d01e7-b6ef-498e-a55e-3dbcfb19945c-utilities\") pod \"certified-operators-dtmhc\" (UID: \"b55d01e7-b6ef-498e-a55e-3dbcfb19945c\") " pod="openshift-marketplace/certified-operators-dtmhc" Dec 01 16:56:32 crc kubenswrapper[4810]: I1201 16:56:32.270874 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b55d01e7-b6ef-498e-a55e-3dbcfb19945c-catalog-content\") pod \"certified-operators-dtmhc\" (UID: \"b55d01e7-b6ef-498e-a55e-3dbcfb19945c\") " pod="openshift-marketplace/certified-operators-dtmhc" Dec 01 16:56:32 crc kubenswrapper[4810]: I1201 16:56:32.270959 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt5kp\" (UniqueName: \"kubernetes.io/projected/b55d01e7-b6ef-498e-a55e-3dbcfb19945c-kube-api-access-bt5kp\") pod \"certified-operators-dtmhc\" (UID: \"b55d01e7-b6ef-498e-a55e-3dbcfb19945c\") " pod="openshift-marketplace/certified-operators-dtmhc" Dec 01 16:56:32 crc kubenswrapper[4810]: I1201 16:56:32.271407 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b55d01e7-b6ef-498e-a55e-3dbcfb19945c-catalog-content\") pod \"certified-operators-dtmhc\" (UID: \"b55d01e7-b6ef-498e-a55e-3dbcfb19945c\") " pod="openshift-marketplace/certified-operators-dtmhc" Dec 01 16:56:32 crc kubenswrapper[4810]: I1201 16:56:32.271778 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b55d01e7-b6ef-498e-a55e-3dbcfb19945c-utilities\") pod \"certified-operators-dtmhc\" (UID: \"b55d01e7-b6ef-498e-a55e-3dbcfb19945c\") " pod="openshift-marketplace/certified-operators-dtmhc" Dec 01 16:56:32 crc kubenswrapper[4810]: I1201 16:56:32.298997 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt5kp\" (UniqueName: \"kubernetes.io/projected/b55d01e7-b6ef-498e-a55e-3dbcfb19945c-kube-api-access-bt5kp\") pod \"certified-operators-dtmhc\" (UID: \"b55d01e7-b6ef-498e-a55e-3dbcfb19945c\") " pod="openshift-marketplace/certified-operators-dtmhc" Dec 01 16:56:32 crc kubenswrapper[4810]: I1201 16:56:32.319084 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtmhc" Dec 01 16:56:32 crc kubenswrapper[4810]: I1201 16:56:32.837272 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dtmhc"] Dec 01 16:56:33 crc kubenswrapper[4810]: I1201 16:56:33.004112 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtmhc" event={"ID":"b55d01e7-b6ef-498e-a55e-3dbcfb19945c","Type":"ContainerStarted","Data":"f4a219105b95af25709b46ef1c1d5f0b4bb4bcaacd4d9abb2c57bfa18f59a8cf"} Dec 01 16:56:34 crc kubenswrapper[4810]: I1201 16:56:34.019751 4810 generic.go:334] "Generic (PLEG): container finished" podID="b55d01e7-b6ef-498e-a55e-3dbcfb19945c" containerID="3487b63310205ca0ffdf003142229bb90c1442155d66eab3b7ee0adb1e2f554d" exitCode=0 Dec 01 16:56:34 crc kubenswrapper[4810]: I1201 16:56:34.019862 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtmhc" event={"ID":"b55d01e7-b6ef-498e-a55e-3dbcfb19945c","Type":"ContainerDied","Data":"3487b63310205ca0ffdf003142229bb90c1442155d66eab3b7ee0adb1e2f554d"} Dec 01 16:56:36 crc kubenswrapper[4810]: I1201 16:56:36.044994 4810 generic.go:334] "Generic (PLEG): container finished" podID="b55d01e7-b6ef-498e-a55e-3dbcfb19945c" containerID="f69912b8b21cfdde303583dbe3adc2f9a57e09af2e1450da4ed4a0caa4b6fe9f" exitCode=0 Dec 01 16:56:36 crc kubenswrapper[4810]: I1201 16:56:36.045083 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtmhc" event={"ID":"b55d01e7-b6ef-498e-a55e-3dbcfb19945c","Type":"ContainerDied","Data":"f69912b8b21cfdde303583dbe3adc2f9a57e09af2e1450da4ed4a0caa4b6fe9f"} Dec 01 16:56:37 crc kubenswrapper[4810]: I1201 16:56:37.067250 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtmhc" event={"ID":"b55d01e7-b6ef-498e-a55e-3dbcfb19945c","Type":"ContainerStarted","Data":"788481719cc6cee3d90a7437ca39081157b3c0a2ad3aef3c28d70d40e9df9278"} Dec 01 16:56:37 crc kubenswrapper[4810]: I1201 16:56:37.092650 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dtmhc" podStartSLOduration=3.580754098 podStartE2EDuration="6.092630372s" podCreationTimestamp="2025-12-01 16:56:31 +0000 UTC" firstStartedPulling="2025-12-01 16:56:34.024263863 +0000 UTC m=+8559.787773476" lastFinishedPulling="2025-12-01 16:56:36.536140147 +0000 UTC m=+8562.299649750" observedRunningTime="2025-12-01 16:56:37.084768808 +0000 UTC m=+8562.848278481" watchObservedRunningTime="2025-12-01 16:56:37.092630372 +0000 UTC m=+8562.856139985" Dec 01 16:56:42 crc kubenswrapper[4810]: I1201 16:56:42.319244 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dtmhc" Dec 01 16:56:42 crc kubenswrapper[4810]: I1201 16:56:42.319571 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dtmhc" Dec 01 16:56:42 crc kubenswrapper[4810]: I1201 16:56:42.393873 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dtmhc" Dec 01 16:56:43 crc kubenswrapper[4810]: I1201 16:56:43.198302 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dtmhc" Dec 01 16:56:43 crc kubenswrapper[4810]: I1201 16:56:43.255094 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dtmhc"] Dec 01 16:56:45 crc kubenswrapper[4810]: I1201 16:56:45.149203 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dtmhc" podUID="b55d01e7-b6ef-498e-a55e-3dbcfb19945c" containerName="registry-server" containerID="cri-o://788481719cc6cee3d90a7437ca39081157b3c0a2ad3aef3c28d70d40e9df9278" gracePeriod=2 Dec 01 16:56:45 crc kubenswrapper[4810]: I1201 16:56:45.629894 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtmhc" Dec 01 16:56:45 crc kubenswrapper[4810]: I1201 16:56:45.770267 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bt5kp\" (UniqueName: \"kubernetes.io/projected/b55d01e7-b6ef-498e-a55e-3dbcfb19945c-kube-api-access-bt5kp\") pod \"b55d01e7-b6ef-498e-a55e-3dbcfb19945c\" (UID: \"b55d01e7-b6ef-498e-a55e-3dbcfb19945c\") " Dec 01 16:56:45 crc kubenswrapper[4810]: I1201 16:56:45.770368 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b55d01e7-b6ef-498e-a55e-3dbcfb19945c-catalog-content\") pod \"b55d01e7-b6ef-498e-a55e-3dbcfb19945c\" (UID: \"b55d01e7-b6ef-498e-a55e-3dbcfb19945c\") " Dec 01 16:56:45 crc kubenswrapper[4810]: I1201 16:56:45.770429 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b55d01e7-b6ef-498e-a55e-3dbcfb19945c-utilities\") pod \"b55d01e7-b6ef-498e-a55e-3dbcfb19945c\" (UID: \"b55d01e7-b6ef-498e-a55e-3dbcfb19945c\") " Dec 01 16:56:45 crc kubenswrapper[4810]: I1201 16:56:45.772248 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b55d01e7-b6ef-498e-a55e-3dbcfb19945c-utilities" (OuterVolumeSpecName: "utilities") pod "b55d01e7-b6ef-498e-a55e-3dbcfb19945c" (UID: "b55d01e7-b6ef-498e-a55e-3dbcfb19945c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:56:45 crc kubenswrapper[4810]: I1201 16:56:45.779460 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b55d01e7-b6ef-498e-a55e-3dbcfb19945c-kube-api-access-bt5kp" (OuterVolumeSpecName: "kube-api-access-bt5kp") pod "b55d01e7-b6ef-498e-a55e-3dbcfb19945c" (UID: "b55d01e7-b6ef-498e-a55e-3dbcfb19945c"). InnerVolumeSpecName "kube-api-access-bt5kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:56:45 crc kubenswrapper[4810]: I1201 16:56:45.873771 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bt5kp\" (UniqueName: \"kubernetes.io/projected/b55d01e7-b6ef-498e-a55e-3dbcfb19945c-kube-api-access-bt5kp\") on node \"crc\" DevicePath \"\"" Dec 01 16:56:45 crc kubenswrapper[4810]: I1201 16:56:45.874158 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b55d01e7-b6ef-498e-a55e-3dbcfb19945c-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:56:46 crc kubenswrapper[4810]: I1201 16:56:46.141782 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b55d01e7-b6ef-498e-a55e-3dbcfb19945c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b55d01e7-b6ef-498e-a55e-3dbcfb19945c" (UID: "b55d01e7-b6ef-498e-a55e-3dbcfb19945c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:56:46 crc kubenswrapper[4810]: I1201 16:56:46.160796 4810 generic.go:334] "Generic (PLEG): container finished" podID="b55d01e7-b6ef-498e-a55e-3dbcfb19945c" containerID="788481719cc6cee3d90a7437ca39081157b3c0a2ad3aef3c28d70d40e9df9278" exitCode=0 Dec 01 16:56:46 crc kubenswrapper[4810]: I1201 16:56:46.160867 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dtmhc" Dec 01 16:56:46 crc kubenswrapper[4810]: I1201 16:56:46.160860 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtmhc" event={"ID":"b55d01e7-b6ef-498e-a55e-3dbcfb19945c","Type":"ContainerDied","Data":"788481719cc6cee3d90a7437ca39081157b3c0a2ad3aef3c28d70d40e9df9278"} Dec 01 16:56:46 crc kubenswrapper[4810]: I1201 16:56:46.162088 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dtmhc" event={"ID":"b55d01e7-b6ef-498e-a55e-3dbcfb19945c","Type":"ContainerDied","Data":"f4a219105b95af25709b46ef1c1d5f0b4bb4bcaacd4d9abb2c57bfa18f59a8cf"} Dec 01 16:56:46 crc kubenswrapper[4810]: I1201 16:56:46.162145 4810 scope.go:117] "RemoveContainer" containerID="788481719cc6cee3d90a7437ca39081157b3c0a2ad3aef3c28d70d40e9df9278" Dec 01 16:56:46 crc kubenswrapper[4810]: I1201 16:56:46.181344 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b55d01e7-b6ef-498e-a55e-3dbcfb19945c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:56:46 crc kubenswrapper[4810]: I1201 16:56:46.194218 4810 scope.go:117] "RemoveContainer" containerID="f69912b8b21cfdde303583dbe3adc2f9a57e09af2e1450da4ed4a0caa4b6fe9f" Dec 01 16:56:46 crc kubenswrapper[4810]: I1201 16:56:46.207620 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dtmhc"] Dec 01 16:56:46 crc kubenswrapper[4810]: I1201 16:56:46.217281 4810 scope.go:117] "RemoveContainer" containerID="3487b63310205ca0ffdf003142229bb90c1442155d66eab3b7ee0adb1e2f554d" Dec 01 16:56:46 crc kubenswrapper[4810]: I1201 16:56:46.220405 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dtmhc"] Dec 01 16:56:46 crc kubenswrapper[4810]: I1201 16:56:46.290837 4810 scope.go:117] "RemoveContainer" containerID="788481719cc6cee3d90a7437ca39081157b3c0a2ad3aef3c28d70d40e9df9278" Dec 01 16:56:46 crc kubenswrapper[4810]: E1201 16:56:46.296642 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"788481719cc6cee3d90a7437ca39081157b3c0a2ad3aef3c28d70d40e9df9278\": container with ID starting with 788481719cc6cee3d90a7437ca39081157b3c0a2ad3aef3c28d70d40e9df9278 not found: ID does not exist" containerID="788481719cc6cee3d90a7437ca39081157b3c0a2ad3aef3c28d70d40e9df9278" Dec 01 16:56:46 crc kubenswrapper[4810]: I1201 16:56:46.296729 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"788481719cc6cee3d90a7437ca39081157b3c0a2ad3aef3c28d70d40e9df9278"} err="failed to get container status \"788481719cc6cee3d90a7437ca39081157b3c0a2ad3aef3c28d70d40e9df9278\": rpc error: code = NotFound desc = could not find container \"788481719cc6cee3d90a7437ca39081157b3c0a2ad3aef3c28d70d40e9df9278\": container with ID starting with 788481719cc6cee3d90a7437ca39081157b3c0a2ad3aef3c28d70d40e9df9278 not found: ID does not exist" Dec 01 16:56:46 crc kubenswrapper[4810]: I1201 16:56:46.296765 4810 scope.go:117] "RemoveContainer" containerID="f69912b8b21cfdde303583dbe3adc2f9a57e09af2e1450da4ed4a0caa4b6fe9f" Dec 01 16:56:46 crc kubenswrapper[4810]: E1201 16:56:46.297132 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f69912b8b21cfdde303583dbe3adc2f9a57e09af2e1450da4ed4a0caa4b6fe9f\": container with ID starting with f69912b8b21cfdde303583dbe3adc2f9a57e09af2e1450da4ed4a0caa4b6fe9f not found: ID does not exist" containerID="f69912b8b21cfdde303583dbe3adc2f9a57e09af2e1450da4ed4a0caa4b6fe9f" Dec 01 16:56:46 crc kubenswrapper[4810]: I1201 16:56:46.297174 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f69912b8b21cfdde303583dbe3adc2f9a57e09af2e1450da4ed4a0caa4b6fe9f"} err="failed to get container status \"f69912b8b21cfdde303583dbe3adc2f9a57e09af2e1450da4ed4a0caa4b6fe9f\": rpc error: code = NotFound desc = could not find container \"f69912b8b21cfdde303583dbe3adc2f9a57e09af2e1450da4ed4a0caa4b6fe9f\": container with ID starting with f69912b8b21cfdde303583dbe3adc2f9a57e09af2e1450da4ed4a0caa4b6fe9f not found: ID does not exist" Dec 01 16:56:46 crc kubenswrapper[4810]: I1201 16:56:46.297202 4810 scope.go:117] "RemoveContainer" containerID="3487b63310205ca0ffdf003142229bb90c1442155d66eab3b7ee0adb1e2f554d" Dec 01 16:56:46 crc kubenswrapper[4810]: E1201 16:56:46.297672 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3487b63310205ca0ffdf003142229bb90c1442155d66eab3b7ee0adb1e2f554d\": container with ID starting with 3487b63310205ca0ffdf003142229bb90c1442155d66eab3b7ee0adb1e2f554d not found: ID does not exist" containerID="3487b63310205ca0ffdf003142229bb90c1442155d66eab3b7ee0adb1e2f554d" Dec 01 16:56:46 crc kubenswrapper[4810]: I1201 16:56:46.297692 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3487b63310205ca0ffdf003142229bb90c1442155d66eab3b7ee0adb1e2f554d"} err="failed to get container status \"3487b63310205ca0ffdf003142229bb90c1442155d66eab3b7ee0adb1e2f554d\": rpc error: code = NotFound desc = could not find container \"3487b63310205ca0ffdf003142229bb90c1442155d66eab3b7ee0adb1e2f554d\": container with ID starting with 3487b63310205ca0ffdf003142229bb90c1442155d66eab3b7ee0adb1e2f554d not found: ID does not exist" Dec 01 16:56:46 crc kubenswrapper[4810]: I1201 16:56:46.503353 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b55d01e7-b6ef-498e-a55e-3dbcfb19945c" path="/var/lib/kubelet/pods/b55d01e7-b6ef-498e-a55e-3dbcfb19945c/volumes" Dec 01 16:57:32 crc kubenswrapper[4810]: I1201 16:57:32.972499 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:57:32 crc kubenswrapper[4810]: I1201 16:57:32.973096 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:58:02 crc kubenswrapper[4810]: I1201 16:58:02.972863 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:58:02 crc kubenswrapper[4810]: I1201 16:58:02.973840 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:58:31 crc kubenswrapper[4810]: I1201 16:58:31.333042 4810 generic.go:334] "Generic (PLEG): container finished" podID="7bfcd0d6-c140-4c51-aa35-8acd8a96d68f" containerID="6e0ce25193ab2d6d0fb6e1a77c303fe57b40d03cdd01125b950c81dbfc9f6ad9" exitCode=0 Dec 01 16:58:31 crc kubenswrapper[4810]: I1201 16:58:31.333080 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" event={"ID":"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f","Type":"ContainerDied","Data":"6e0ce25193ab2d6d0fb6e1a77c303fe57b40d03cdd01125b950c81dbfc9f6ad9"} Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.888952 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.928462 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cell1-combined-ca-bundle\") pod \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.928704 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-ssh-key\") pod \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.928792 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-migration-ssh-key-1\") pod \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.928818 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-inventory\") pod \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.928939 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cells-global-config-0\") pod \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.928976 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cell1-compute-config-1\") pod \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.929023 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztcr7\" (UniqueName: \"kubernetes.io/projected/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-kube-api-access-ztcr7\") pod \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.929143 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cell1-compute-config-0\") pod \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.929250 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-migration-ssh-key-0\") pod \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\" (UID: \"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f\") " Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.936586 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "7bfcd0d6-c140-4c51-aa35-8acd8a96d68f" (UID: "7bfcd0d6-c140-4c51-aa35-8acd8a96d68f"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.936882 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-kube-api-access-ztcr7" (OuterVolumeSpecName: "kube-api-access-ztcr7") pod "7bfcd0d6-c140-4c51-aa35-8acd8a96d68f" (UID: "7bfcd0d6-c140-4c51-aa35-8acd8a96d68f"). InnerVolumeSpecName "kube-api-access-ztcr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.962456 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "7bfcd0d6-c140-4c51-aa35-8acd8a96d68f" (UID: "7bfcd0d6-c140-4c51-aa35-8acd8a96d68f"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.968688 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "7bfcd0d6-c140-4c51-aa35-8acd8a96d68f" (UID: "7bfcd0d6-c140-4c51-aa35-8acd8a96d68f"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.968815 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-inventory" (OuterVolumeSpecName: "inventory") pod "7bfcd0d6-c140-4c51-aa35-8acd8a96d68f" (UID: "7bfcd0d6-c140-4c51-aa35-8acd8a96d68f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.970116 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7bfcd0d6-c140-4c51-aa35-8acd8a96d68f" (UID: "7bfcd0d6-c140-4c51-aa35-8acd8a96d68f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.971707 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "7bfcd0d6-c140-4c51-aa35-8acd8a96d68f" (UID: "7bfcd0d6-c140-4c51-aa35-8acd8a96d68f"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.971804 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.971869 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.971930 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.972978 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"58ed6efdf6a9eee8fc2c0ff2afbf931ed73ba5c194f897a984932d52cef19b8d"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.973061 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://58ed6efdf6a9eee8fc2c0ff2afbf931ed73ba5c194f897a984932d52cef19b8d" gracePeriod=600 Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.989406 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "7bfcd0d6-c140-4c51-aa35-8acd8a96d68f" (UID: "7bfcd0d6-c140-4c51-aa35-8acd8a96d68f"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:58:32 crc kubenswrapper[4810]: I1201 16:58:32.999068 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "7bfcd0d6-c140-4c51-aa35-8acd8a96d68f" (UID: "7bfcd0d6-c140-4c51-aa35-8acd8a96d68f"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.032213 4810 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.032253 4810 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.032266 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.032280 4810 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.032293 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.032304 4810 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.032314 4810 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.032326 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztcr7\" (UniqueName: \"kubernetes.io/projected/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-kube-api-access-ztcr7\") on node \"crc\" DevicePath \"\"" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.032337 4810 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7bfcd0d6-c140-4c51-aa35-8acd8a96d68f-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.354293 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" event={"ID":"7bfcd0d6-c140-4c51-aa35-8acd8a96d68f","Type":"ContainerDied","Data":"388eec5629f089a426f76fd3ec4a0eddb8479c73c4049047201254a643a878ee"} Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.354664 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="388eec5629f089a426f76fd3ec4a0eddb8479c73c4049047201254a643a878ee" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.354334 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-8cbjx" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.356736 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="58ed6efdf6a9eee8fc2c0ff2afbf931ed73ba5c194f897a984932d52cef19b8d" exitCode=0 Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.356774 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"58ed6efdf6a9eee8fc2c0ff2afbf931ed73ba5c194f897a984932d52cef19b8d"} Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.356800 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798"} Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.356817 4810 scope.go:117] "RemoveContainer" containerID="c7a8274a6ece1d98030b50bcac59fe0a289cdf7d56bdf1b77781fec8041e27c3" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.467134 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-kp8l4"] Dec 01 16:58:33 crc kubenswrapper[4810]: E1201 16:58:33.467716 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b55d01e7-b6ef-498e-a55e-3dbcfb19945c" containerName="extract-utilities" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.467740 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b55d01e7-b6ef-498e-a55e-3dbcfb19945c" containerName="extract-utilities" Dec 01 16:58:33 crc kubenswrapper[4810]: E1201 16:58:33.467770 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b55d01e7-b6ef-498e-a55e-3dbcfb19945c" containerName="registry-server" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.467780 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b55d01e7-b6ef-498e-a55e-3dbcfb19945c" containerName="registry-server" Dec 01 16:58:33 crc kubenswrapper[4810]: E1201 16:58:33.467816 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bfcd0d6-c140-4c51-aa35-8acd8a96d68f" containerName="nova-cell1-openstack-openstack-cell1" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.467825 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bfcd0d6-c140-4c51-aa35-8acd8a96d68f" containerName="nova-cell1-openstack-openstack-cell1" Dec 01 16:58:33 crc kubenswrapper[4810]: E1201 16:58:33.467852 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b55d01e7-b6ef-498e-a55e-3dbcfb19945c" containerName="extract-content" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.467859 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b55d01e7-b6ef-498e-a55e-3dbcfb19945c" containerName="extract-content" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.468120 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bfcd0d6-c140-4c51-aa35-8acd8a96d68f" containerName="nova-cell1-openstack-openstack-cell1" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.468150 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b55d01e7-b6ef-498e-a55e-3dbcfb19945c" containerName="registry-server" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.469150 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.473568 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.474970 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.475235 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.477021 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.478249 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-kp8l4"] Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.478765 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.542456 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.542589 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ssh-key\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.542612 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.542682 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.542745 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.544088 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wfkk\" (UniqueName: \"kubernetes.io/projected/d35d5ae0-9a24-4816-80d6-b25a71f73a52-kube-api-access-5wfkk\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.544198 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-inventory\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.645790 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.646226 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wfkk\" (UniqueName: \"kubernetes.io/projected/d35d5ae0-9a24-4816-80d6-b25a71f73a52-kube-api-access-5wfkk\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.646276 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-inventory\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.646322 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.646443 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ssh-key\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.646501 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.646565 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.651565 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.651670 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-inventory\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.651844 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ssh-key\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.652691 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.654719 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.656750 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.664742 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wfkk\" (UniqueName: \"kubernetes.io/projected/d35d5ae0-9a24-4816-80d6-b25a71f73a52-kube-api-access-5wfkk\") pod \"telemetry-openstack-openstack-cell1-kp8l4\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:33 crc kubenswrapper[4810]: I1201 16:58:33.807255 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 16:58:34 crc kubenswrapper[4810]: I1201 16:58:34.365495 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-kp8l4"] Dec 01 16:58:34 crc kubenswrapper[4810]: W1201 16:58:34.375099 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd35d5ae0_9a24_4816_80d6_b25a71f73a52.slice/crio-f0098268a87051d8f6a79407d8ddf6d4a0e1cd8a0bb338543aa452bdc5ff3ee1 WatchSource:0}: Error finding container f0098268a87051d8f6a79407d8ddf6d4a0e1cd8a0bb338543aa452bdc5ff3ee1: Status 404 returned error can't find the container with id f0098268a87051d8f6a79407d8ddf6d4a0e1cd8a0bb338543aa452bdc5ff3ee1 Dec 01 16:58:35 crc kubenswrapper[4810]: I1201 16:58:35.378940 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" event={"ID":"d35d5ae0-9a24-4816-80d6-b25a71f73a52","Type":"ContainerStarted","Data":"c2ce945f965947592dcc1f2b828b0439dd8c5b403df8f63a1dee658b773c5744"} Dec 01 16:58:35 crc kubenswrapper[4810]: I1201 16:58:35.379525 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" event={"ID":"d35d5ae0-9a24-4816-80d6-b25a71f73a52","Type":"ContainerStarted","Data":"f0098268a87051d8f6a79407d8ddf6d4a0e1cd8a0bb338543aa452bdc5ff3ee1"} Dec 01 16:58:35 crc kubenswrapper[4810]: I1201 16:58:35.399758 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" podStartSLOduration=1.866356584 podStartE2EDuration="2.399740453s" podCreationTimestamp="2025-12-01 16:58:33 +0000 UTC" firstStartedPulling="2025-12-01 16:58:34.377270776 +0000 UTC m=+8680.140780379" lastFinishedPulling="2025-12-01 16:58:34.910654595 +0000 UTC m=+8680.674164248" observedRunningTime="2025-12-01 16:58:35.395520179 +0000 UTC m=+8681.159029772" watchObservedRunningTime="2025-12-01 16:58:35.399740453 +0000 UTC m=+8681.163250056" Dec 01 16:58:45 crc kubenswrapper[4810]: I1201 16:58:45.159150 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cjqkf"] Dec 01 16:58:45 crc kubenswrapper[4810]: I1201 16:58:45.161809 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cjqkf" Dec 01 16:58:45 crc kubenswrapper[4810]: I1201 16:58:45.171135 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cjqkf"] Dec 01 16:58:45 crc kubenswrapper[4810]: I1201 16:58:45.210081 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec-catalog-content\") pod \"community-operators-cjqkf\" (UID: \"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec\") " pod="openshift-marketplace/community-operators-cjqkf" Dec 01 16:58:45 crc kubenswrapper[4810]: I1201 16:58:45.210172 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec-utilities\") pod \"community-operators-cjqkf\" (UID: \"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec\") " pod="openshift-marketplace/community-operators-cjqkf" Dec 01 16:58:45 crc kubenswrapper[4810]: I1201 16:58:45.210192 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fvfn\" (UniqueName: \"kubernetes.io/projected/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec-kube-api-access-6fvfn\") pod \"community-operators-cjqkf\" (UID: \"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec\") " pod="openshift-marketplace/community-operators-cjqkf" Dec 01 16:58:45 crc kubenswrapper[4810]: I1201 16:58:45.312567 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec-catalog-content\") pod \"community-operators-cjqkf\" (UID: \"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec\") " pod="openshift-marketplace/community-operators-cjqkf" Dec 01 16:58:45 crc kubenswrapper[4810]: I1201 16:58:45.312764 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec-utilities\") pod \"community-operators-cjqkf\" (UID: \"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec\") " pod="openshift-marketplace/community-operators-cjqkf" Dec 01 16:58:45 crc kubenswrapper[4810]: I1201 16:58:45.312787 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fvfn\" (UniqueName: \"kubernetes.io/projected/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec-kube-api-access-6fvfn\") pod \"community-operators-cjqkf\" (UID: \"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec\") " pod="openshift-marketplace/community-operators-cjqkf" Dec 01 16:58:45 crc kubenswrapper[4810]: I1201 16:58:45.313024 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec-catalog-content\") pod \"community-operators-cjqkf\" (UID: \"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec\") " pod="openshift-marketplace/community-operators-cjqkf" Dec 01 16:58:45 crc kubenswrapper[4810]: I1201 16:58:45.313322 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec-utilities\") pod \"community-operators-cjqkf\" (UID: \"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec\") " pod="openshift-marketplace/community-operators-cjqkf" Dec 01 16:58:45 crc kubenswrapper[4810]: I1201 16:58:45.334355 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fvfn\" (UniqueName: \"kubernetes.io/projected/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec-kube-api-access-6fvfn\") pod \"community-operators-cjqkf\" (UID: \"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec\") " pod="openshift-marketplace/community-operators-cjqkf" Dec 01 16:58:45 crc kubenswrapper[4810]: I1201 16:58:45.488581 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cjqkf" Dec 01 16:58:46 crc kubenswrapper[4810]: I1201 16:58:46.045590 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cjqkf"] Dec 01 16:58:46 crc kubenswrapper[4810]: W1201 16:58:46.046668 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec5f0b12_ac30_43d1_9a00_cbc68d86e9ec.slice/crio-72f93f447c4dc1812c402a119f92ebeff6e5aa6646cd9b233f13c6b6fa3550ac WatchSource:0}: Error finding container 72f93f447c4dc1812c402a119f92ebeff6e5aa6646cd9b233f13c6b6fa3550ac: Status 404 returned error can't find the container with id 72f93f447c4dc1812c402a119f92ebeff6e5aa6646cd9b233f13c6b6fa3550ac Dec 01 16:58:46 crc kubenswrapper[4810]: I1201 16:58:46.493691 4810 generic.go:334] "Generic (PLEG): container finished" podID="ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec" containerID="3f6890fe75b46bf2495c20bee8e66c6d328e1915214811992bff58ef7b06bab6" exitCode=0 Dec 01 16:58:46 crc kubenswrapper[4810]: I1201 16:58:46.506512 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjqkf" event={"ID":"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec","Type":"ContainerDied","Data":"3f6890fe75b46bf2495c20bee8e66c6d328e1915214811992bff58ef7b06bab6"} Dec 01 16:58:46 crc kubenswrapper[4810]: I1201 16:58:46.506564 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjqkf" event={"ID":"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec","Type":"ContainerStarted","Data":"72f93f447c4dc1812c402a119f92ebeff6e5aa6646cd9b233f13c6b6fa3550ac"} Dec 01 16:58:48 crc kubenswrapper[4810]: I1201 16:58:48.511065 4810 generic.go:334] "Generic (PLEG): container finished" podID="ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec" containerID="0bb8fd8ffbf34a6e2a3ed468ea6676d98acbdd92523d52f36b07ae074325bcd5" exitCode=0 Dec 01 16:58:48 crc kubenswrapper[4810]: I1201 16:58:48.511158 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjqkf" event={"ID":"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec","Type":"ContainerDied","Data":"0bb8fd8ffbf34a6e2a3ed468ea6676d98acbdd92523d52f36b07ae074325bcd5"} Dec 01 16:58:49 crc kubenswrapper[4810]: I1201 16:58:49.523871 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjqkf" event={"ID":"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec","Type":"ContainerStarted","Data":"54e68b6704c91f2a9406a23008e8591c7bfa079b137f5791c23962bb1974b0a4"} Dec 01 16:58:49 crc kubenswrapper[4810]: I1201 16:58:49.544265 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cjqkf" podStartSLOduration=1.9485475650000001 podStartE2EDuration="4.544241329s" podCreationTimestamp="2025-12-01 16:58:45 +0000 UTC" firstStartedPulling="2025-12-01 16:58:46.495220555 +0000 UTC m=+8692.258730148" lastFinishedPulling="2025-12-01 16:58:49.090914309 +0000 UTC m=+8694.854423912" observedRunningTime="2025-12-01 16:58:49.540660773 +0000 UTC m=+8695.304170456" watchObservedRunningTime="2025-12-01 16:58:49.544241329 +0000 UTC m=+8695.307750932" Dec 01 16:58:55 crc kubenswrapper[4810]: I1201 16:58:55.489391 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cjqkf" Dec 01 16:58:55 crc kubenswrapper[4810]: I1201 16:58:55.490083 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cjqkf" Dec 01 16:58:55 crc kubenswrapper[4810]: I1201 16:58:55.916391 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cjqkf" Dec 01 16:58:56 crc kubenswrapper[4810]: I1201 16:58:56.680325 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cjqkf" Dec 01 16:58:56 crc kubenswrapper[4810]: I1201 16:58:56.750457 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cjqkf"] Dec 01 16:58:58 crc kubenswrapper[4810]: I1201 16:58:58.638978 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cjqkf" podUID="ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec" containerName="registry-server" containerID="cri-o://54e68b6704c91f2a9406a23008e8591c7bfa079b137f5791c23962bb1974b0a4" gracePeriod=2 Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.639916 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cjqkf" Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.651698 4810 generic.go:334] "Generic (PLEG): container finished" podID="ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec" containerID="54e68b6704c91f2a9406a23008e8591c7bfa079b137f5791c23962bb1974b0a4" exitCode=0 Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.651739 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjqkf" event={"ID":"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec","Type":"ContainerDied","Data":"54e68b6704c91f2a9406a23008e8591c7bfa079b137f5791c23962bb1974b0a4"} Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.651781 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cjqkf" event={"ID":"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec","Type":"ContainerDied","Data":"72f93f447c4dc1812c402a119f92ebeff6e5aa6646cd9b233f13c6b6fa3550ac"} Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.651800 4810 scope.go:117] "RemoveContainer" containerID="54e68b6704c91f2a9406a23008e8591c7bfa079b137f5791c23962bb1974b0a4" Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.651808 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cjqkf" Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.701103 4810 scope.go:117] "RemoveContainer" containerID="0bb8fd8ffbf34a6e2a3ed468ea6676d98acbdd92523d52f36b07ae074325bcd5" Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.726960 4810 scope.go:117] "RemoveContainer" containerID="3f6890fe75b46bf2495c20bee8e66c6d328e1915214811992bff58ef7b06bab6" Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.750292 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec-utilities\") pod \"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec\" (UID: \"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec\") " Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.750380 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec-catalog-content\") pod \"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec\" (UID: \"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec\") " Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.750551 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fvfn\" (UniqueName: \"kubernetes.io/projected/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec-kube-api-access-6fvfn\") pod \"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec\" (UID: \"ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec\") " Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.752251 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec-utilities" (OuterVolumeSpecName: "utilities") pod "ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec" (UID: "ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.757726 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec-kube-api-access-6fvfn" (OuterVolumeSpecName: "kube-api-access-6fvfn") pod "ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec" (UID: "ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec"). InnerVolumeSpecName "kube-api-access-6fvfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.776867 4810 scope.go:117] "RemoveContainer" containerID="54e68b6704c91f2a9406a23008e8591c7bfa079b137f5791c23962bb1974b0a4" Dec 01 16:58:59 crc kubenswrapper[4810]: E1201 16:58:59.777353 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54e68b6704c91f2a9406a23008e8591c7bfa079b137f5791c23962bb1974b0a4\": container with ID starting with 54e68b6704c91f2a9406a23008e8591c7bfa079b137f5791c23962bb1974b0a4 not found: ID does not exist" containerID="54e68b6704c91f2a9406a23008e8591c7bfa079b137f5791c23962bb1974b0a4" Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.777406 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54e68b6704c91f2a9406a23008e8591c7bfa079b137f5791c23962bb1974b0a4"} err="failed to get container status \"54e68b6704c91f2a9406a23008e8591c7bfa079b137f5791c23962bb1974b0a4\": rpc error: code = NotFound desc = could not find container \"54e68b6704c91f2a9406a23008e8591c7bfa079b137f5791c23962bb1974b0a4\": container with ID starting with 54e68b6704c91f2a9406a23008e8591c7bfa079b137f5791c23962bb1974b0a4 not found: ID does not exist" Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.777445 4810 scope.go:117] "RemoveContainer" containerID="0bb8fd8ffbf34a6e2a3ed468ea6676d98acbdd92523d52f36b07ae074325bcd5" Dec 01 16:58:59 crc kubenswrapper[4810]: E1201 16:58:59.777759 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bb8fd8ffbf34a6e2a3ed468ea6676d98acbdd92523d52f36b07ae074325bcd5\": container with ID starting with 0bb8fd8ffbf34a6e2a3ed468ea6676d98acbdd92523d52f36b07ae074325bcd5 not found: ID does not exist" containerID="0bb8fd8ffbf34a6e2a3ed468ea6676d98acbdd92523d52f36b07ae074325bcd5" Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.777790 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bb8fd8ffbf34a6e2a3ed468ea6676d98acbdd92523d52f36b07ae074325bcd5"} err="failed to get container status \"0bb8fd8ffbf34a6e2a3ed468ea6676d98acbdd92523d52f36b07ae074325bcd5\": rpc error: code = NotFound desc = could not find container \"0bb8fd8ffbf34a6e2a3ed468ea6676d98acbdd92523d52f36b07ae074325bcd5\": container with ID starting with 0bb8fd8ffbf34a6e2a3ed468ea6676d98acbdd92523d52f36b07ae074325bcd5 not found: ID does not exist" Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.777812 4810 scope.go:117] "RemoveContainer" containerID="3f6890fe75b46bf2495c20bee8e66c6d328e1915214811992bff58ef7b06bab6" Dec 01 16:58:59 crc kubenswrapper[4810]: E1201 16:58:59.778245 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f6890fe75b46bf2495c20bee8e66c6d328e1915214811992bff58ef7b06bab6\": container with ID starting with 3f6890fe75b46bf2495c20bee8e66c6d328e1915214811992bff58ef7b06bab6 not found: ID does not exist" containerID="3f6890fe75b46bf2495c20bee8e66c6d328e1915214811992bff58ef7b06bab6" Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.778313 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f6890fe75b46bf2495c20bee8e66c6d328e1915214811992bff58ef7b06bab6"} err="failed to get container status \"3f6890fe75b46bf2495c20bee8e66c6d328e1915214811992bff58ef7b06bab6\": rpc error: code = NotFound desc = could not find container \"3f6890fe75b46bf2495c20bee8e66c6d328e1915214811992bff58ef7b06bab6\": container with ID starting with 3f6890fe75b46bf2495c20bee8e66c6d328e1915214811992bff58ef7b06bab6 not found: ID does not exist" Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.813531 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec" (UID: "ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.853718 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.853750 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.853761 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fvfn\" (UniqueName: \"kubernetes.io/projected/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec-kube-api-access-6fvfn\") on node \"crc\" DevicePath \"\"" Dec 01 16:58:59 crc kubenswrapper[4810]: I1201 16:58:59.995833 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cjqkf"] Dec 01 16:59:00 crc kubenswrapper[4810]: I1201 16:59:00.005083 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cjqkf"] Dec 01 16:59:00 crc kubenswrapper[4810]: I1201 16:59:00.505244 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec" path="/var/lib/kubelet/pods/ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec/volumes" Dec 01 16:59:36 crc kubenswrapper[4810]: I1201 16:59:36.743689 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lrznn"] Dec 01 16:59:36 crc kubenswrapper[4810]: E1201 16:59:36.744660 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec" containerName="extract-content" Dec 01 16:59:36 crc kubenswrapper[4810]: I1201 16:59:36.744673 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec" containerName="extract-content" Dec 01 16:59:36 crc kubenswrapper[4810]: E1201 16:59:36.744726 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec" containerName="extract-utilities" Dec 01 16:59:36 crc kubenswrapper[4810]: I1201 16:59:36.744732 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec" containerName="extract-utilities" Dec 01 16:59:36 crc kubenswrapper[4810]: E1201 16:59:36.744740 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec" containerName="registry-server" Dec 01 16:59:36 crc kubenswrapper[4810]: I1201 16:59:36.744747 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec" containerName="registry-server" Dec 01 16:59:36 crc kubenswrapper[4810]: I1201 16:59:36.744943 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec5f0b12-ac30-43d1-9a00-cbc68d86e9ec" containerName="registry-server" Dec 01 16:59:36 crc kubenswrapper[4810]: I1201 16:59:36.746571 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrznn" Dec 01 16:59:36 crc kubenswrapper[4810]: I1201 16:59:36.764908 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lrznn"] Dec 01 16:59:36 crc kubenswrapper[4810]: I1201 16:59:36.842671 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061-utilities\") pod \"redhat-operators-lrznn\" (UID: \"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061\") " pod="openshift-marketplace/redhat-operators-lrznn" Dec 01 16:59:36 crc kubenswrapper[4810]: I1201 16:59:36.842738 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncj74\" (UniqueName: \"kubernetes.io/projected/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061-kube-api-access-ncj74\") pod \"redhat-operators-lrznn\" (UID: \"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061\") " pod="openshift-marketplace/redhat-operators-lrznn" Dec 01 16:59:36 crc kubenswrapper[4810]: I1201 16:59:36.842775 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061-catalog-content\") pod \"redhat-operators-lrznn\" (UID: \"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061\") " pod="openshift-marketplace/redhat-operators-lrznn" Dec 01 16:59:36 crc kubenswrapper[4810]: I1201 16:59:36.944778 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061-utilities\") pod \"redhat-operators-lrznn\" (UID: \"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061\") " pod="openshift-marketplace/redhat-operators-lrznn" Dec 01 16:59:36 crc kubenswrapper[4810]: I1201 16:59:36.944857 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncj74\" (UniqueName: \"kubernetes.io/projected/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061-kube-api-access-ncj74\") pod \"redhat-operators-lrznn\" (UID: \"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061\") " pod="openshift-marketplace/redhat-operators-lrznn" Dec 01 16:59:36 crc kubenswrapper[4810]: I1201 16:59:36.944886 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061-catalog-content\") pod \"redhat-operators-lrznn\" (UID: \"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061\") " pod="openshift-marketplace/redhat-operators-lrznn" Dec 01 16:59:36 crc kubenswrapper[4810]: I1201 16:59:36.945380 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061-utilities\") pod \"redhat-operators-lrznn\" (UID: \"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061\") " pod="openshift-marketplace/redhat-operators-lrznn" Dec 01 16:59:36 crc kubenswrapper[4810]: I1201 16:59:36.945409 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061-catalog-content\") pod \"redhat-operators-lrznn\" (UID: \"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061\") " pod="openshift-marketplace/redhat-operators-lrznn" Dec 01 16:59:36 crc kubenswrapper[4810]: I1201 16:59:36.967139 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncj74\" (UniqueName: \"kubernetes.io/projected/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061-kube-api-access-ncj74\") pod \"redhat-operators-lrznn\" (UID: \"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061\") " pod="openshift-marketplace/redhat-operators-lrznn" Dec 01 16:59:37 crc kubenswrapper[4810]: I1201 16:59:37.079814 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrznn" Dec 01 16:59:37 crc kubenswrapper[4810]: I1201 16:59:37.587771 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lrznn"] Dec 01 16:59:38 crc kubenswrapper[4810]: I1201 16:59:38.186784 4810 generic.go:334] "Generic (PLEG): container finished" podID="fb5a3628-f3e0-48a2-8a1b-aa7b55a90061" containerID="e82f05018561b4d4cf4b1c4bbc593e8a17c5a0f86201661a9c3c4ba66b7924c8" exitCode=0 Dec 01 16:59:38 crc kubenswrapper[4810]: I1201 16:59:38.187096 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrznn" event={"ID":"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061","Type":"ContainerDied","Data":"e82f05018561b4d4cf4b1c4bbc593e8a17c5a0f86201661a9c3c4ba66b7924c8"} Dec 01 16:59:38 crc kubenswrapper[4810]: I1201 16:59:38.187123 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrznn" event={"ID":"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061","Type":"ContainerStarted","Data":"b70c3867de47373641061dd93068783dac23be6bc001ca0ff820cdccc1146ae6"} Dec 01 16:59:39 crc kubenswrapper[4810]: I1201 16:59:39.201975 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrznn" event={"ID":"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061","Type":"ContainerStarted","Data":"570ee0008782d15d49dc83dc418342d56bdb7b004dc355d7749a2121d29757bd"} Dec 01 16:59:42 crc kubenswrapper[4810]: I1201 16:59:42.233979 4810 generic.go:334] "Generic (PLEG): container finished" podID="fb5a3628-f3e0-48a2-8a1b-aa7b55a90061" containerID="570ee0008782d15d49dc83dc418342d56bdb7b004dc355d7749a2121d29757bd" exitCode=0 Dec 01 16:59:42 crc kubenswrapper[4810]: I1201 16:59:42.234047 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrznn" event={"ID":"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061","Type":"ContainerDied","Data":"570ee0008782d15d49dc83dc418342d56bdb7b004dc355d7749a2121d29757bd"} Dec 01 16:59:43 crc kubenswrapper[4810]: I1201 16:59:43.247099 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrznn" event={"ID":"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061","Type":"ContainerStarted","Data":"d620497c6a8e3ca7677b0dcf14966d1872aa9b89b49aa276f61894d641a554b2"} Dec 01 16:59:43 crc kubenswrapper[4810]: I1201 16:59:43.287953 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lrznn" podStartSLOduration=2.7218434289999998 podStartE2EDuration="7.287930058s" podCreationTimestamp="2025-12-01 16:59:36 +0000 UTC" firstStartedPulling="2025-12-01 16:59:38.1884446 +0000 UTC m=+8743.951954203" lastFinishedPulling="2025-12-01 16:59:42.754531229 +0000 UTC m=+8748.518040832" observedRunningTime="2025-12-01 16:59:43.275839851 +0000 UTC m=+8749.039349454" watchObservedRunningTime="2025-12-01 16:59:43.287930058 +0000 UTC m=+8749.051439661" Dec 01 16:59:47 crc kubenswrapper[4810]: I1201 16:59:47.080499 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lrznn" Dec 01 16:59:47 crc kubenswrapper[4810]: I1201 16:59:47.080818 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lrznn" Dec 01 16:59:48 crc kubenswrapper[4810]: I1201 16:59:48.134754 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lrznn" podUID="fb5a3628-f3e0-48a2-8a1b-aa7b55a90061" containerName="registry-server" probeResult="failure" output=< Dec 01 16:59:48 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Dec 01 16:59:48 crc kubenswrapper[4810]: > Dec 01 16:59:57 crc kubenswrapper[4810]: I1201 16:59:57.149524 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lrznn" Dec 01 16:59:57 crc kubenswrapper[4810]: I1201 16:59:57.452135 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lrznn" Dec 01 16:59:57 crc kubenswrapper[4810]: I1201 16:59:57.507123 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lrznn"] Dec 01 16:59:58 crc kubenswrapper[4810]: I1201 16:59:58.407151 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lrznn" podUID="fb5a3628-f3e0-48a2-8a1b-aa7b55a90061" containerName="registry-server" containerID="cri-o://d620497c6a8e3ca7677b0dcf14966d1872aa9b89b49aa276f61894d641a554b2" gracePeriod=2 Dec 01 16:59:59 crc kubenswrapper[4810]: I1201 16:59:59.429012 4810 generic.go:334] "Generic (PLEG): container finished" podID="fb5a3628-f3e0-48a2-8a1b-aa7b55a90061" containerID="d620497c6a8e3ca7677b0dcf14966d1872aa9b89b49aa276f61894d641a554b2" exitCode=0 Dec 01 16:59:59 crc kubenswrapper[4810]: I1201 16:59:59.429207 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrznn" event={"ID":"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061","Type":"ContainerDied","Data":"d620497c6a8e3ca7677b0dcf14966d1872aa9b89b49aa276f61894d641a554b2"} Dec 01 16:59:59 crc kubenswrapper[4810]: I1201 16:59:59.429352 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lrznn" event={"ID":"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061","Type":"ContainerDied","Data":"b70c3867de47373641061dd93068783dac23be6bc001ca0ff820cdccc1146ae6"} Dec 01 16:59:59 crc kubenswrapper[4810]: I1201 16:59:59.429377 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b70c3867de47373641061dd93068783dac23be6bc001ca0ff820cdccc1146ae6" Dec 01 16:59:59 crc kubenswrapper[4810]: I1201 16:59:59.600871 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrznn" Dec 01 16:59:59 crc kubenswrapper[4810]: I1201 16:59:59.705743 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061-utilities\") pod \"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061\" (UID: \"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061\") " Dec 01 16:59:59 crc kubenswrapper[4810]: I1201 16:59:59.705913 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061-catalog-content\") pod \"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061\" (UID: \"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061\") " Dec 01 16:59:59 crc kubenswrapper[4810]: I1201 16:59:59.705959 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncj74\" (UniqueName: \"kubernetes.io/projected/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061-kube-api-access-ncj74\") pod \"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061\" (UID: \"fb5a3628-f3e0-48a2-8a1b-aa7b55a90061\") " Dec 01 16:59:59 crc kubenswrapper[4810]: I1201 16:59:59.707070 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061-utilities" (OuterVolumeSpecName: "utilities") pod "fb5a3628-f3e0-48a2-8a1b-aa7b55a90061" (UID: "fb5a3628-f3e0-48a2-8a1b-aa7b55a90061"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:59:59 crc kubenswrapper[4810]: I1201 16:59:59.711953 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061-kube-api-access-ncj74" (OuterVolumeSpecName: "kube-api-access-ncj74") pod "fb5a3628-f3e0-48a2-8a1b-aa7b55a90061" (UID: "fb5a3628-f3e0-48a2-8a1b-aa7b55a90061"). InnerVolumeSpecName "kube-api-access-ncj74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 16:59:59 crc kubenswrapper[4810]: I1201 16:59:59.807896 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 16:59:59 crc kubenswrapper[4810]: I1201 16:59:59.807934 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncj74\" (UniqueName: \"kubernetes.io/projected/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061-kube-api-access-ncj74\") on node \"crc\" DevicePath \"\"" Dec 01 16:59:59 crc kubenswrapper[4810]: I1201 16:59:59.824153 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fb5a3628-f3e0-48a2-8a1b-aa7b55a90061" (UID: "fb5a3628-f3e0-48a2-8a1b-aa7b55a90061"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 16:59:59 crc kubenswrapper[4810]: I1201 16:59:59.909708 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.160288 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7"] Dec 01 17:00:00 crc kubenswrapper[4810]: E1201 17:00:00.160978 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb5a3628-f3e0-48a2-8a1b-aa7b55a90061" containerName="extract-content" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.161006 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb5a3628-f3e0-48a2-8a1b-aa7b55a90061" containerName="extract-content" Dec 01 17:00:00 crc kubenswrapper[4810]: E1201 17:00:00.161020 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb5a3628-f3e0-48a2-8a1b-aa7b55a90061" containerName="registry-server" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.161027 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb5a3628-f3e0-48a2-8a1b-aa7b55a90061" containerName="registry-server" Dec 01 17:00:00 crc kubenswrapper[4810]: E1201 17:00:00.161036 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb5a3628-f3e0-48a2-8a1b-aa7b55a90061" containerName="extract-utilities" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.161043 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb5a3628-f3e0-48a2-8a1b-aa7b55a90061" containerName="extract-utilities" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.161253 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb5a3628-f3e0-48a2-8a1b-aa7b55a90061" containerName="registry-server" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.162010 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.167772 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.170921 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.181664 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7"] Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.214782 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2wsj\" (UniqueName: \"kubernetes.io/projected/f8b531aa-593f-474c-8520-63df40c2395b-kube-api-access-v2wsj\") pod \"collect-profiles-29410140-xchr7\" (UID: \"f8b531aa-593f-474c-8520-63df40c2395b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.214932 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8b531aa-593f-474c-8520-63df40c2395b-config-volume\") pod \"collect-profiles-29410140-xchr7\" (UID: \"f8b531aa-593f-474c-8520-63df40c2395b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.214997 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8b531aa-593f-474c-8520-63df40c2395b-secret-volume\") pod \"collect-profiles-29410140-xchr7\" (UID: \"f8b531aa-593f-474c-8520-63df40c2395b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.316343 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2wsj\" (UniqueName: \"kubernetes.io/projected/f8b531aa-593f-474c-8520-63df40c2395b-kube-api-access-v2wsj\") pod \"collect-profiles-29410140-xchr7\" (UID: \"f8b531aa-593f-474c-8520-63df40c2395b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.316940 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8b531aa-593f-474c-8520-63df40c2395b-config-volume\") pod \"collect-profiles-29410140-xchr7\" (UID: \"f8b531aa-593f-474c-8520-63df40c2395b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.317061 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8b531aa-593f-474c-8520-63df40c2395b-secret-volume\") pod \"collect-profiles-29410140-xchr7\" (UID: \"f8b531aa-593f-474c-8520-63df40c2395b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.317870 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8b531aa-593f-474c-8520-63df40c2395b-config-volume\") pod \"collect-profiles-29410140-xchr7\" (UID: \"f8b531aa-593f-474c-8520-63df40c2395b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.324069 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8b531aa-593f-474c-8520-63df40c2395b-secret-volume\") pod \"collect-profiles-29410140-xchr7\" (UID: \"f8b531aa-593f-474c-8520-63df40c2395b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.332728 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2wsj\" (UniqueName: \"kubernetes.io/projected/f8b531aa-593f-474c-8520-63df40c2395b-kube-api-access-v2wsj\") pod \"collect-profiles-29410140-xchr7\" (UID: \"f8b531aa-593f-474c-8520-63df40c2395b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.438974 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lrznn" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.480773 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lrznn"] Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.482585 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7" Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.510866 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lrznn"] Dec 01 17:00:00 crc kubenswrapper[4810]: I1201 17:00:00.963036 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7"] Dec 01 17:00:01 crc kubenswrapper[4810]: W1201 17:00:01.257651 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8b531aa_593f_474c_8520_63df40c2395b.slice/crio-2fb06f823326093db483586a7b664dc110c9c11c007cade8412b0fccfee1be59 WatchSource:0}: Error finding container 2fb06f823326093db483586a7b664dc110c9c11c007cade8412b0fccfee1be59: Status 404 returned error can't find the container with id 2fb06f823326093db483586a7b664dc110c9c11c007cade8412b0fccfee1be59 Dec 01 17:00:01 crc kubenswrapper[4810]: I1201 17:00:01.459696 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7" event={"ID":"f8b531aa-593f-474c-8520-63df40c2395b","Type":"ContainerStarted","Data":"2fb06f823326093db483586a7b664dc110c9c11c007cade8412b0fccfee1be59"} Dec 01 17:00:02 crc kubenswrapper[4810]: I1201 17:00:02.471029 4810 generic.go:334] "Generic (PLEG): container finished" podID="f8b531aa-593f-474c-8520-63df40c2395b" containerID="a51f8e9ebbb50478ee9c1436fbbd0623a0a83251bdc94b10e1e30b81995923da" exitCode=0 Dec 01 17:00:02 crc kubenswrapper[4810]: I1201 17:00:02.471104 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7" event={"ID":"f8b531aa-593f-474c-8520-63df40c2395b","Type":"ContainerDied","Data":"a51f8e9ebbb50478ee9c1436fbbd0623a0a83251bdc94b10e1e30b81995923da"} Dec 01 17:00:02 crc kubenswrapper[4810]: I1201 17:00:02.503142 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb5a3628-f3e0-48a2-8a1b-aa7b55a90061" path="/var/lib/kubelet/pods/fb5a3628-f3e0-48a2-8a1b-aa7b55a90061/volumes" Dec 01 17:00:03 crc kubenswrapper[4810]: I1201 17:00:03.865869 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7" Dec 01 17:00:03 crc kubenswrapper[4810]: I1201 17:00:03.902110 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8b531aa-593f-474c-8520-63df40c2395b-secret-volume\") pod \"f8b531aa-593f-474c-8520-63df40c2395b\" (UID: \"f8b531aa-593f-474c-8520-63df40c2395b\") " Dec 01 17:00:03 crc kubenswrapper[4810]: I1201 17:00:03.902425 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8b531aa-593f-474c-8520-63df40c2395b-config-volume\") pod \"f8b531aa-593f-474c-8520-63df40c2395b\" (UID: \"f8b531aa-593f-474c-8520-63df40c2395b\") " Dec 01 17:00:03 crc kubenswrapper[4810]: I1201 17:00:03.902754 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2wsj\" (UniqueName: \"kubernetes.io/projected/f8b531aa-593f-474c-8520-63df40c2395b-kube-api-access-v2wsj\") pod \"f8b531aa-593f-474c-8520-63df40c2395b\" (UID: \"f8b531aa-593f-474c-8520-63df40c2395b\") " Dec 01 17:00:03 crc kubenswrapper[4810]: I1201 17:00:03.907247 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8b531aa-593f-474c-8520-63df40c2395b-config-volume" (OuterVolumeSpecName: "config-volume") pod "f8b531aa-593f-474c-8520-63df40c2395b" (UID: "f8b531aa-593f-474c-8520-63df40c2395b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:00:03 crc kubenswrapper[4810]: I1201 17:00:03.910433 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8b531aa-593f-474c-8520-63df40c2395b-kube-api-access-v2wsj" (OuterVolumeSpecName: "kube-api-access-v2wsj") pod "f8b531aa-593f-474c-8520-63df40c2395b" (UID: "f8b531aa-593f-474c-8520-63df40c2395b"). InnerVolumeSpecName "kube-api-access-v2wsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:00:03 crc kubenswrapper[4810]: I1201 17:00:03.910587 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8b531aa-593f-474c-8520-63df40c2395b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f8b531aa-593f-474c-8520-63df40c2395b" (UID: "f8b531aa-593f-474c-8520-63df40c2395b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:00:04 crc kubenswrapper[4810]: I1201 17:00:04.005325 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2wsj\" (UniqueName: \"kubernetes.io/projected/f8b531aa-593f-474c-8520-63df40c2395b-kube-api-access-v2wsj\") on node \"crc\" DevicePath \"\"" Dec 01 17:00:04 crc kubenswrapper[4810]: I1201 17:00:04.005367 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f8b531aa-593f-474c-8520-63df40c2395b-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 17:00:04 crc kubenswrapper[4810]: I1201 17:00:04.005380 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f8b531aa-593f-474c-8520-63df40c2395b-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 17:00:04 crc kubenswrapper[4810]: I1201 17:00:04.504177 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7" Dec 01 17:00:04 crc kubenswrapper[4810]: I1201 17:00:04.507016 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410140-xchr7" event={"ID":"f8b531aa-593f-474c-8520-63df40c2395b","Type":"ContainerDied","Data":"2fb06f823326093db483586a7b664dc110c9c11c007cade8412b0fccfee1be59"} Dec 01 17:00:04 crc kubenswrapper[4810]: I1201 17:00:04.507086 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fb06f823326093db483586a7b664dc110c9c11c007cade8412b0fccfee1be59" Dec 01 17:00:04 crc kubenswrapper[4810]: I1201 17:00:04.945054 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q"] Dec 01 17:00:04 crc kubenswrapper[4810]: I1201 17:00:04.959210 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410095-b7b6q"] Dec 01 17:00:06 crc kubenswrapper[4810]: I1201 17:00:06.505442 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ef74083-dfac-48d7-9f66-77dd31086f68" path="/var/lib/kubelet/pods/8ef74083-dfac-48d7-9f66-77dd31086f68/volumes" Dec 01 17:00:17 crc kubenswrapper[4810]: I1201 17:00:17.841770 4810 scope.go:117] "RemoveContainer" containerID="07f9b8592d9274a5c91042bff038a1f8b35cf054f99282d72115a68796255d82" Dec 01 17:00:56 crc kubenswrapper[4810]: I1201 17:00:56.077990 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-522ww"] Dec 01 17:00:56 crc kubenswrapper[4810]: E1201 17:00:56.079151 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8b531aa-593f-474c-8520-63df40c2395b" containerName="collect-profiles" Dec 01 17:00:56 crc kubenswrapper[4810]: I1201 17:00:56.079173 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8b531aa-593f-474c-8520-63df40c2395b" containerName="collect-profiles" Dec 01 17:00:56 crc kubenswrapper[4810]: I1201 17:00:56.079508 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8b531aa-593f-474c-8520-63df40c2395b" containerName="collect-profiles" Dec 01 17:00:56 crc kubenswrapper[4810]: I1201 17:00:56.084970 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-522ww" Dec 01 17:00:56 crc kubenswrapper[4810]: I1201 17:00:56.099890 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-522ww"] Dec 01 17:00:56 crc kubenswrapper[4810]: I1201 17:00:56.210410 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9bl6\" (UniqueName: \"kubernetes.io/projected/c78c4f30-a933-4cef-8b91-a5af5a3112b8-kube-api-access-m9bl6\") pod \"redhat-marketplace-522ww\" (UID: \"c78c4f30-a933-4cef-8b91-a5af5a3112b8\") " pod="openshift-marketplace/redhat-marketplace-522ww" Dec 01 17:00:56 crc kubenswrapper[4810]: I1201 17:00:56.210775 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c78c4f30-a933-4cef-8b91-a5af5a3112b8-catalog-content\") pod \"redhat-marketplace-522ww\" (UID: \"c78c4f30-a933-4cef-8b91-a5af5a3112b8\") " pod="openshift-marketplace/redhat-marketplace-522ww" Dec 01 17:00:56 crc kubenswrapper[4810]: I1201 17:00:56.210926 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c78c4f30-a933-4cef-8b91-a5af5a3112b8-utilities\") pod \"redhat-marketplace-522ww\" (UID: \"c78c4f30-a933-4cef-8b91-a5af5a3112b8\") " pod="openshift-marketplace/redhat-marketplace-522ww" Dec 01 17:00:56 crc kubenswrapper[4810]: I1201 17:00:56.313107 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9bl6\" (UniqueName: \"kubernetes.io/projected/c78c4f30-a933-4cef-8b91-a5af5a3112b8-kube-api-access-m9bl6\") pod \"redhat-marketplace-522ww\" (UID: \"c78c4f30-a933-4cef-8b91-a5af5a3112b8\") " pod="openshift-marketplace/redhat-marketplace-522ww" Dec 01 17:00:56 crc kubenswrapper[4810]: I1201 17:00:56.313208 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c78c4f30-a933-4cef-8b91-a5af5a3112b8-catalog-content\") pod \"redhat-marketplace-522ww\" (UID: \"c78c4f30-a933-4cef-8b91-a5af5a3112b8\") " pod="openshift-marketplace/redhat-marketplace-522ww" Dec 01 17:00:56 crc kubenswrapper[4810]: I1201 17:00:56.313236 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c78c4f30-a933-4cef-8b91-a5af5a3112b8-utilities\") pod \"redhat-marketplace-522ww\" (UID: \"c78c4f30-a933-4cef-8b91-a5af5a3112b8\") " pod="openshift-marketplace/redhat-marketplace-522ww" Dec 01 17:00:56 crc kubenswrapper[4810]: I1201 17:00:56.314000 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c78c4f30-a933-4cef-8b91-a5af5a3112b8-utilities\") pod \"redhat-marketplace-522ww\" (UID: \"c78c4f30-a933-4cef-8b91-a5af5a3112b8\") " pod="openshift-marketplace/redhat-marketplace-522ww" Dec 01 17:00:56 crc kubenswrapper[4810]: I1201 17:00:56.314081 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c78c4f30-a933-4cef-8b91-a5af5a3112b8-catalog-content\") pod \"redhat-marketplace-522ww\" (UID: \"c78c4f30-a933-4cef-8b91-a5af5a3112b8\") " pod="openshift-marketplace/redhat-marketplace-522ww" Dec 01 17:00:56 crc kubenswrapper[4810]: I1201 17:00:56.340596 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9bl6\" (UniqueName: \"kubernetes.io/projected/c78c4f30-a933-4cef-8b91-a5af5a3112b8-kube-api-access-m9bl6\") pod \"redhat-marketplace-522ww\" (UID: \"c78c4f30-a933-4cef-8b91-a5af5a3112b8\") " pod="openshift-marketplace/redhat-marketplace-522ww" Dec 01 17:00:56 crc kubenswrapper[4810]: I1201 17:00:56.416415 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-522ww" Dec 01 17:00:56 crc kubenswrapper[4810]: I1201 17:00:56.962764 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-522ww"] Dec 01 17:00:58 crc kubenswrapper[4810]: I1201 17:00:58.154330 4810 generic.go:334] "Generic (PLEG): container finished" podID="c78c4f30-a933-4cef-8b91-a5af5a3112b8" containerID="6bf38fec9cbe3505db30a7c614a9faf2cc973fe338153fbb3ed19112168043c0" exitCode=0 Dec 01 17:00:58 crc kubenswrapper[4810]: I1201 17:00:58.154391 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-522ww" event={"ID":"c78c4f30-a933-4cef-8b91-a5af5a3112b8","Type":"ContainerDied","Data":"6bf38fec9cbe3505db30a7c614a9faf2cc973fe338153fbb3ed19112168043c0"} Dec 01 17:00:58 crc kubenswrapper[4810]: I1201 17:00:58.154899 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-522ww" event={"ID":"c78c4f30-a933-4cef-8b91-a5af5a3112b8","Type":"ContainerStarted","Data":"0767d8e5b92aab0fe6f869f7c1abf9fb09c398cee2c4bccbfe8700aea8d26441"} Dec 01 17:00:58 crc kubenswrapper[4810]: I1201 17:00:58.157713 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 17:01:00 crc kubenswrapper[4810]: I1201 17:01:00.157780 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29410141-lkmhs"] Dec 01 17:01:00 crc kubenswrapper[4810]: I1201 17:01:00.159716 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410141-lkmhs" Dec 01 17:01:00 crc kubenswrapper[4810]: I1201 17:01:00.202383 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-522ww" event={"ID":"c78c4f30-a933-4cef-8b91-a5af5a3112b8","Type":"ContainerStarted","Data":"f670591f0cfa8049a49dfb0d64706f7010df3137300af6fd2c1685246107a2c5"} Dec 01 17:01:00 crc kubenswrapper[4810]: I1201 17:01:00.212657 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29410141-lkmhs"] Dec 01 17:01:00 crc kubenswrapper[4810]: I1201 17:01:00.304268 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a37de29-2b57-4f07-b0d2-cf186488012e-combined-ca-bundle\") pod \"keystone-cron-29410141-lkmhs\" (UID: \"4a37de29-2b57-4f07-b0d2-cf186488012e\") " pod="openstack/keystone-cron-29410141-lkmhs" Dec 01 17:01:00 crc kubenswrapper[4810]: I1201 17:01:00.304329 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a37de29-2b57-4f07-b0d2-cf186488012e-fernet-keys\") pod \"keystone-cron-29410141-lkmhs\" (UID: \"4a37de29-2b57-4f07-b0d2-cf186488012e\") " pod="openstack/keystone-cron-29410141-lkmhs" Dec 01 17:01:00 crc kubenswrapper[4810]: I1201 17:01:00.304604 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a37de29-2b57-4f07-b0d2-cf186488012e-config-data\") pod \"keystone-cron-29410141-lkmhs\" (UID: \"4a37de29-2b57-4f07-b0d2-cf186488012e\") " pod="openstack/keystone-cron-29410141-lkmhs" Dec 01 17:01:00 crc kubenswrapper[4810]: I1201 17:01:00.304711 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bzf8\" (UniqueName: \"kubernetes.io/projected/4a37de29-2b57-4f07-b0d2-cf186488012e-kube-api-access-4bzf8\") pod \"keystone-cron-29410141-lkmhs\" (UID: \"4a37de29-2b57-4f07-b0d2-cf186488012e\") " pod="openstack/keystone-cron-29410141-lkmhs" Dec 01 17:01:00 crc kubenswrapper[4810]: I1201 17:01:00.406626 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a37de29-2b57-4f07-b0d2-cf186488012e-combined-ca-bundle\") pod \"keystone-cron-29410141-lkmhs\" (UID: \"4a37de29-2b57-4f07-b0d2-cf186488012e\") " pod="openstack/keystone-cron-29410141-lkmhs" Dec 01 17:01:00 crc kubenswrapper[4810]: I1201 17:01:00.406684 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a37de29-2b57-4f07-b0d2-cf186488012e-fernet-keys\") pod \"keystone-cron-29410141-lkmhs\" (UID: \"4a37de29-2b57-4f07-b0d2-cf186488012e\") " pod="openstack/keystone-cron-29410141-lkmhs" Dec 01 17:01:00 crc kubenswrapper[4810]: I1201 17:01:00.406742 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a37de29-2b57-4f07-b0d2-cf186488012e-config-data\") pod \"keystone-cron-29410141-lkmhs\" (UID: \"4a37de29-2b57-4f07-b0d2-cf186488012e\") " pod="openstack/keystone-cron-29410141-lkmhs" Dec 01 17:01:00 crc kubenswrapper[4810]: I1201 17:01:00.406765 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bzf8\" (UniqueName: \"kubernetes.io/projected/4a37de29-2b57-4f07-b0d2-cf186488012e-kube-api-access-4bzf8\") pod \"keystone-cron-29410141-lkmhs\" (UID: \"4a37de29-2b57-4f07-b0d2-cf186488012e\") " pod="openstack/keystone-cron-29410141-lkmhs" Dec 01 17:01:00 crc kubenswrapper[4810]: I1201 17:01:00.415221 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a37de29-2b57-4f07-b0d2-cf186488012e-combined-ca-bundle\") pod \"keystone-cron-29410141-lkmhs\" (UID: \"4a37de29-2b57-4f07-b0d2-cf186488012e\") " pod="openstack/keystone-cron-29410141-lkmhs" Dec 01 17:01:00 crc kubenswrapper[4810]: I1201 17:01:00.419499 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a37de29-2b57-4f07-b0d2-cf186488012e-fernet-keys\") pod \"keystone-cron-29410141-lkmhs\" (UID: \"4a37de29-2b57-4f07-b0d2-cf186488012e\") " pod="openstack/keystone-cron-29410141-lkmhs" Dec 01 17:01:00 crc kubenswrapper[4810]: I1201 17:01:00.434400 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a37de29-2b57-4f07-b0d2-cf186488012e-config-data\") pod \"keystone-cron-29410141-lkmhs\" (UID: \"4a37de29-2b57-4f07-b0d2-cf186488012e\") " pod="openstack/keystone-cron-29410141-lkmhs" Dec 01 17:01:00 crc kubenswrapper[4810]: I1201 17:01:00.440239 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bzf8\" (UniqueName: \"kubernetes.io/projected/4a37de29-2b57-4f07-b0d2-cf186488012e-kube-api-access-4bzf8\") pod \"keystone-cron-29410141-lkmhs\" (UID: \"4a37de29-2b57-4f07-b0d2-cf186488012e\") " pod="openstack/keystone-cron-29410141-lkmhs" Dec 01 17:01:00 crc kubenswrapper[4810]: I1201 17:01:00.477462 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410141-lkmhs" Dec 01 17:01:01 crc kubenswrapper[4810]: I1201 17:01:01.015334 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29410141-lkmhs"] Dec 01 17:01:01 crc kubenswrapper[4810]: I1201 17:01:01.216824 4810 generic.go:334] "Generic (PLEG): container finished" podID="c78c4f30-a933-4cef-8b91-a5af5a3112b8" containerID="f670591f0cfa8049a49dfb0d64706f7010df3137300af6fd2c1685246107a2c5" exitCode=0 Dec 01 17:01:01 crc kubenswrapper[4810]: I1201 17:01:01.216901 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-522ww" event={"ID":"c78c4f30-a933-4cef-8b91-a5af5a3112b8","Type":"ContainerDied","Data":"f670591f0cfa8049a49dfb0d64706f7010df3137300af6fd2c1685246107a2c5"} Dec 01 17:01:02 crc kubenswrapper[4810]: I1201 17:01:02.227951 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-522ww" event={"ID":"c78c4f30-a933-4cef-8b91-a5af5a3112b8","Type":"ContainerStarted","Data":"006ad8c85f1a37314866492d48b05c3f2bc6b1516c4ced8c8f644b26f692f02c"} Dec 01 17:01:02 crc kubenswrapper[4810]: I1201 17:01:02.231172 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410141-lkmhs" event={"ID":"4a37de29-2b57-4f07-b0d2-cf186488012e","Type":"ContainerStarted","Data":"c83ea3d547f9537d8dde467bf7b504dc187e3a79bf6857a14cff2c5f22552a97"} Dec 01 17:01:02 crc kubenswrapper[4810]: I1201 17:01:02.231209 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410141-lkmhs" event={"ID":"4a37de29-2b57-4f07-b0d2-cf186488012e","Type":"ContainerStarted","Data":"75013b32c0a8eca86f73d34706ab22e28d9de0e55e33a73d66895aa64176b079"} Dec 01 17:01:02 crc kubenswrapper[4810]: I1201 17:01:02.257597 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-522ww" podStartSLOduration=2.601666537 podStartE2EDuration="6.257580821s" podCreationTimestamp="2025-12-01 17:00:56 +0000 UTC" firstStartedPulling="2025-12-01 17:00:58.157306079 +0000 UTC m=+8823.920815682" lastFinishedPulling="2025-12-01 17:01:01.813220343 +0000 UTC m=+8827.576729966" observedRunningTime="2025-12-01 17:01:02.247704592 +0000 UTC m=+8828.011214205" watchObservedRunningTime="2025-12-01 17:01:02.257580821 +0000 UTC m=+8828.021090414" Dec 01 17:01:02 crc kubenswrapper[4810]: I1201 17:01:02.972638 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:01:02 crc kubenswrapper[4810]: I1201 17:01:02.972986 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:01:04 crc kubenswrapper[4810]: I1201 17:01:04.251905 4810 generic.go:334] "Generic (PLEG): container finished" podID="4a37de29-2b57-4f07-b0d2-cf186488012e" containerID="c83ea3d547f9537d8dde467bf7b504dc187e3a79bf6857a14cff2c5f22552a97" exitCode=0 Dec 01 17:01:04 crc kubenswrapper[4810]: I1201 17:01:04.252005 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410141-lkmhs" event={"ID":"4a37de29-2b57-4f07-b0d2-cf186488012e","Type":"ContainerDied","Data":"c83ea3d547f9537d8dde467bf7b504dc187e3a79bf6857a14cff2c5f22552a97"} Dec 01 17:01:05 crc kubenswrapper[4810]: I1201 17:01:05.589800 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410141-lkmhs" Dec 01 17:01:05 crc kubenswrapper[4810]: I1201 17:01:05.628767 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a37de29-2b57-4f07-b0d2-cf186488012e-config-data\") pod \"4a37de29-2b57-4f07-b0d2-cf186488012e\" (UID: \"4a37de29-2b57-4f07-b0d2-cf186488012e\") " Dec 01 17:01:05 crc kubenswrapper[4810]: I1201 17:01:05.628836 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bzf8\" (UniqueName: \"kubernetes.io/projected/4a37de29-2b57-4f07-b0d2-cf186488012e-kube-api-access-4bzf8\") pod \"4a37de29-2b57-4f07-b0d2-cf186488012e\" (UID: \"4a37de29-2b57-4f07-b0d2-cf186488012e\") " Dec 01 17:01:05 crc kubenswrapper[4810]: I1201 17:01:05.628867 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a37de29-2b57-4f07-b0d2-cf186488012e-combined-ca-bundle\") pod \"4a37de29-2b57-4f07-b0d2-cf186488012e\" (UID: \"4a37de29-2b57-4f07-b0d2-cf186488012e\") " Dec 01 17:01:05 crc kubenswrapper[4810]: I1201 17:01:05.628937 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a37de29-2b57-4f07-b0d2-cf186488012e-fernet-keys\") pod \"4a37de29-2b57-4f07-b0d2-cf186488012e\" (UID: \"4a37de29-2b57-4f07-b0d2-cf186488012e\") " Dec 01 17:01:05 crc kubenswrapper[4810]: I1201 17:01:05.634630 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a37de29-2b57-4f07-b0d2-cf186488012e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4a37de29-2b57-4f07-b0d2-cf186488012e" (UID: "4a37de29-2b57-4f07-b0d2-cf186488012e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:01:05 crc kubenswrapper[4810]: I1201 17:01:05.647659 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a37de29-2b57-4f07-b0d2-cf186488012e-kube-api-access-4bzf8" (OuterVolumeSpecName: "kube-api-access-4bzf8") pod "4a37de29-2b57-4f07-b0d2-cf186488012e" (UID: "4a37de29-2b57-4f07-b0d2-cf186488012e"). InnerVolumeSpecName "kube-api-access-4bzf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:01:05 crc kubenswrapper[4810]: I1201 17:01:05.670964 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a37de29-2b57-4f07-b0d2-cf186488012e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a37de29-2b57-4f07-b0d2-cf186488012e" (UID: "4a37de29-2b57-4f07-b0d2-cf186488012e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:01:05 crc kubenswrapper[4810]: I1201 17:01:05.693454 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a37de29-2b57-4f07-b0d2-cf186488012e-config-data" (OuterVolumeSpecName: "config-data") pod "4a37de29-2b57-4f07-b0d2-cf186488012e" (UID: "4a37de29-2b57-4f07-b0d2-cf186488012e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:01:05 crc kubenswrapper[4810]: I1201 17:01:05.734368 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bzf8\" (UniqueName: \"kubernetes.io/projected/4a37de29-2b57-4f07-b0d2-cf186488012e-kube-api-access-4bzf8\") on node \"crc\" DevicePath \"\"" Dec 01 17:01:05 crc kubenswrapper[4810]: I1201 17:01:05.734413 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a37de29-2b57-4f07-b0d2-cf186488012e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:01:05 crc kubenswrapper[4810]: I1201 17:01:05.734426 4810 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a37de29-2b57-4f07-b0d2-cf186488012e-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 17:01:05 crc kubenswrapper[4810]: I1201 17:01:05.734439 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a37de29-2b57-4f07-b0d2-cf186488012e-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:01:06 crc kubenswrapper[4810]: I1201 17:01:06.271727 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410141-lkmhs" event={"ID":"4a37de29-2b57-4f07-b0d2-cf186488012e","Type":"ContainerDied","Data":"75013b32c0a8eca86f73d34706ab22e28d9de0e55e33a73d66895aa64176b079"} Dec 01 17:01:06 crc kubenswrapper[4810]: I1201 17:01:06.271777 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410141-lkmhs" Dec 01 17:01:06 crc kubenswrapper[4810]: I1201 17:01:06.271777 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75013b32c0a8eca86f73d34706ab22e28d9de0e55e33a73d66895aa64176b079" Dec 01 17:01:06 crc kubenswrapper[4810]: I1201 17:01:06.417391 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-522ww" Dec 01 17:01:06 crc kubenswrapper[4810]: I1201 17:01:06.417776 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-522ww" Dec 01 17:01:06 crc kubenswrapper[4810]: I1201 17:01:06.467085 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-522ww" Dec 01 17:01:07 crc kubenswrapper[4810]: I1201 17:01:07.346953 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-522ww" Dec 01 17:01:07 crc kubenswrapper[4810]: I1201 17:01:07.405733 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-522ww"] Dec 01 17:01:09 crc kubenswrapper[4810]: I1201 17:01:09.299127 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-522ww" podUID="c78c4f30-a933-4cef-8b91-a5af5a3112b8" containerName="registry-server" containerID="cri-o://006ad8c85f1a37314866492d48b05c3f2bc6b1516c4ced8c8f644b26f692f02c" gracePeriod=2 Dec 01 17:01:10 crc kubenswrapper[4810]: I1201 17:01:10.309625 4810 generic.go:334] "Generic (PLEG): container finished" podID="c78c4f30-a933-4cef-8b91-a5af5a3112b8" containerID="006ad8c85f1a37314866492d48b05c3f2bc6b1516c4ced8c8f644b26f692f02c" exitCode=0 Dec 01 17:01:10 crc kubenswrapper[4810]: I1201 17:01:10.309742 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-522ww" event={"ID":"c78c4f30-a933-4cef-8b91-a5af5a3112b8","Type":"ContainerDied","Data":"006ad8c85f1a37314866492d48b05c3f2bc6b1516c4ced8c8f644b26f692f02c"} Dec 01 17:01:10 crc kubenswrapper[4810]: I1201 17:01:10.309929 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-522ww" event={"ID":"c78c4f30-a933-4cef-8b91-a5af5a3112b8","Type":"ContainerDied","Data":"0767d8e5b92aab0fe6f869f7c1abf9fb09c398cee2c4bccbfe8700aea8d26441"} Dec 01 17:01:10 crc kubenswrapper[4810]: I1201 17:01:10.309945 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0767d8e5b92aab0fe6f869f7c1abf9fb09c398cee2c4bccbfe8700aea8d26441" Dec 01 17:01:10 crc kubenswrapper[4810]: I1201 17:01:10.607591 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-522ww" Dec 01 17:01:10 crc kubenswrapper[4810]: I1201 17:01:10.673456 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9bl6\" (UniqueName: \"kubernetes.io/projected/c78c4f30-a933-4cef-8b91-a5af5a3112b8-kube-api-access-m9bl6\") pod \"c78c4f30-a933-4cef-8b91-a5af5a3112b8\" (UID: \"c78c4f30-a933-4cef-8b91-a5af5a3112b8\") " Dec 01 17:01:10 crc kubenswrapper[4810]: I1201 17:01:10.673565 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c78c4f30-a933-4cef-8b91-a5af5a3112b8-catalog-content\") pod \"c78c4f30-a933-4cef-8b91-a5af5a3112b8\" (UID: \"c78c4f30-a933-4cef-8b91-a5af5a3112b8\") " Dec 01 17:01:10 crc kubenswrapper[4810]: I1201 17:01:10.673630 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c78c4f30-a933-4cef-8b91-a5af5a3112b8-utilities\") pod \"c78c4f30-a933-4cef-8b91-a5af5a3112b8\" (UID: \"c78c4f30-a933-4cef-8b91-a5af5a3112b8\") " Dec 01 17:01:10 crc kubenswrapper[4810]: I1201 17:01:10.675960 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c78c4f30-a933-4cef-8b91-a5af5a3112b8-utilities" (OuterVolumeSpecName: "utilities") pod "c78c4f30-a933-4cef-8b91-a5af5a3112b8" (UID: "c78c4f30-a933-4cef-8b91-a5af5a3112b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:01:10 crc kubenswrapper[4810]: I1201 17:01:10.681291 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c78c4f30-a933-4cef-8b91-a5af5a3112b8-kube-api-access-m9bl6" (OuterVolumeSpecName: "kube-api-access-m9bl6") pod "c78c4f30-a933-4cef-8b91-a5af5a3112b8" (UID: "c78c4f30-a933-4cef-8b91-a5af5a3112b8"). InnerVolumeSpecName "kube-api-access-m9bl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:01:10 crc kubenswrapper[4810]: I1201 17:01:10.695102 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c78c4f30-a933-4cef-8b91-a5af5a3112b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c78c4f30-a933-4cef-8b91-a5af5a3112b8" (UID: "c78c4f30-a933-4cef-8b91-a5af5a3112b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:01:10 crc kubenswrapper[4810]: I1201 17:01:10.775858 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c78c4f30-a933-4cef-8b91-a5af5a3112b8-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:01:10 crc kubenswrapper[4810]: I1201 17:01:10.775899 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9bl6\" (UniqueName: \"kubernetes.io/projected/c78c4f30-a933-4cef-8b91-a5af5a3112b8-kube-api-access-m9bl6\") on node \"crc\" DevicePath \"\"" Dec 01 17:01:10 crc kubenswrapper[4810]: I1201 17:01:10.775911 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c78c4f30-a933-4cef-8b91-a5af5a3112b8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:01:11 crc kubenswrapper[4810]: I1201 17:01:11.318151 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-522ww" Dec 01 17:01:11 crc kubenswrapper[4810]: I1201 17:01:11.351640 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-522ww"] Dec 01 17:01:11 crc kubenswrapper[4810]: I1201 17:01:11.361585 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-522ww"] Dec 01 17:01:12 crc kubenswrapper[4810]: I1201 17:01:12.521847 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c78c4f30-a933-4cef-8b91-a5af5a3112b8" path="/var/lib/kubelet/pods/c78c4f30-a933-4cef-8b91-a5af5a3112b8/volumes" Dec 01 17:01:32 crc kubenswrapper[4810]: I1201 17:01:32.972373 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:01:32 crc kubenswrapper[4810]: I1201 17:01:32.973085 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:02:02 crc kubenswrapper[4810]: I1201 17:02:02.972532 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:02:02 crc kubenswrapper[4810]: I1201 17:02:02.973333 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:02:02 crc kubenswrapper[4810]: I1201 17:02:02.973417 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 17:02:02 crc kubenswrapper[4810]: I1201 17:02:02.974739 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 17:02:02 crc kubenswrapper[4810]: I1201 17:02:02.974834 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" gracePeriod=600 Dec 01 17:02:03 crc kubenswrapper[4810]: E1201 17:02:03.112438 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:02:03 crc kubenswrapper[4810]: I1201 17:02:03.907068 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" exitCode=0 Dec 01 17:02:03 crc kubenswrapper[4810]: I1201 17:02:03.907175 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798"} Dec 01 17:02:03 crc kubenswrapper[4810]: I1201 17:02:03.907410 4810 scope.go:117] "RemoveContainer" containerID="58ed6efdf6a9eee8fc2c0ff2afbf931ed73ba5c194f897a984932d52cef19b8d" Dec 01 17:02:03 crc kubenswrapper[4810]: I1201 17:02:03.908526 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:02:03 crc kubenswrapper[4810]: E1201 17:02:03.909104 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:02:18 crc kubenswrapper[4810]: I1201 17:02:18.491146 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:02:18 crc kubenswrapper[4810]: E1201 17:02:18.493466 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:02:29 crc kubenswrapper[4810]: I1201 17:02:29.492537 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:02:29 crc kubenswrapper[4810]: E1201 17:02:29.493302 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:02:41 crc kubenswrapper[4810]: I1201 17:02:41.490950 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:02:41 crc kubenswrapper[4810]: E1201 17:02:41.491892 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:02:49 crc kubenswrapper[4810]: I1201 17:02:49.425338 4810 generic.go:334] "Generic (PLEG): container finished" podID="d35d5ae0-9a24-4816-80d6-b25a71f73a52" containerID="c2ce945f965947592dcc1f2b828b0439dd8c5b403df8f63a1dee658b773c5744" exitCode=0 Dec 01 17:02:49 crc kubenswrapper[4810]: I1201 17:02:49.425389 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" event={"ID":"d35d5ae0-9a24-4816-80d6-b25a71f73a52","Type":"ContainerDied","Data":"c2ce945f965947592dcc1f2b828b0439dd8c5b403df8f63a1dee658b773c5744"} Dec 01 17:02:50 crc kubenswrapper[4810]: I1201 17:02:50.984324 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.087052 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ceilometer-compute-config-data-0\") pod \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.087391 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ceilometer-compute-config-data-1\") pod \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.087561 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-inventory\") pod \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.087700 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ssh-key\") pod \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.087830 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wfkk\" (UniqueName: \"kubernetes.io/projected/d35d5ae0-9a24-4816-80d6-b25a71f73a52-kube-api-access-5wfkk\") pod \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.087926 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ceilometer-compute-config-data-2\") pod \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.088057 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-telemetry-combined-ca-bundle\") pod \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\" (UID: \"d35d5ae0-9a24-4816-80d6-b25a71f73a52\") " Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.094235 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d35d5ae0-9a24-4816-80d6-b25a71f73a52" (UID: "d35d5ae0-9a24-4816-80d6-b25a71f73a52"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.100745 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d35d5ae0-9a24-4816-80d6-b25a71f73a52-kube-api-access-5wfkk" (OuterVolumeSpecName: "kube-api-access-5wfkk") pod "d35d5ae0-9a24-4816-80d6-b25a71f73a52" (UID: "d35d5ae0-9a24-4816-80d6-b25a71f73a52"). InnerVolumeSpecName "kube-api-access-5wfkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.117321 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d35d5ae0-9a24-4816-80d6-b25a71f73a52" (UID: "d35d5ae0-9a24-4816-80d6-b25a71f73a52"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.118766 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "d35d5ae0-9a24-4816-80d6-b25a71f73a52" (UID: "d35d5ae0-9a24-4816-80d6-b25a71f73a52"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.119279 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "d35d5ae0-9a24-4816-80d6-b25a71f73a52" (UID: "d35d5ae0-9a24-4816-80d6-b25a71f73a52"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.125766 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "d35d5ae0-9a24-4816-80d6-b25a71f73a52" (UID: "d35d5ae0-9a24-4816-80d6-b25a71f73a52"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.132705 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-inventory" (OuterVolumeSpecName: "inventory") pod "d35d5ae0-9a24-4816-80d6-b25a71f73a52" (UID: "d35d5ae0-9a24-4816-80d6-b25a71f73a52"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.190705 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.190777 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.190792 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wfkk\" (UniqueName: \"kubernetes.io/projected/d35d5ae0-9a24-4816-80d6-b25a71f73a52-kube-api-access-5wfkk\") on node \"crc\" DevicePath \"\"" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.190831 4810 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.190845 4810 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.190859 4810 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.190871 4810 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d35d5ae0-9a24-4816-80d6-b25a71f73a52-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.462009 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" event={"ID":"d35d5ae0-9a24-4816-80d6-b25a71f73a52","Type":"ContainerDied","Data":"f0098268a87051d8f6a79407d8ddf6d4a0e1cd8a0bb338543aa452bdc5ff3ee1"} Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.462047 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0098268a87051d8f6a79407d8ddf6d4a0e1cd8a0bb338543aa452bdc5ff3ee1" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.462063 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-kp8l4" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.553966 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-2csms"] Dec 01 17:02:51 crc kubenswrapper[4810]: E1201 17:02:51.556835 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c78c4f30-a933-4cef-8b91-a5af5a3112b8" containerName="extract-utilities" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.556864 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c78c4f30-a933-4cef-8b91-a5af5a3112b8" containerName="extract-utilities" Dec 01 17:02:51 crc kubenswrapper[4810]: E1201 17:02:51.556893 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d35d5ae0-9a24-4816-80d6-b25a71f73a52" containerName="telemetry-openstack-openstack-cell1" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.556900 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d35d5ae0-9a24-4816-80d6-b25a71f73a52" containerName="telemetry-openstack-openstack-cell1" Dec 01 17:02:51 crc kubenswrapper[4810]: E1201 17:02:51.556921 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c78c4f30-a933-4cef-8b91-a5af5a3112b8" containerName="extract-content" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.556927 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c78c4f30-a933-4cef-8b91-a5af5a3112b8" containerName="extract-content" Dec 01 17:02:51 crc kubenswrapper[4810]: E1201 17:02:51.556941 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c78c4f30-a933-4cef-8b91-a5af5a3112b8" containerName="registry-server" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.556946 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c78c4f30-a933-4cef-8b91-a5af5a3112b8" containerName="registry-server" Dec 01 17:02:51 crc kubenswrapper[4810]: E1201 17:02:51.556954 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a37de29-2b57-4f07-b0d2-cf186488012e" containerName="keystone-cron" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.556960 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a37de29-2b57-4f07-b0d2-cf186488012e" containerName="keystone-cron" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.557253 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a37de29-2b57-4f07-b0d2-cf186488012e" containerName="keystone-cron" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.557282 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d35d5ae0-9a24-4816-80d6-b25a71f73a52" containerName="telemetry-openstack-openstack-cell1" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.557293 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c78c4f30-a933-4cef-8b91-a5af5a3112b8" containerName="registry-server" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.558110 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.563784 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.564113 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.564293 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.564741 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.564751 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.577546 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-2csms"] Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.597608 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-2csms\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.597955 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-2csms\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.598289 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnzx8\" (UniqueName: \"kubernetes.io/projected/791f52f5-96d0-4949-813d-c86ddf995d9f-kube-api-access-lnzx8\") pod \"neutron-sriov-openstack-openstack-cell1-2csms\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.598428 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-2csms\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.598584 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-2csms\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.699398 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-2csms\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.699788 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-2csms\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.699985 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnzx8\" (UniqueName: \"kubernetes.io/projected/791f52f5-96d0-4949-813d-c86ddf995d9f-kube-api-access-lnzx8\") pod \"neutron-sriov-openstack-openstack-cell1-2csms\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.700150 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-2csms\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.700257 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-2csms\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.705548 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-2csms\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.705580 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-2csms\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.705900 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-2csms\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.706609 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-2csms\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.722921 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnzx8\" (UniqueName: \"kubernetes.io/projected/791f52f5-96d0-4949-813d-c86ddf995d9f-kube-api-access-lnzx8\") pod \"neutron-sriov-openstack-openstack-cell1-2csms\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:02:51 crc kubenswrapper[4810]: I1201 17:02:51.885853 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:02:52 crc kubenswrapper[4810]: I1201 17:02:52.456513 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-2csms"] Dec 01 17:02:52 crc kubenswrapper[4810]: I1201 17:02:52.482504 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" event={"ID":"791f52f5-96d0-4949-813d-c86ddf995d9f","Type":"ContainerStarted","Data":"622d3e857ed736cde08aedd878fa70c7a8fbca0f3ce2a6ee5043a03eef046c85"} Dec 01 17:02:54 crc kubenswrapper[4810]: I1201 17:02:54.502173 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:02:54 crc kubenswrapper[4810]: E1201 17:02:54.504115 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:02:54 crc kubenswrapper[4810]: I1201 17:02:54.514575 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" event={"ID":"791f52f5-96d0-4949-813d-c86ddf995d9f","Type":"ContainerStarted","Data":"b7b1c0757aeae82bb5adf4b85b26ce7b9d8b20605579acfb6674513f0e587a8c"} Dec 01 17:02:54 crc kubenswrapper[4810]: I1201 17:02:54.549218 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" podStartSLOduration=2.673152651 podStartE2EDuration="3.549198731s" podCreationTimestamp="2025-12-01 17:02:51 +0000 UTC" firstStartedPulling="2025-12-01 17:02:52.463240505 +0000 UTC m=+8938.226750148" lastFinishedPulling="2025-12-01 17:02:53.339286615 +0000 UTC m=+8939.102796228" observedRunningTime="2025-12-01 17:02:54.540815024 +0000 UTC m=+8940.304324647" watchObservedRunningTime="2025-12-01 17:02:54.549198731 +0000 UTC m=+8940.312708334" Dec 01 17:03:08 crc kubenswrapper[4810]: I1201 17:03:08.492227 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:03:08 crc kubenswrapper[4810]: E1201 17:03:08.493064 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:03:23 crc kubenswrapper[4810]: I1201 17:03:23.492761 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:03:23 crc kubenswrapper[4810]: E1201 17:03:23.494147 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:03:34 crc kubenswrapper[4810]: I1201 17:03:34.498560 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:03:34 crc kubenswrapper[4810]: E1201 17:03:34.499760 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:03:46 crc kubenswrapper[4810]: I1201 17:03:46.491343 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:03:46 crc kubenswrapper[4810]: E1201 17:03:46.492346 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:04:00 crc kubenswrapper[4810]: I1201 17:04:00.492145 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:04:00 crc kubenswrapper[4810]: E1201 17:04:00.493821 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:04:13 crc kubenswrapper[4810]: I1201 17:04:13.492063 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:04:13 crc kubenswrapper[4810]: E1201 17:04:13.493150 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:04:28 crc kubenswrapper[4810]: I1201 17:04:28.494105 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:04:28 crc kubenswrapper[4810]: E1201 17:04:28.495071 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:04:40 crc kubenswrapper[4810]: I1201 17:04:40.491323 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:04:40 crc kubenswrapper[4810]: E1201 17:04:40.492212 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:04:51 crc kubenswrapper[4810]: I1201 17:04:51.491637 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:04:51 crc kubenswrapper[4810]: E1201 17:04:51.492429 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:05:02 crc kubenswrapper[4810]: I1201 17:05:02.494068 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:05:02 crc kubenswrapper[4810]: E1201 17:05:02.495204 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:05:16 crc kubenswrapper[4810]: I1201 17:05:16.491309 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:05:16 crc kubenswrapper[4810]: E1201 17:05:16.492351 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:05:30 crc kubenswrapper[4810]: I1201 17:05:30.491089 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:05:30 crc kubenswrapper[4810]: E1201 17:05:30.492811 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:05:45 crc kubenswrapper[4810]: I1201 17:05:45.492694 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:05:45 crc kubenswrapper[4810]: E1201 17:05:45.495743 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:06:00 crc kubenswrapper[4810]: I1201 17:06:00.491981 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:06:00 crc kubenswrapper[4810]: E1201 17:06:00.492740 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:06:11 crc kubenswrapper[4810]: I1201 17:06:11.491398 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:06:11 crc kubenswrapper[4810]: E1201 17:06:11.492413 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:06:18 crc kubenswrapper[4810]: I1201 17:06:18.019719 4810 scope.go:117] "RemoveContainer" containerID="e82f05018561b4d4cf4b1c4bbc593e8a17c5a0f86201661a9c3c4ba66b7924c8" Dec 01 17:06:18 crc kubenswrapper[4810]: I1201 17:06:18.064507 4810 scope.go:117] "RemoveContainer" containerID="570ee0008782d15d49dc83dc418342d56bdb7b004dc355d7749a2121d29757bd" Dec 01 17:06:18 crc kubenswrapper[4810]: I1201 17:06:18.143137 4810 scope.go:117] "RemoveContainer" containerID="d620497c6a8e3ca7677b0dcf14966d1872aa9b89b49aa276f61894d641a554b2" Dec 01 17:06:22 crc kubenswrapper[4810]: I1201 17:06:22.492250 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:06:22 crc kubenswrapper[4810]: E1201 17:06:22.493818 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:06:36 crc kubenswrapper[4810]: I1201 17:06:36.491422 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:06:36 crc kubenswrapper[4810]: E1201 17:06:36.492426 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:06:48 crc kubenswrapper[4810]: I1201 17:06:48.491616 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:06:48 crc kubenswrapper[4810]: E1201 17:06:48.492858 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:07:01 crc kubenswrapper[4810]: I1201 17:07:01.491997 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:07:01 crc kubenswrapper[4810]: E1201 17:07:01.493154 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:07:12 crc kubenswrapper[4810]: I1201 17:07:12.491927 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:07:13 crc kubenswrapper[4810]: I1201 17:07:13.786067 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"74a89642a2db884ad4f530c9c1ea28d67bf41093533dd8bb660e27a80e52f743"} Dec 01 17:07:18 crc kubenswrapper[4810]: I1201 17:07:18.225016 4810 scope.go:117] "RemoveContainer" containerID="f670591f0cfa8049a49dfb0d64706f7010df3137300af6fd2c1685246107a2c5" Dec 01 17:07:18 crc kubenswrapper[4810]: I1201 17:07:18.283719 4810 scope.go:117] "RemoveContainer" containerID="006ad8c85f1a37314866492d48b05c3f2bc6b1516c4ced8c8f644b26f692f02c" Dec 01 17:07:18 crc kubenswrapper[4810]: I1201 17:07:18.344395 4810 scope.go:117] "RemoveContainer" containerID="6bf38fec9cbe3505db30a7c614a9faf2cc973fe338153fbb3ed19112168043c0" Dec 01 17:08:41 crc kubenswrapper[4810]: I1201 17:08:41.527077 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-79646c549d-jxbn6" podUID="e9c91a11-9c09-49e7-b228-b37eddc05cee" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 01 17:09:32 crc kubenswrapper[4810]: I1201 17:09:32.972339 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:09:32 crc kubenswrapper[4810]: I1201 17:09:32.973005 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:10:02 crc kubenswrapper[4810]: I1201 17:10:02.972685 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:10:02 crc kubenswrapper[4810]: I1201 17:10:02.973328 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:10:19 crc kubenswrapper[4810]: I1201 17:10:19.976591 4810 generic.go:334] "Generic (PLEG): container finished" podID="791f52f5-96d0-4949-813d-c86ddf995d9f" containerID="b7b1c0757aeae82bb5adf4b85b26ce7b9d8b20605579acfb6674513f0e587a8c" exitCode=0 Dec 01 17:10:19 crc kubenswrapper[4810]: I1201 17:10:19.976674 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" event={"ID":"791f52f5-96d0-4949-813d-c86ddf995d9f","Type":"ContainerDied","Data":"b7b1c0757aeae82bb5adf4b85b26ce7b9d8b20605579acfb6674513f0e587a8c"} Dec 01 17:10:21 crc kubenswrapper[4810]: I1201 17:10:21.625653 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:10:21 crc kubenswrapper[4810]: I1201 17:10:21.723637 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-neutron-sriov-combined-ca-bundle\") pod \"791f52f5-96d0-4949-813d-c86ddf995d9f\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " Dec 01 17:10:21 crc kubenswrapper[4810]: I1201 17:10:21.723822 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnzx8\" (UniqueName: \"kubernetes.io/projected/791f52f5-96d0-4949-813d-c86ddf995d9f-kube-api-access-lnzx8\") pod \"791f52f5-96d0-4949-813d-c86ddf995d9f\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " Dec 01 17:10:21 crc kubenswrapper[4810]: I1201 17:10:21.723890 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-ssh-key\") pod \"791f52f5-96d0-4949-813d-c86ddf995d9f\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " Dec 01 17:10:21 crc kubenswrapper[4810]: I1201 17:10:21.723913 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-neutron-sriov-agent-neutron-config-0\") pod \"791f52f5-96d0-4949-813d-c86ddf995d9f\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " Dec 01 17:10:21 crc kubenswrapper[4810]: I1201 17:10:21.723959 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-inventory\") pod \"791f52f5-96d0-4949-813d-c86ddf995d9f\" (UID: \"791f52f5-96d0-4949-813d-c86ddf995d9f\") " Dec 01 17:10:21 crc kubenswrapper[4810]: I1201 17:10:21.730499 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "791f52f5-96d0-4949-813d-c86ddf995d9f" (UID: "791f52f5-96d0-4949-813d-c86ddf995d9f"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:10:21 crc kubenswrapper[4810]: I1201 17:10:21.731421 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/791f52f5-96d0-4949-813d-c86ddf995d9f-kube-api-access-lnzx8" (OuterVolumeSpecName: "kube-api-access-lnzx8") pod "791f52f5-96d0-4949-813d-c86ddf995d9f" (UID: "791f52f5-96d0-4949-813d-c86ddf995d9f"). InnerVolumeSpecName "kube-api-access-lnzx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:10:21 crc kubenswrapper[4810]: I1201 17:10:21.755597 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-inventory" (OuterVolumeSpecName: "inventory") pod "791f52f5-96d0-4949-813d-c86ddf995d9f" (UID: "791f52f5-96d0-4949-813d-c86ddf995d9f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:10:21 crc kubenswrapper[4810]: I1201 17:10:21.761681 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "791f52f5-96d0-4949-813d-c86ddf995d9f" (UID: "791f52f5-96d0-4949-813d-c86ddf995d9f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:10:21 crc kubenswrapper[4810]: I1201 17:10:21.767958 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "791f52f5-96d0-4949-813d-c86ddf995d9f" (UID: "791f52f5-96d0-4949-813d-c86ddf995d9f"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:10:21 crc kubenswrapper[4810]: I1201 17:10:21.826507 4810 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:10:21 crc kubenswrapper[4810]: I1201 17:10:21.826540 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnzx8\" (UniqueName: \"kubernetes.io/projected/791f52f5-96d0-4949-813d-c86ddf995d9f-kube-api-access-lnzx8\") on node \"crc\" DevicePath \"\"" Dec 01 17:10:21 crc kubenswrapper[4810]: I1201 17:10:21.826556 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:10:21 crc kubenswrapper[4810]: I1201 17:10:21.826564 4810 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:10:21 crc kubenswrapper[4810]: I1201 17:10:21.826573 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/791f52f5-96d0-4949-813d-c86ddf995d9f-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.009084 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" event={"ID":"791f52f5-96d0-4949-813d-c86ddf995d9f","Type":"ContainerDied","Data":"622d3e857ed736cde08aedd878fa70c7a8fbca0f3ce2a6ee5043a03eef046c85"} Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.009132 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="622d3e857ed736cde08aedd878fa70c7a8fbca0f3ce2a6ee5043a03eef046c85" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.009151 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-2csms" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.104848 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn"] Dec 01 17:10:22 crc kubenswrapper[4810]: E1201 17:10:22.105358 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="791f52f5-96d0-4949-813d-c86ddf995d9f" containerName="neutron-sriov-openstack-openstack-cell1" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.105380 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="791f52f5-96d0-4949-813d-c86ddf995d9f" containerName="neutron-sriov-openstack-openstack-cell1" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.105716 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="791f52f5-96d0-4949-813d-c86ddf995d9f" containerName="neutron-sriov-openstack-openstack-cell1" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.106819 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.109573 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.110573 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.110590 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.111023 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.111079 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.125923 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn"] Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.131146 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-jjwfn\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.131257 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-jjwfn\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.131309 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwkl9\" (UniqueName: \"kubernetes.io/projected/db74fde5-b56b-45ca-bac8-23a2adac45f5-kube-api-access-bwkl9\") pod \"neutron-dhcp-openstack-openstack-cell1-jjwfn\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.131347 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-jjwfn\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.131533 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-jjwfn\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.233175 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-jjwfn\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.233281 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-jjwfn\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.233332 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwkl9\" (UniqueName: \"kubernetes.io/projected/db74fde5-b56b-45ca-bac8-23a2adac45f5-kube-api-access-bwkl9\") pod \"neutron-dhcp-openstack-openstack-cell1-jjwfn\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.233372 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-jjwfn\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.233407 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-jjwfn\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.237095 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-jjwfn\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.244935 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-jjwfn\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.245078 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-jjwfn\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.245307 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-jjwfn\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.248607 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwkl9\" (UniqueName: \"kubernetes.io/projected/db74fde5-b56b-45ca-bac8-23a2adac45f5-kube-api-access-bwkl9\") pod \"neutron-dhcp-openstack-openstack-cell1-jjwfn\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:10:22 crc kubenswrapper[4810]: I1201 17:10:22.422552 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:10:23 crc kubenswrapper[4810]: I1201 17:10:23.062032 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn"] Dec 01 17:10:23 crc kubenswrapper[4810]: I1201 17:10:23.070083 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 17:10:24 crc kubenswrapper[4810]: I1201 17:10:24.048336 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" event={"ID":"db74fde5-b56b-45ca-bac8-23a2adac45f5","Type":"ContainerStarted","Data":"3f42fe60a0f842f184bdcd2bc1396267503e116dd31958122b50a12f64161e45"} Dec 01 17:10:25 crc kubenswrapper[4810]: I1201 17:10:25.062152 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" event={"ID":"db74fde5-b56b-45ca-bac8-23a2adac45f5","Type":"ContainerStarted","Data":"2a75a21849ab4d7c054903e34a092c57fc11a8a4275ec412fe4703632afb3e65"} Dec 01 17:10:25 crc kubenswrapper[4810]: I1201 17:10:25.088706 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" podStartSLOduration=2.370696582 podStartE2EDuration="3.088685969s" podCreationTimestamp="2025-12-01 17:10:22 +0000 UTC" firstStartedPulling="2025-12-01 17:10:23.069777251 +0000 UTC m=+9388.833286864" lastFinishedPulling="2025-12-01 17:10:23.787766638 +0000 UTC m=+9389.551276251" observedRunningTime="2025-12-01 17:10:25.081877425 +0000 UTC m=+9390.845387068" watchObservedRunningTime="2025-12-01 17:10:25.088685969 +0000 UTC m=+9390.852195572" Dec 01 17:10:32 crc kubenswrapper[4810]: I1201 17:10:32.972410 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:10:32 crc kubenswrapper[4810]: I1201 17:10:32.972956 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:10:32 crc kubenswrapper[4810]: I1201 17:10:32.973005 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 17:10:32 crc kubenswrapper[4810]: I1201 17:10:32.973987 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"74a89642a2db884ad4f530c9c1ea28d67bf41093533dd8bb660e27a80e52f743"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 17:10:32 crc kubenswrapper[4810]: I1201 17:10:32.974044 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://74a89642a2db884ad4f530c9c1ea28d67bf41093533dd8bb660e27a80e52f743" gracePeriod=600 Dec 01 17:10:33 crc kubenswrapper[4810]: I1201 17:10:33.150767 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="74a89642a2db884ad4f530c9c1ea28d67bf41093533dd8bb660e27a80e52f743" exitCode=0 Dec 01 17:10:33 crc kubenswrapper[4810]: I1201 17:10:33.150837 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"74a89642a2db884ad4f530c9c1ea28d67bf41093533dd8bb660e27a80e52f743"} Dec 01 17:10:33 crc kubenswrapper[4810]: I1201 17:10:33.150884 4810 scope.go:117] "RemoveContainer" containerID="885debea9aa7d9e445fc510cd3f081a831f8e1fef48561813cacfe07de120798" Dec 01 17:10:34 crc kubenswrapper[4810]: I1201 17:10:34.187724 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb"} Dec 01 17:12:02 crc kubenswrapper[4810]: I1201 17:12:02.457045 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f78gq"] Dec 01 17:12:02 crc kubenswrapper[4810]: I1201 17:12:02.466733 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f78gq" Dec 01 17:12:02 crc kubenswrapper[4810]: I1201 17:12:02.475826 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f78gq"] Dec 01 17:12:02 crc kubenswrapper[4810]: I1201 17:12:02.600025 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czjk4\" (UniqueName: \"kubernetes.io/projected/f5b93b2f-57da-440e-b275-4e0fffb27cf5-kube-api-access-czjk4\") pod \"redhat-marketplace-f78gq\" (UID: \"f5b93b2f-57da-440e-b275-4e0fffb27cf5\") " pod="openshift-marketplace/redhat-marketplace-f78gq" Dec 01 17:12:02 crc kubenswrapper[4810]: I1201 17:12:02.600557 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5b93b2f-57da-440e-b275-4e0fffb27cf5-catalog-content\") pod \"redhat-marketplace-f78gq\" (UID: \"f5b93b2f-57da-440e-b275-4e0fffb27cf5\") " pod="openshift-marketplace/redhat-marketplace-f78gq" Dec 01 17:12:02 crc kubenswrapper[4810]: I1201 17:12:02.600928 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5b93b2f-57da-440e-b275-4e0fffb27cf5-utilities\") pod \"redhat-marketplace-f78gq\" (UID: \"f5b93b2f-57da-440e-b275-4e0fffb27cf5\") " pod="openshift-marketplace/redhat-marketplace-f78gq" Dec 01 17:12:02 crc kubenswrapper[4810]: I1201 17:12:02.703225 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5b93b2f-57da-440e-b275-4e0fffb27cf5-utilities\") pod \"redhat-marketplace-f78gq\" (UID: \"f5b93b2f-57da-440e-b275-4e0fffb27cf5\") " pod="openshift-marketplace/redhat-marketplace-f78gq" Dec 01 17:12:02 crc kubenswrapper[4810]: I1201 17:12:02.703275 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czjk4\" (UniqueName: \"kubernetes.io/projected/f5b93b2f-57da-440e-b275-4e0fffb27cf5-kube-api-access-czjk4\") pod \"redhat-marketplace-f78gq\" (UID: \"f5b93b2f-57da-440e-b275-4e0fffb27cf5\") " pod="openshift-marketplace/redhat-marketplace-f78gq" Dec 01 17:12:02 crc kubenswrapper[4810]: I1201 17:12:02.703356 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5b93b2f-57da-440e-b275-4e0fffb27cf5-catalog-content\") pod \"redhat-marketplace-f78gq\" (UID: \"f5b93b2f-57da-440e-b275-4e0fffb27cf5\") " pod="openshift-marketplace/redhat-marketplace-f78gq" Dec 01 17:12:02 crc kubenswrapper[4810]: I1201 17:12:02.703955 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5b93b2f-57da-440e-b275-4e0fffb27cf5-catalog-content\") pod \"redhat-marketplace-f78gq\" (UID: \"f5b93b2f-57da-440e-b275-4e0fffb27cf5\") " pod="openshift-marketplace/redhat-marketplace-f78gq" Dec 01 17:12:02 crc kubenswrapper[4810]: I1201 17:12:02.703949 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5b93b2f-57da-440e-b275-4e0fffb27cf5-utilities\") pod \"redhat-marketplace-f78gq\" (UID: \"f5b93b2f-57da-440e-b275-4e0fffb27cf5\") " pod="openshift-marketplace/redhat-marketplace-f78gq" Dec 01 17:12:02 crc kubenswrapper[4810]: I1201 17:12:02.722233 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czjk4\" (UniqueName: \"kubernetes.io/projected/f5b93b2f-57da-440e-b275-4e0fffb27cf5-kube-api-access-czjk4\") pod \"redhat-marketplace-f78gq\" (UID: \"f5b93b2f-57da-440e-b275-4e0fffb27cf5\") " pod="openshift-marketplace/redhat-marketplace-f78gq" Dec 01 17:12:02 crc kubenswrapper[4810]: I1201 17:12:02.800819 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f78gq" Dec 01 17:12:03 crc kubenswrapper[4810]: I1201 17:12:03.295869 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f78gq"] Dec 01 17:12:04 crc kubenswrapper[4810]: I1201 17:12:04.169876 4810 generic.go:334] "Generic (PLEG): container finished" podID="f5b93b2f-57da-440e-b275-4e0fffb27cf5" containerID="0489966c7de182ebe9bafe99b40889de9b6ad4c8cceb296f50933ce430c442c8" exitCode=0 Dec 01 17:12:04 crc kubenswrapper[4810]: I1201 17:12:04.169944 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f78gq" event={"ID":"f5b93b2f-57da-440e-b275-4e0fffb27cf5","Type":"ContainerDied","Data":"0489966c7de182ebe9bafe99b40889de9b6ad4c8cceb296f50933ce430c442c8"} Dec 01 17:12:04 crc kubenswrapper[4810]: I1201 17:12:04.170133 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f78gq" event={"ID":"f5b93b2f-57da-440e-b275-4e0fffb27cf5","Type":"ContainerStarted","Data":"b51b72d3559fb8b6140f7610d218ca9e7f1e0ca13f8c52d7898ce57d2fe5a37b"} Dec 01 17:12:10 crc kubenswrapper[4810]: I1201 17:12:10.249810 4810 generic.go:334] "Generic (PLEG): container finished" podID="f5b93b2f-57da-440e-b275-4e0fffb27cf5" containerID="84520c7529fd8b70caabdf18fd28b8f829b35cea3e7b829b4f6afc8464e97cb8" exitCode=0 Dec 01 17:12:10 crc kubenswrapper[4810]: I1201 17:12:10.249864 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f78gq" event={"ID":"f5b93b2f-57da-440e-b275-4e0fffb27cf5","Type":"ContainerDied","Data":"84520c7529fd8b70caabdf18fd28b8f829b35cea3e7b829b4f6afc8464e97cb8"} Dec 01 17:12:11 crc kubenswrapper[4810]: I1201 17:12:11.262873 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f78gq" event={"ID":"f5b93b2f-57da-440e-b275-4e0fffb27cf5","Type":"ContainerStarted","Data":"cbdef598587f2b114deeaac3b6e103048e08f4a9a003311ad6b002e460910b82"} Dec 01 17:12:11 crc kubenswrapper[4810]: I1201 17:12:11.283863 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f78gq" podStartSLOduration=2.77353957 podStartE2EDuration="9.283844428s" podCreationTimestamp="2025-12-01 17:12:02 +0000 UTC" firstStartedPulling="2025-12-01 17:12:04.172700085 +0000 UTC m=+9489.936209678" lastFinishedPulling="2025-12-01 17:12:10.683004943 +0000 UTC m=+9496.446514536" observedRunningTime="2025-12-01 17:12:11.278552156 +0000 UTC m=+9497.042061759" watchObservedRunningTime="2025-12-01 17:12:11.283844428 +0000 UTC m=+9497.047354031" Dec 01 17:12:12 crc kubenswrapper[4810]: I1201 17:12:12.801744 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f78gq" Dec 01 17:12:12 crc kubenswrapper[4810]: I1201 17:12:12.802358 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f78gq" Dec 01 17:12:12 crc kubenswrapper[4810]: I1201 17:12:12.887379 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f78gq" Dec 01 17:12:22 crc kubenswrapper[4810]: I1201 17:12:22.861252 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f78gq" Dec 01 17:12:22 crc kubenswrapper[4810]: I1201 17:12:22.918432 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f78gq"] Dec 01 17:12:23 crc kubenswrapper[4810]: I1201 17:12:23.395282 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f78gq" podUID="f5b93b2f-57da-440e-b275-4e0fffb27cf5" containerName="registry-server" containerID="cri-o://cbdef598587f2b114deeaac3b6e103048e08f4a9a003311ad6b002e460910b82" gracePeriod=2 Dec 01 17:12:23 crc kubenswrapper[4810]: I1201 17:12:23.945672 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f78gq" Dec 01 17:12:24 crc kubenswrapper[4810]: I1201 17:12:24.022072 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5b93b2f-57da-440e-b275-4e0fffb27cf5-utilities\") pod \"f5b93b2f-57da-440e-b275-4e0fffb27cf5\" (UID: \"f5b93b2f-57da-440e-b275-4e0fffb27cf5\") " Dec 01 17:12:24 crc kubenswrapper[4810]: I1201 17:12:24.022123 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czjk4\" (UniqueName: \"kubernetes.io/projected/f5b93b2f-57da-440e-b275-4e0fffb27cf5-kube-api-access-czjk4\") pod \"f5b93b2f-57da-440e-b275-4e0fffb27cf5\" (UID: \"f5b93b2f-57da-440e-b275-4e0fffb27cf5\") " Dec 01 17:12:24 crc kubenswrapper[4810]: I1201 17:12:24.022312 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5b93b2f-57da-440e-b275-4e0fffb27cf5-catalog-content\") pod \"f5b93b2f-57da-440e-b275-4e0fffb27cf5\" (UID: \"f5b93b2f-57da-440e-b275-4e0fffb27cf5\") " Dec 01 17:12:24 crc kubenswrapper[4810]: I1201 17:12:24.023450 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5b93b2f-57da-440e-b275-4e0fffb27cf5-utilities" (OuterVolumeSpecName: "utilities") pod "f5b93b2f-57da-440e-b275-4e0fffb27cf5" (UID: "f5b93b2f-57da-440e-b275-4e0fffb27cf5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:12:24 crc kubenswrapper[4810]: I1201 17:12:24.035190 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5b93b2f-57da-440e-b275-4e0fffb27cf5-kube-api-access-czjk4" (OuterVolumeSpecName: "kube-api-access-czjk4") pod "f5b93b2f-57da-440e-b275-4e0fffb27cf5" (UID: "f5b93b2f-57da-440e-b275-4e0fffb27cf5"). InnerVolumeSpecName "kube-api-access-czjk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:12:24 crc kubenswrapper[4810]: I1201 17:12:24.042120 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5b93b2f-57da-440e-b275-4e0fffb27cf5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5b93b2f-57da-440e-b275-4e0fffb27cf5" (UID: "f5b93b2f-57da-440e-b275-4e0fffb27cf5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:12:24 crc kubenswrapper[4810]: I1201 17:12:24.125084 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5b93b2f-57da-440e-b275-4e0fffb27cf5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:12:24 crc kubenswrapper[4810]: I1201 17:12:24.125127 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5b93b2f-57da-440e-b275-4e0fffb27cf5-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:12:24 crc kubenswrapper[4810]: I1201 17:12:24.125138 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czjk4\" (UniqueName: \"kubernetes.io/projected/f5b93b2f-57da-440e-b275-4e0fffb27cf5-kube-api-access-czjk4\") on node \"crc\" DevicePath \"\"" Dec 01 17:12:24 crc kubenswrapper[4810]: I1201 17:12:24.408337 4810 generic.go:334] "Generic (PLEG): container finished" podID="f5b93b2f-57da-440e-b275-4e0fffb27cf5" containerID="cbdef598587f2b114deeaac3b6e103048e08f4a9a003311ad6b002e460910b82" exitCode=0 Dec 01 17:12:24 crc kubenswrapper[4810]: I1201 17:12:24.408408 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f78gq" Dec 01 17:12:24 crc kubenswrapper[4810]: I1201 17:12:24.408407 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f78gq" event={"ID":"f5b93b2f-57da-440e-b275-4e0fffb27cf5","Type":"ContainerDied","Data":"cbdef598587f2b114deeaac3b6e103048e08f4a9a003311ad6b002e460910b82"} Dec 01 17:12:24 crc kubenswrapper[4810]: I1201 17:12:24.408521 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f78gq" event={"ID":"f5b93b2f-57da-440e-b275-4e0fffb27cf5","Type":"ContainerDied","Data":"b51b72d3559fb8b6140f7610d218ca9e7f1e0ca13f8c52d7898ce57d2fe5a37b"} Dec 01 17:12:24 crc kubenswrapper[4810]: I1201 17:12:24.408573 4810 scope.go:117] "RemoveContainer" containerID="cbdef598587f2b114deeaac3b6e103048e08f4a9a003311ad6b002e460910b82" Dec 01 17:12:24 crc kubenswrapper[4810]: I1201 17:12:24.433507 4810 scope.go:117] "RemoveContainer" containerID="84520c7529fd8b70caabdf18fd28b8f829b35cea3e7b829b4f6afc8464e97cb8" Dec 01 17:12:24 crc kubenswrapper[4810]: I1201 17:12:24.445334 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f78gq"] Dec 01 17:12:24 crc kubenswrapper[4810]: I1201 17:12:24.454847 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f78gq"] Dec 01 17:12:24 crc kubenswrapper[4810]: I1201 17:12:24.502267 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5b93b2f-57da-440e-b275-4e0fffb27cf5" path="/var/lib/kubelet/pods/f5b93b2f-57da-440e-b275-4e0fffb27cf5/volumes" Dec 01 17:12:24 crc kubenswrapper[4810]: E1201 17:12:24.522975 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5b93b2f_57da_440e_b275_4e0fffb27cf5.slice\": RecentStats: unable to find data in memory cache]" Dec 01 17:12:25 crc kubenswrapper[4810]: I1201 17:12:25.065682 4810 scope.go:117] "RemoveContainer" containerID="0489966c7de182ebe9bafe99b40889de9b6ad4c8cceb296f50933ce430c442c8" Dec 01 17:12:25 crc kubenswrapper[4810]: I1201 17:12:25.164735 4810 scope.go:117] "RemoveContainer" containerID="cbdef598587f2b114deeaac3b6e103048e08f4a9a003311ad6b002e460910b82" Dec 01 17:12:25 crc kubenswrapper[4810]: E1201 17:12:25.165589 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbdef598587f2b114deeaac3b6e103048e08f4a9a003311ad6b002e460910b82\": container with ID starting with cbdef598587f2b114deeaac3b6e103048e08f4a9a003311ad6b002e460910b82 not found: ID does not exist" containerID="cbdef598587f2b114deeaac3b6e103048e08f4a9a003311ad6b002e460910b82" Dec 01 17:12:25 crc kubenswrapper[4810]: I1201 17:12:25.165632 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbdef598587f2b114deeaac3b6e103048e08f4a9a003311ad6b002e460910b82"} err="failed to get container status \"cbdef598587f2b114deeaac3b6e103048e08f4a9a003311ad6b002e460910b82\": rpc error: code = NotFound desc = could not find container \"cbdef598587f2b114deeaac3b6e103048e08f4a9a003311ad6b002e460910b82\": container with ID starting with cbdef598587f2b114deeaac3b6e103048e08f4a9a003311ad6b002e460910b82 not found: ID does not exist" Dec 01 17:12:25 crc kubenswrapper[4810]: I1201 17:12:25.165650 4810 scope.go:117] "RemoveContainer" containerID="84520c7529fd8b70caabdf18fd28b8f829b35cea3e7b829b4f6afc8464e97cb8" Dec 01 17:12:25 crc kubenswrapper[4810]: E1201 17:12:25.166096 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84520c7529fd8b70caabdf18fd28b8f829b35cea3e7b829b4f6afc8464e97cb8\": container with ID starting with 84520c7529fd8b70caabdf18fd28b8f829b35cea3e7b829b4f6afc8464e97cb8 not found: ID does not exist" containerID="84520c7529fd8b70caabdf18fd28b8f829b35cea3e7b829b4f6afc8464e97cb8" Dec 01 17:12:25 crc kubenswrapper[4810]: I1201 17:12:25.166131 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84520c7529fd8b70caabdf18fd28b8f829b35cea3e7b829b4f6afc8464e97cb8"} err="failed to get container status \"84520c7529fd8b70caabdf18fd28b8f829b35cea3e7b829b4f6afc8464e97cb8\": rpc error: code = NotFound desc = could not find container \"84520c7529fd8b70caabdf18fd28b8f829b35cea3e7b829b4f6afc8464e97cb8\": container with ID starting with 84520c7529fd8b70caabdf18fd28b8f829b35cea3e7b829b4f6afc8464e97cb8 not found: ID does not exist" Dec 01 17:12:25 crc kubenswrapper[4810]: I1201 17:12:25.166199 4810 scope.go:117] "RemoveContainer" containerID="0489966c7de182ebe9bafe99b40889de9b6ad4c8cceb296f50933ce430c442c8" Dec 01 17:12:25 crc kubenswrapper[4810]: E1201 17:12:25.166562 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0489966c7de182ebe9bafe99b40889de9b6ad4c8cceb296f50933ce430c442c8\": container with ID starting with 0489966c7de182ebe9bafe99b40889de9b6ad4c8cceb296f50933ce430c442c8 not found: ID does not exist" containerID="0489966c7de182ebe9bafe99b40889de9b6ad4c8cceb296f50933ce430c442c8" Dec 01 17:12:25 crc kubenswrapper[4810]: I1201 17:12:25.166593 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0489966c7de182ebe9bafe99b40889de9b6ad4c8cceb296f50933ce430c442c8"} err="failed to get container status \"0489966c7de182ebe9bafe99b40889de9b6ad4c8cceb296f50933ce430c442c8\": rpc error: code = NotFound desc = could not find container \"0489966c7de182ebe9bafe99b40889de9b6ad4c8cceb296f50933ce430c442c8\": container with ID starting with 0489966c7de182ebe9bafe99b40889de9b6ad4c8cceb296f50933ce430c442c8 not found: ID does not exist" Dec 01 17:13:02 crc kubenswrapper[4810]: I1201 17:13:02.972117 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:13:02 crc kubenswrapper[4810]: I1201 17:13:02.975135 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:13:32 crc kubenswrapper[4810]: I1201 17:13:32.972181 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:13:32 crc kubenswrapper[4810]: I1201 17:13:32.972835 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:13:33 crc kubenswrapper[4810]: I1201 17:13:33.447882 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xr754"] Dec 01 17:13:33 crc kubenswrapper[4810]: E1201 17:13:33.448432 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5b93b2f-57da-440e-b275-4e0fffb27cf5" containerName="extract-content" Dec 01 17:13:33 crc kubenswrapper[4810]: I1201 17:13:33.448453 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5b93b2f-57da-440e-b275-4e0fffb27cf5" containerName="extract-content" Dec 01 17:13:33 crc kubenswrapper[4810]: E1201 17:13:33.448562 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5b93b2f-57da-440e-b275-4e0fffb27cf5" containerName="registry-server" Dec 01 17:13:33 crc kubenswrapper[4810]: I1201 17:13:33.448572 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5b93b2f-57da-440e-b275-4e0fffb27cf5" containerName="registry-server" Dec 01 17:13:33 crc kubenswrapper[4810]: E1201 17:13:33.448596 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5b93b2f-57da-440e-b275-4e0fffb27cf5" containerName="extract-utilities" Dec 01 17:13:33 crc kubenswrapper[4810]: I1201 17:13:33.448604 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5b93b2f-57da-440e-b275-4e0fffb27cf5" containerName="extract-utilities" Dec 01 17:13:33 crc kubenswrapper[4810]: I1201 17:13:33.448875 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5b93b2f-57da-440e-b275-4e0fffb27cf5" containerName="registry-server" Dec 01 17:13:33 crc kubenswrapper[4810]: I1201 17:13:33.450902 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xr754" Dec 01 17:13:33 crc kubenswrapper[4810]: I1201 17:13:33.463297 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xr754"] Dec 01 17:13:33 crc kubenswrapper[4810]: I1201 17:13:33.594313 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2skhv\" (UniqueName: \"kubernetes.io/projected/1b5689ce-d989-4c61-bd56-bfb326e4f400-kube-api-access-2skhv\") pod \"redhat-operators-xr754\" (UID: \"1b5689ce-d989-4c61-bd56-bfb326e4f400\") " pod="openshift-marketplace/redhat-operators-xr754" Dec 01 17:13:33 crc kubenswrapper[4810]: I1201 17:13:33.594519 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b5689ce-d989-4c61-bd56-bfb326e4f400-utilities\") pod \"redhat-operators-xr754\" (UID: \"1b5689ce-d989-4c61-bd56-bfb326e4f400\") " pod="openshift-marketplace/redhat-operators-xr754" Dec 01 17:13:33 crc kubenswrapper[4810]: I1201 17:13:33.594629 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b5689ce-d989-4c61-bd56-bfb326e4f400-catalog-content\") pod \"redhat-operators-xr754\" (UID: \"1b5689ce-d989-4c61-bd56-bfb326e4f400\") " pod="openshift-marketplace/redhat-operators-xr754" Dec 01 17:13:33 crc kubenswrapper[4810]: I1201 17:13:33.696350 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b5689ce-d989-4c61-bd56-bfb326e4f400-utilities\") pod \"redhat-operators-xr754\" (UID: \"1b5689ce-d989-4c61-bd56-bfb326e4f400\") " pod="openshift-marketplace/redhat-operators-xr754" Dec 01 17:13:33 crc kubenswrapper[4810]: I1201 17:13:33.696441 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b5689ce-d989-4c61-bd56-bfb326e4f400-catalog-content\") pod \"redhat-operators-xr754\" (UID: \"1b5689ce-d989-4c61-bd56-bfb326e4f400\") " pod="openshift-marketplace/redhat-operators-xr754" Dec 01 17:13:33 crc kubenswrapper[4810]: I1201 17:13:33.696559 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2skhv\" (UniqueName: \"kubernetes.io/projected/1b5689ce-d989-4c61-bd56-bfb326e4f400-kube-api-access-2skhv\") pod \"redhat-operators-xr754\" (UID: \"1b5689ce-d989-4c61-bd56-bfb326e4f400\") " pod="openshift-marketplace/redhat-operators-xr754" Dec 01 17:13:33 crc kubenswrapper[4810]: I1201 17:13:33.696976 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b5689ce-d989-4c61-bd56-bfb326e4f400-catalog-content\") pod \"redhat-operators-xr754\" (UID: \"1b5689ce-d989-4c61-bd56-bfb326e4f400\") " pod="openshift-marketplace/redhat-operators-xr754" Dec 01 17:13:33 crc kubenswrapper[4810]: I1201 17:13:33.697043 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b5689ce-d989-4c61-bd56-bfb326e4f400-utilities\") pod \"redhat-operators-xr754\" (UID: \"1b5689ce-d989-4c61-bd56-bfb326e4f400\") " pod="openshift-marketplace/redhat-operators-xr754" Dec 01 17:13:33 crc kubenswrapper[4810]: I1201 17:13:33.718353 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2skhv\" (UniqueName: \"kubernetes.io/projected/1b5689ce-d989-4c61-bd56-bfb326e4f400-kube-api-access-2skhv\") pod \"redhat-operators-xr754\" (UID: \"1b5689ce-d989-4c61-bd56-bfb326e4f400\") " pod="openshift-marketplace/redhat-operators-xr754" Dec 01 17:13:33 crc kubenswrapper[4810]: I1201 17:13:33.775121 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xr754" Dec 01 17:13:34 crc kubenswrapper[4810]: I1201 17:13:34.293074 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xr754"] Dec 01 17:13:35 crc kubenswrapper[4810]: I1201 17:13:35.250417 4810 generic.go:334] "Generic (PLEG): container finished" podID="1b5689ce-d989-4c61-bd56-bfb326e4f400" containerID="0559901b78b0cd012f501328ad92fe70f29bc9af035496d41dcfc9abfe5021c5" exitCode=0 Dec 01 17:13:35 crc kubenswrapper[4810]: I1201 17:13:35.251091 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr754" event={"ID":"1b5689ce-d989-4c61-bd56-bfb326e4f400","Type":"ContainerDied","Data":"0559901b78b0cd012f501328ad92fe70f29bc9af035496d41dcfc9abfe5021c5"} Dec 01 17:13:35 crc kubenswrapper[4810]: I1201 17:13:35.251128 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr754" event={"ID":"1b5689ce-d989-4c61-bd56-bfb326e4f400","Type":"ContainerStarted","Data":"573cf129a25a3eb9579d426d3f88d17d12e915cf1fe47de0eb00bcb0841da3b3"} Dec 01 17:13:37 crc kubenswrapper[4810]: I1201 17:13:37.273771 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr754" event={"ID":"1b5689ce-d989-4c61-bd56-bfb326e4f400","Type":"ContainerStarted","Data":"779332e398e851f2511da038623f6779b0675ea84c7d3c5f59488ea2799f9f11"} Dec 01 17:13:39 crc kubenswrapper[4810]: I1201 17:13:39.297795 4810 generic.go:334] "Generic (PLEG): container finished" podID="1b5689ce-d989-4c61-bd56-bfb326e4f400" containerID="779332e398e851f2511da038623f6779b0675ea84c7d3c5f59488ea2799f9f11" exitCode=0 Dec 01 17:13:39 crc kubenswrapper[4810]: I1201 17:13:39.297867 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr754" event={"ID":"1b5689ce-d989-4c61-bd56-bfb326e4f400","Type":"ContainerDied","Data":"779332e398e851f2511da038623f6779b0675ea84c7d3c5f59488ea2799f9f11"} Dec 01 17:13:39 crc kubenswrapper[4810]: I1201 17:13:39.618598 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-th5zw"] Dec 01 17:13:39 crc kubenswrapper[4810]: I1201 17:13:39.621178 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-th5zw" Dec 01 17:13:39 crc kubenswrapper[4810]: I1201 17:13:39.646332 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-th5zw"] Dec 01 17:13:39 crc kubenswrapper[4810]: I1201 17:13:39.719975 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e62dc78d-0d09-4b02-a206-30b306bb4807-catalog-content\") pod \"community-operators-th5zw\" (UID: \"e62dc78d-0d09-4b02-a206-30b306bb4807\") " pod="openshift-marketplace/community-operators-th5zw" Dec 01 17:13:39 crc kubenswrapper[4810]: I1201 17:13:39.720403 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw9ln\" (UniqueName: \"kubernetes.io/projected/e62dc78d-0d09-4b02-a206-30b306bb4807-kube-api-access-sw9ln\") pod \"community-operators-th5zw\" (UID: \"e62dc78d-0d09-4b02-a206-30b306bb4807\") " pod="openshift-marketplace/community-operators-th5zw" Dec 01 17:13:39 crc kubenswrapper[4810]: I1201 17:13:39.720452 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e62dc78d-0d09-4b02-a206-30b306bb4807-utilities\") pod \"community-operators-th5zw\" (UID: \"e62dc78d-0d09-4b02-a206-30b306bb4807\") " pod="openshift-marketplace/community-operators-th5zw" Dec 01 17:13:39 crc kubenswrapper[4810]: I1201 17:13:39.823213 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e62dc78d-0d09-4b02-a206-30b306bb4807-catalog-content\") pod \"community-operators-th5zw\" (UID: \"e62dc78d-0d09-4b02-a206-30b306bb4807\") " pod="openshift-marketplace/community-operators-th5zw" Dec 01 17:13:39 crc kubenswrapper[4810]: I1201 17:13:39.823262 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw9ln\" (UniqueName: \"kubernetes.io/projected/e62dc78d-0d09-4b02-a206-30b306bb4807-kube-api-access-sw9ln\") pod \"community-operators-th5zw\" (UID: \"e62dc78d-0d09-4b02-a206-30b306bb4807\") " pod="openshift-marketplace/community-operators-th5zw" Dec 01 17:13:39 crc kubenswrapper[4810]: I1201 17:13:39.823290 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e62dc78d-0d09-4b02-a206-30b306bb4807-utilities\") pod \"community-operators-th5zw\" (UID: \"e62dc78d-0d09-4b02-a206-30b306bb4807\") " pod="openshift-marketplace/community-operators-th5zw" Dec 01 17:13:39 crc kubenswrapper[4810]: I1201 17:13:39.823901 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e62dc78d-0d09-4b02-a206-30b306bb4807-utilities\") pod \"community-operators-th5zw\" (UID: \"e62dc78d-0d09-4b02-a206-30b306bb4807\") " pod="openshift-marketplace/community-operators-th5zw" Dec 01 17:13:39 crc kubenswrapper[4810]: I1201 17:13:39.823937 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e62dc78d-0d09-4b02-a206-30b306bb4807-catalog-content\") pod \"community-operators-th5zw\" (UID: \"e62dc78d-0d09-4b02-a206-30b306bb4807\") " pod="openshift-marketplace/community-operators-th5zw" Dec 01 17:13:39 crc kubenswrapper[4810]: I1201 17:13:39.844379 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw9ln\" (UniqueName: \"kubernetes.io/projected/e62dc78d-0d09-4b02-a206-30b306bb4807-kube-api-access-sw9ln\") pod \"community-operators-th5zw\" (UID: \"e62dc78d-0d09-4b02-a206-30b306bb4807\") " pod="openshift-marketplace/community-operators-th5zw" Dec 01 17:13:39 crc kubenswrapper[4810]: I1201 17:13:39.952385 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-th5zw" Dec 01 17:13:40 crc kubenswrapper[4810]: I1201 17:13:40.311839 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr754" event={"ID":"1b5689ce-d989-4c61-bd56-bfb326e4f400","Type":"ContainerStarted","Data":"a48d5021407e5d6117a6af3b2465d47cabbe4a0a28c30443f23333b522cd8c58"} Dec 01 17:13:40 crc kubenswrapper[4810]: I1201 17:13:40.336853 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xr754" podStartSLOduration=2.886477532 podStartE2EDuration="7.336834518s" podCreationTimestamp="2025-12-01 17:13:33 +0000 UTC" firstStartedPulling="2025-12-01 17:13:35.258035954 +0000 UTC m=+9581.021545557" lastFinishedPulling="2025-12-01 17:13:39.70839294 +0000 UTC m=+9585.471902543" observedRunningTime="2025-12-01 17:13:40.328059312 +0000 UTC m=+9586.091568915" watchObservedRunningTime="2025-12-01 17:13:40.336834518 +0000 UTC m=+9586.100344111" Dec 01 17:13:40 crc kubenswrapper[4810]: I1201 17:13:40.483834 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-th5zw"] Dec 01 17:13:41 crc kubenswrapper[4810]: I1201 17:13:41.323354 4810 generic.go:334] "Generic (PLEG): container finished" podID="e62dc78d-0d09-4b02-a206-30b306bb4807" containerID="62c39a567f2291d9818e88002ef72a32340ec416b05cede67822323150f30e54" exitCode=0 Dec 01 17:13:41 crc kubenswrapper[4810]: I1201 17:13:41.323455 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-th5zw" event={"ID":"e62dc78d-0d09-4b02-a206-30b306bb4807","Type":"ContainerDied","Data":"62c39a567f2291d9818e88002ef72a32340ec416b05cede67822323150f30e54"} Dec 01 17:13:41 crc kubenswrapper[4810]: I1201 17:13:41.323916 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-th5zw" event={"ID":"e62dc78d-0d09-4b02-a206-30b306bb4807","Type":"ContainerStarted","Data":"1155f7f671351a037fd80c82bf000c3f151e849985676a2ea497e06055859b49"} Dec 01 17:13:41 crc kubenswrapper[4810]: I1201 17:13:41.816035 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9s8r9"] Dec 01 17:13:41 crc kubenswrapper[4810]: I1201 17:13:41.818878 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9s8r9" Dec 01 17:13:41 crc kubenswrapper[4810]: I1201 17:13:41.850230 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9s8r9"] Dec 01 17:13:41 crc kubenswrapper[4810]: I1201 17:13:41.967804 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhzf5\" (UniqueName: \"kubernetes.io/projected/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4-kube-api-access-vhzf5\") pod \"certified-operators-9s8r9\" (UID: \"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4\") " pod="openshift-marketplace/certified-operators-9s8r9" Dec 01 17:13:41 crc kubenswrapper[4810]: I1201 17:13:41.967867 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4-catalog-content\") pod \"certified-operators-9s8r9\" (UID: \"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4\") " pod="openshift-marketplace/certified-operators-9s8r9" Dec 01 17:13:41 crc kubenswrapper[4810]: I1201 17:13:41.967893 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4-utilities\") pod \"certified-operators-9s8r9\" (UID: \"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4\") " pod="openshift-marketplace/certified-operators-9s8r9" Dec 01 17:13:42 crc kubenswrapper[4810]: I1201 17:13:42.069729 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhzf5\" (UniqueName: \"kubernetes.io/projected/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4-kube-api-access-vhzf5\") pod \"certified-operators-9s8r9\" (UID: \"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4\") " pod="openshift-marketplace/certified-operators-9s8r9" Dec 01 17:13:42 crc kubenswrapper[4810]: I1201 17:13:42.070058 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4-catalog-content\") pod \"certified-operators-9s8r9\" (UID: \"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4\") " pod="openshift-marketplace/certified-operators-9s8r9" Dec 01 17:13:42 crc kubenswrapper[4810]: I1201 17:13:42.070181 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4-utilities\") pod \"certified-operators-9s8r9\" (UID: \"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4\") " pod="openshift-marketplace/certified-operators-9s8r9" Dec 01 17:13:42 crc kubenswrapper[4810]: I1201 17:13:42.070489 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4-catalog-content\") pod \"certified-operators-9s8r9\" (UID: \"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4\") " pod="openshift-marketplace/certified-operators-9s8r9" Dec 01 17:13:42 crc kubenswrapper[4810]: I1201 17:13:42.070752 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4-utilities\") pod \"certified-operators-9s8r9\" (UID: \"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4\") " pod="openshift-marketplace/certified-operators-9s8r9" Dec 01 17:13:42 crc kubenswrapper[4810]: I1201 17:13:42.090532 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhzf5\" (UniqueName: \"kubernetes.io/projected/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4-kube-api-access-vhzf5\") pod \"certified-operators-9s8r9\" (UID: \"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4\") " pod="openshift-marketplace/certified-operators-9s8r9" Dec 01 17:13:42 crc kubenswrapper[4810]: I1201 17:13:42.143623 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9s8r9" Dec 01 17:13:42 crc kubenswrapper[4810]: I1201 17:13:42.736725 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9s8r9"] Dec 01 17:13:42 crc kubenswrapper[4810]: W1201 17:13:42.749624 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7e5bcc3_ecf1_4481_a8fa_2ba4069289e4.slice/crio-4f6a6d002e523dbc26b2d32deaeac12a335f8601eb0ccf7810ab95772d1f85bd WatchSource:0}: Error finding container 4f6a6d002e523dbc26b2d32deaeac12a335f8601eb0ccf7810ab95772d1f85bd: Status 404 returned error can't find the container with id 4f6a6d002e523dbc26b2d32deaeac12a335f8601eb0ccf7810ab95772d1f85bd Dec 01 17:13:43 crc kubenswrapper[4810]: I1201 17:13:43.344217 4810 generic.go:334] "Generic (PLEG): container finished" podID="c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4" containerID="f4f84f4bfbaf711ff6860d90f8836a7e3401c2fd2e76055fbdad4cb35c38477e" exitCode=0 Dec 01 17:13:43 crc kubenswrapper[4810]: I1201 17:13:43.344510 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9s8r9" event={"ID":"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4","Type":"ContainerDied","Data":"f4f84f4bfbaf711ff6860d90f8836a7e3401c2fd2e76055fbdad4cb35c38477e"} Dec 01 17:13:43 crc kubenswrapper[4810]: I1201 17:13:43.344536 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9s8r9" event={"ID":"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4","Type":"ContainerStarted","Data":"4f6a6d002e523dbc26b2d32deaeac12a335f8601eb0ccf7810ab95772d1f85bd"} Dec 01 17:13:43 crc kubenswrapper[4810]: I1201 17:13:43.776659 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xr754" Dec 01 17:13:43 crc kubenswrapper[4810]: I1201 17:13:43.777019 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xr754" Dec 01 17:13:45 crc kubenswrapper[4810]: I1201 17:13:45.973979 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xr754" Dec 01 17:13:46 crc kubenswrapper[4810]: I1201 17:13:46.037774 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xr754" Dec 01 17:13:47 crc kubenswrapper[4810]: I1201 17:13:47.078129 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xr754"] Dec 01 17:13:47 crc kubenswrapper[4810]: I1201 17:13:47.382515 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xr754" podUID="1b5689ce-d989-4c61-bd56-bfb326e4f400" containerName="registry-server" containerID="cri-o://a48d5021407e5d6117a6af3b2465d47cabbe4a0a28c30443f23333b522cd8c58" gracePeriod=2 Dec 01 17:13:48 crc kubenswrapper[4810]: I1201 17:13:48.402131 4810 generic.go:334] "Generic (PLEG): container finished" podID="1b5689ce-d989-4c61-bd56-bfb326e4f400" containerID="a48d5021407e5d6117a6af3b2465d47cabbe4a0a28c30443f23333b522cd8c58" exitCode=0 Dec 01 17:13:48 crc kubenswrapper[4810]: I1201 17:13:48.402185 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr754" event={"ID":"1b5689ce-d989-4c61-bd56-bfb326e4f400","Type":"ContainerDied","Data":"a48d5021407e5d6117a6af3b2465d47cabbe4a0a28c30443f23333b522cd8c58"} Dec 01 17:13:48 crc kubenswrapper[4810]: I1201 17:13:48.652896 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xr754" Dec 01 17:13:48 crc kubenswrapper[4810]: I1201 17:13:48.815364 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b5689ce-d989-4c61-bd56-bfb326e4f400-catalog-content\") pod \"1b5689ce-d989-4c61-bd56-bfb326e4f400\" (UID: \"1b5689ce-d989-4c61-bd56-bfb326e4f400\") " Dec 01 17:13:48 crc kubenswrapper[4810]: I1201 17:13:48.815515 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2skhv\" (UniqueName: \"kubernetes.io/projected/1b5689ce-d989-4c61-bd56-bfb326e4f400-kube-api-access-2skhv\") pod \"1b5689ce-d989-4c61-bd56-bfb326e4f400\" (UID: \"1b5689ce-d989-4c61-bd56-bfb326e4f400\") " Dec 01 17:13:48 crc kubenswrapper[4810]: I1201 17:13:48.815613 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b5689ce-d989-4c61-bd56-bfb326e4f400-utilities\") pod \"1b5689ce-d989-4c61-bd56-bfb326e4f400\" (UID: \"1b5689ce-d989-4c61-bd56-bfb326e4f400\") " Dec 01 17:13:48 crc kubenswrapper[4810]: I1201 17:13:48.816665 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b5689ce-d989-4c61-bd56-bfb326e4f400-utilities" (OuterVolumeSpecName: "utilities") pod "1b5689ce-d989-4c61-bd56-bfb326e4f400" (UID: "1b5689ce-d989-4c61-bd56-bfb326e4f400"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:13:48 crc kubenswrapper[4810]: I1201 17:13:48.822710 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b5689ce-d989-4c61-bd56-bfb326e4f400-kube-api-access-2skhv" (OuterVolumeSpecName: "kube-api-access-2skhv") pod "1b5689ce-d989-4c61-bd56-bfb326e4f400" (UID: "1b5689ce-d989-4c61-bd56-bfb326e4f400"). InnerVolumeSpecName "kube-api-access-2skhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:13:48 crc kubenswrapper[4810]: I1201 17:13:48.954256 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2skhv\" (UniqueName: \"kubernetes.io/projected/1b5689ce-d989-4c61-bd56-bfb326e4f400-kube-api-access-2skhv\") on node \"crc\" DevicePath \"\"" Dec 01 17:13:48 crc kubenswrapper[4810]: I1201 17:13:48.954311 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b5689ce-d989-4c61-bd56-bfb326e4f400-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:13:49 crc kubenswrapper[4810]: I1201 17:13:49.030781 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b5689ce-d989-4c61-bd56-bfb326e4f400-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b5689ce-d989-4c61-bd56-bfb326e4f400" (UID: "1b5689ce-d989-4c61-bd56-bfb326e4f400"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:13:49 crc kubenswrapper[4810]: I1201 17:13:49.056780 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b5689ce-d989-4c61-bd56-bfb326e4f400-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:13:49 crc kubenswrapper[4810]: I1201 17:13:49.419686 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr754" event={"ID":"1b5689ce-d989-4c61-bd56-bfb326e4f400","Type":"ContainerDied","Data":"573cf129a25a3eb9579d426d3f88d17d12e915cf1fe47de0eb00bcb0841da3b3"} Dec 01 17:13:49 crc kubenswrapper[4810]: I1201 17:13:49.419722 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xr754" Dec 01 17:13:49 crc kubenswrapper[4810]: I1201 17:13:49.419864 4810 scope.go:117] "RemoveContainer" containerID="a48d5021407e5d6117a6af3b2465d47cabbe4a0a28c30443f23333b522cd8c58" Dec 01 17:13:49 crc kubenswrapper[4810]: I1201 17:13:49.422505 4810 generic.go:334] "Generic (PLEG): container finished" podID="c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4" containerID="7bc72316cd4e8379f85514128156260e727ba94cda3abbe90b95abafd81d0012" exitCode=0 Dec 01 17:13:49 crc kubenswrapper[4810]: I1201 17:13:49.422664 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9s8r9" event={"ID":"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4","Type":"ContainerDied","Data":"7bc72316cd4e8379f85514128156260e727ba94cda3abbe90b95abafd81d0012"} Dec 01 17:13:49 crc kubenswrapper[4810]: I1201 17:13:49.427613 4810 generic.go:334] "Generic (PLEG): container finished" podID="e62dc78d-0d09-4b02-a206-30b306bb4807" containerID="15b34452741fd42b62bd15be38f226ca1893e6123ae9a84b9ac26045a1ffd99c" exitCode=0 Dec 01 17:13:49 crc kubenswrapper[4810]: I1201 17:13:49.427653 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-th5zw" event={"ID":"e62dc78d-0d09-4b02-a206-30b306bb4807","Type":"ContainerDied","Data":"15b34452741fd42b62bd15be38f226ca1893e6123ae9a84b9ac26045a1ffd99c"} Dec 01 17:13:49 crc kubenswrapper[4810]: I1201 17:13:49.467096 4810 scope.go:117] "RemoveContainer" containerID="779332e398e851f2511da038623f6779b0675ea84c7d3c5f59488ea2799f9f11" Dec 01 17:13:49 crc kubenswrapper[4810]: I1201 17:13:49.478527 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xr754"] Dec 01 17:13:49 crc kubenswrapper[4810]: I1201 17:13:49.492665 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xr754"] Dec 01 17:13:49 crc kubenswrapper[4810]: I1201 17:13:49.493699 4810 scope.go:117] "RemoveContainer" containerID="0559901b78b0cd012f501328ad92fe70f29bc9af035496d41dcfc9abfe5021c5" Dec 01 17:13:50 crc kubenswrapper[4810]: I1201 17:13:50.512890 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b5689ce-d989-4c61-bd56-bfb326e4f400" path="/var/lib/kubelet/pods/1b5689ce-d989-4c61-bd56-bfb326e4f400/volumes" Dec 01 17:13:51 crc kubenswrapper[4810]: I1201 17:13:51.451380 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9s8r9" event={"ID":"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4","Type":"ContainerStarted","Data":"6905810317c9e447908ec1a11d9c857a1073b0c334937f279a2fcd9c8bdc9fab"} Dec 01 17:13:51 crc kubenswrapper[4810]: I1201 17:13:51.454393 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-th5zw" event={"ID":"e62dc78d-0d09-4b02-a206-30b306bb4807","Type":"ContainerStarted","Data":"7d025189743cf0f62168d909a47489866aff3d2f9cf217a3a9d12d062c2f0650"} Dec 01 17:13:51 crc kubenswrapper[4810]: I1201 17:13:51.481209 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9s8r9" podStartSLOduration=2.776340835 podStartE2EDuration="10.481186972s" podCreationTimestamp="2025-12-01 17:13:41 +0000 UTC" firstStartedPulling="2025-12-01 17:13:43.345889795 +0000 UTC m=+9589.109399398" lastFinishedPulling="2025-12-01 17:13:51.050735932 +0000 UTC m=+9596.814245535" observedRunningTime="2025-12-01 17:13:51.472315754 +0000 UTC m=+9597.235825367" watchObservedRunningTime="2025-12-01 17:13:51.481186972 +0000 UTC m=+9597.244696585" Dec 01 17:13:51 crc kubenswrapper[4810]: I1201 17:13:51.508984 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-th5zw" podStartSLOduration=2.851618093 podStartE2EDuration="12.50896333s" podCreationTimestamp="2025-12-01 17:13:39 +0000 UTC" firstStartedPulling="2025-12-01 17:13:41.325404405 +0000 UTC m=+9587.088914008" lastFinishedPulling="2025-12-01 17:13:50.982749642 +0000 UTC m=+9596.746259245" observedRunningTime="2025-12-01 17:13:51.502715392 +0000 UTC m=+9597.266225005" watchObservedRunningTime="2025-12-01 17:13:51.50896333 +0000 UTC m=+9597.272472943" Dec 01 17:13:52 crc kubenswrapper[4810]: I1201 17:13:52.144603 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9s8r9" Dec 01 17:13:52 crc kubenswrapper[4810]: I1201 17:13:52.144690 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9s8r9" Dec 01 17:13:53 crc kubenswrapper[4810]: I1201 17:13:53.212017 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-9s8r9" podUID="c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4" containerName="registry-server" probeResult="failure" output=< Dec 01 17:13:53 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Dec 01 17:13:53 crc kubenswrapper[4810]: > Dec 01 17:13:57 crc kubenswrapper[4810]: I1201 17:13:57.520747 4810 generic.go:334] "Generic (PLEG): container finished" podID="db74fde5-b56b-45ca-bac8-23a2adac45f5" containerID="2a75a21849ab4d7c054903e34a092c57fc11a8a4275ec412fe4703632afb3e65" exitCode=0 Dec 01 17:13:57 crc kubenswrapper[4810]: I1201 17:13:57.520867 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" event={"ID":"db74fde5-b56b-45ca-bac8-23a2adac45f5","Type":"ContainerDied","Data":"2a75a21849ab4d7c054903e34a092c57fc11a8a4275ec412fe4703632afb3e65"} Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.045423 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.130659 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-neutron-dhcp-agent-neutron-config-0\") pod \"db74fde5-b56b-45ca-bac8-23a2adac45f5\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.130785 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-neutron-dhcp-combined-ca-bundle\") pod \"db74fde5-b56b-45ca-bac8-23a2adac45f5\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.130829 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwkl9\" (UniqueName: \"kubernetes.io/projected/db74fde5-b56b-45ca-bac8-23a2adac45f5-kube-api-access-bwkl9\") pod \"db74fde5-b56b-45ca-bac8-23a2adac45f5\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.130906 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-inventory\") pod \"db74fde5-b56b-45ca-bac8-23a2adac45f5\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.131059 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-ssh-key\") pod \"db74fde5-b56b-45ca-bac8-23a2adac45f5\" (UID: \"db74fde5-b56b-45ca-bac8-23a2adac45f5\") " Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.156424 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db74fde5-b56b-45ca-bac8-23a2adac45f5-kube-api-access-bwkl9" (OuterVolumeSpecName: "kube-api-access-bwkl9") pod "db74fde5-b56b-45ca-bac8-23a2adac45f5" (UID: "db74fde5-b56b-45ca-bac8-23a2adac45f5"). InnerVolumeSpecName "kube-api-access-bwkl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.176665 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "db74fde5-b56b-45ca-bac8-23a2adac45f5" (UID: "db74fde5-b56b-45ca-bac8-23a2adac45f5"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.243090 4810 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.243143 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwkl9\" (UniqueName: \"kubernetes.io/projected/db74fde5-b56b-45ca-bac8-23a2adac45f5-kube-api-access-bwkl9\") on node \"crc\" DevicePath \"\"" Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.274662 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-inventory" (OuterVolumeSpecName: "inventory") pod "db74fde5-b56b-45ca-bac8-23a2adac45f5" (UID: "db74fde5-b56b-45ca-bac8-23a2adac45f5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.280149 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "db74fde5-b56b-45ca-bac8-23a2adac45f5" (UID: "db74fde5-b56b-45ca-bac8-23a2adac45f5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.353270 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.353606 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.354396 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "db74fde5-b56b-45ca-bac8-23a2adac45f5" (UID: "db74fde5-b56b-45ca-bac8-23a2adac45f5"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.457250 4810 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/db74fde5-b56b-45ca-bac8-23a2adac45f5-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.544686 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" event={"ID":"db74fde5-b56b-45ca-bac8-23a2adac45f5","Type":"ContainerDied","Data":"3f42fe60a0f842f184bdcd2bc1396267503e116dd31958122b50a12f64161e45"} Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.544738 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f42fe60a0f842f184bdcd2bc1396267503e116dd31958122b50a12f64161e45" Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.544759 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jjwfn" Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.952901 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-th5zw" Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.953060 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-th5zw" Dec 01 17:13:59 crc kubenswrapper[4810]: I1201 17:13:59.999263 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-th5zw" Dec 01 17:14:00 crc kubenswrapper[4810]: I1201 17:14:00.616468 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-th5zw" Dec 01 17:14:00 crc kubenswrapper[4810]: I1201 17:14:00.695252 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-th5zw"] Dec 01 17:14:00 crc kubenswrapper[4810]: I1201 17:14:00.737953 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tk26t"] Dec 01 17:14:00 crc kubenswrapper[4810]: I1201 17:14:00.738187 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tk26t" podUID="f81e6c36-a33b-4fe5-b8b5-8dac22d2b294" containerName="registry-server" containerID="cri-o://74dbdf8d08b57b7688e6482d0e8774d960ad696ee84fc73c595f66a8de30e45a" gracePeriod=2 Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.440413 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tk26t" Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.534111 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zm94z\" (UniqueName: \"kubernetes.io/projected/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294-kube-api-access-zm94z\") pod \"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294\" (UID: \"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294\") " Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.534216 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294-utilities\") pod \"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294\" (UID: \"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294\") " Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.534328 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294-catalog-content\") pod \"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294\" (UID: \"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294\") " Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.535327 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294-utilities" (OuterVolumeSpecName: "utilities") pod "f81e6c36-a33b-4fe5-b8b5-8dac22d2b294" (UID: "f81e6c36-a33b-4fe5-b8b5-8dac22d2b294"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.539903 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294-kube-api-access-zm94z" (OuterVolumeSpecName: "kube-api-access-zm94z") pod "f81e6c36-a33b-4fe5-b8b5-8dac22d2b294" (UID: "f81e6c36-a33b-4fe5-b8b5-8dac22d2b294"). InnerVolumeSpecName "kube-api-access-zm94z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.567923 4810 generic.go:334] "Generic (PLEG): container finished" podID="f81e6c36-a33b-4fe5-b8b5-8dac22d2b294" containerID="74dbdf8d08b57b7688e6482d0e8774d960ad696ee84fc73c595f66a8de30e45a" exitCode=0 Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.569083 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tk26t" Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.569095 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tk26t" event={"ID":"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294","Type":"ContainerDied","Data":"74dbdf8d08b57b7688e6482d0e8774d960ad696ee84fc73c595f66a8de30e45a"} Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.569281 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tk26t" event={"ID":"f81e6c36-a33b-4fe5-b8b5-8dac22d2b294","Type":"ContainerDied","Data":"fc8c8e65ed9702afefd6dc4f4a1c2b17878727012a04dca2314c72d9c522893f"} Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.569299 4810 scope.go:117] "RemoveContainer" containerID="74dbdf8d08b57b7688e6482d0e8774d960ad696ee84fc73c595f66a8de30e45a" Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.584812 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f81e6c36-a33b-4fe5-b8b5-8dac22d2b294" (UID: "f81e6c36-a33b-4fe5-b8b5-8dac22d2b294"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.599596 4810 scope.go:117] "RemoveContainer" containerID="0e27029a7946e008d5d71e53dafcf6dba211589fcfefa68211bfa645dd869121" Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.623694 4810 scope.go:117] "RemoveContainer" containerID="8e1e5a8499f3aa9f99b42a7ba70317e61e3a44304043401205662bce97cd81ed" Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.636489 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zm94z\" (UniqueName: \"kubernetes.io/projected/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294-kube-api-access-zm94z\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.636513 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.636524 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.670419 4810 scope.go:117] "RemoveContainer" containerID="74dbdf8d08b57b7688e6482d0e8774d960ad696ee84fc73c595f66a8de30e45a" Dec 01 17:14:01 crc kubenswrapper[4810]: E1201 17:14:01.670946 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74dbdf8d08b57b7688e6482d0e8774d960ad696ee84fc73c595f66a8de30e45a\": container with ID starting with 74dbdf8d08b57b7688e6482d0e8774d960ad696ee84fc73c595f66a8de30e45a not found: ID does not exist" containerID="74dbdf8d08b57b7688e6482d0e8774d960ad696ee84fc73c595f66a8de30e45a" Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.670992 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74dbdf8d08b57b7688e6482d0e8774d960ad696ee84fc73c595f66a8de30e45a"} err="failed to get container status \"74dbdf8d08b57b7688e6482d0e8774d960ad696ee84fc73c595f66a8de30e45a\": rpc error: code = NotFound desc = could not find container \"74dbdf8d08b57b7688e6482d0e8774d960ad696ee84fc73c595f66a8de30e45a\": container with ID starting with 74dbdf8d08b57b7688e6482d0e8774d960ad696ee84fc73c595f66a8de30e45a not found: ID does not exist" Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.671024 4810 scope.go:117] "RemoveContainer" containerID="0e27029a7946e008d5d71e53dafcf6dba211589fcfefa68211bfa645dd869121" Dec 01 17:14:01 crc kubenswrapper[4810]: E1201 17:14:01.671357 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e27029a7946e008d5d71e53dafcf6dba211589fcfefa68211bfa645dd869121\": container with ID starting with 0e27029a7946e008d5d71e53dafcf6dba211589fcfefa68211bfa645dd869121 not found: ID does not exist" containerID="0e27029a7946e008d5d71e53dafcf6dba211589fcfefa68211bfa645dd869121" Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.671407 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e27029a7946e008d5d71e53dafcf6dba211589fcfefa68211bfa645dd869121"} err="failed to get container status \"0e27029a7946e008d5d71e53dafcf6dba211589fcfefa68211bfa645dd869121\": rpc error: code = NotFound desc = could not find container \"0e27029a7946e008d5d71e53dafcf6dba211589fcfefa68211bfa645dd869121\": container with ID starting with 0e27029a7946e008d5d71e53dafcf6dba211589fcfefa68211bfa645dd869121 not found: ID does not exist" Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.671447 4810 scope.go:117] "RemoveContainer" containerID="8e1e5a8499f3aa9f99b42a7ba70317e61e3a44304043401205662bce97cd81ed" Dec 01 17:14:01 crc kubenswrapper[4810]: E1201 17:14:01.671776 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e1e5a8499f3aa9f99b42a7ba70317e61e3a44304043401205662bce97cd81ed\": container with ID starting with 8e1e5a8499f3aa9f99b42a7ba70317e61e3a44304043401205662bce97cd81ed not found: ID does not exist" containerID="8e1e5a8499f3aa9f99b42a7ba70317e61e3a44304043401205662bce97cd81ed" Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.671814 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e1e5a8499f3aa9f99b42a7ba70317e61e3a44304043401205662bce97cd81ed"} err="failed to get container status \"8e1e5a8499f3aa9f99b42a7ba70317e61e3a44304043401205662bce97cd81ed\": rpc error: code = NotFound desc = could not find container \"8e1e5a8499f3aa9f99b42a7ba70317e61e3a44304043401205662bce97cd81ed\": container with ID starting with 8e1e5a8499f3aa9f99b42a7ba70317e61e3a44304043401205662bce97cd81ed not found: ID does not exist" Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.903807 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tk26t"] Dec 01 17:14:01 crc kubenswrapper[4810]: I1201 17:14:01.911909 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tk26t"] Dec 01 17:14:02 crc kubenswrapper[4810]: I1201 17:14:02.226843 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9s8r9" Dec 01 17:14:02 crc kubenswrapper[4810]: I1201 17:14:02.275033 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9s8r9" Dec 01 17:14:02 crc kubenswrapper[4810]: I1201 17:14:02.514257 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f81e6c36-a33b-4fe5-b8b5-8dac22d2b294" path="/var/lib/kubelet/pods/f81e6c36-a33b-4fe5-b8b5-8dac22d2b294/volumes" Dec 01 17:14:02 crc kubenswrapper[4810]: I1201 17:14:02.972460 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:14:02 crc kubenswrapper[4810]: I1201 17:14:02.973063 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:14:02 crc kubenswrapper[4810]: I1201 17:14:02.973127 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 17:14:02 crc kubenswrapper[4810]: I1201 17:14:02.974185 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 17:14:02 crc kubenswrapper[4810]: I1201 17:14:02.974323 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" gracePeriod=600 Dec 01 17:14:03 crc kubenswrapper[4810]: E1201 17:14:03.127946 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:14:03 crc kubenswrapper[4810]: I1201 17:14:03.603814 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" exitCode=0 Dec 01 17:14:03 crc kubenswrapper[4810]: I1201 17:14:03.603899 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb"} Dec 01 17:14:03 crc kubenswrapper[4810]: I1201 17:14:03.604236 4810 scope.go:117] "RemoveContainer" containerID="74a89642a2db884ad4f530c9c1ea28d67bf41093533dd8bb660e27a80e52f743" Dec 01 17:14:03 crc kubenswrapper[4810]: I1201 17:14:03.605169 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:14:03 crc kubenswrapper[4810]: E1201 17:14:03.605669 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:14:06 crc kubenswrapper[4810]: I1201 17:14:06.661206 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9s8r9"] Dec 01 17:14:06 crc kubenswrapper[4810]: I1201 17:14:06.662277 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9s8r9" podUID="c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4" containerName="registry-server" containerID="cri-o://6905810317c9e447908ec1a11d9c857a1073b0c334937f279a2fcd9c8bdc9fab" gracePeriod=2 Dec 01 17:14:07 crc kubenswrapper[4810]: I1201 17:14:07.144726 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9s8r9" Dec 01 17:14:07 crc kubenswrapper[4810]: I1201 17:14:07.259694 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4-catalog-content\") pod \"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4\" (UID: \"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4\") " Dec 01 17:14:07 crc kubenswrapper[4810]: I1201 17:14:07.259798 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhzf5\" (UniqueName: \"kubernetes.io/projected/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4-kube-api-access-vhzf5\") pod \"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4\" (UID: \"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4\") " Dec 01 17:14:07 crc kubenswrapper[4810]: I1201 17:14:07.260029 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4-utilities\") pod \"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4\" (UID: \"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4\") " Dec 01 17:14:07 crc kubenswrapper[4810]: I1201 17:14:07.260861 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4-utilities" (OuterVolumeSpecName: "utilities") pod "c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4" (UID: "c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:14:07 crc kubenswrapper[4810]: I1201 17:14:07.267378 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4-kube-api-access-vhzf5" (OuterVolumeSpecName: "kube-api-access-vhzf5") pod "c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4" (UID: "c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4"). InnerVolumeSpecName "kube-api-access-vhzf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:14:07 crc kubenswrapper[4810]: I1201 17:14:07.330015 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4" (UID: "c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:14:07 crc kubenswrapper[4810]: I1201 17:14:07.364560 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:07 crc kubenswrapper[4810]: I1201 17:14:07.364600 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:07 crc kubenswrapper[4810]: I1201 17:14:07.364613 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhzf5\" (UniqueName: \"kubernetes.io/projected/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4-kube-api-access-vhzf5\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:07 crc kubenswrapper[4810]: I1201 17:14:07.646658 4810 generic.go:334] "Generic (PLEG): container finished" podID="c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4" containerID="6905810317c9e447908ec1a11d9c857a1073b0c334937f279a2fcd9c8bdc9fab" exitCode=0 Dec 01 17:14:07 crc kubenswrapper[4810]: I1201 17:14:07.646716 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9s8r9" event={"ID":"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4","Type":"ContainerDied","Data":"6905810317c9e447908ec1a11d9c857a1073b0c334937f279a2fcd9c8bdc9fab"} Dec 01 17:14:07 crc kubenswrapper[4810]: I1201 17:14:07.647011 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9s8r9" event={"ID":"c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4","Type":"ContainerDied","Data":"4f6a6d002e523dbc26b2d32deaeac12a335f8601eb0ccf7810ab95772d1f85bd"} Dec 01 17:14:07 crc kubenswrapper[4810]: I1201 17:14:07.647032 4810 scope.go:117] "RemoveContainer" containerID="6905810317c9e447908ec1a11d9c857a1073b0c334937f279a2fcd9c8bdc9fab" Dec 01 17:14:07 crc kubenswrapper[4810]: I1201 17:14:07.646759 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9s8r9" Dec 01 17:14:07 crc kubenswrapper[4810]: I1201 17:14:07.676450 4810 scope.go:117] "RemoveContainer" containerID="7bc72316cd4e8379f85514128156260e727ba94cda3abbe90b95abafd81d0012" Dec 01 17:14:07 crc kubenswrapper[4810]: I1201 17:14:07.686329 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9s8r9"] Dec 01 17:14:07 crc kubenswrapper[4810]: I1201 17:14:07.697549 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9s8r9"] Dec 01 17:14:08 crc kubenswrapper[4810]: I1201 17:14:08.164069 4810 scope.go:117] "RemoveContainer" containerID="f4f84f4bfbaf711ff6860d90f8836a7e3401c2fd2e76055fbdad4cb35c38477e" Dec 01 17:14:08 crc kubenswrapper[4810]: I1201 17:14:08.212127 4810 scope.go:117] "RemoveContainer" containerID="6905810317c9e447908ec1a11d9c857a1073b0c334937f279a2fcd9c8bdc9fab" Dec 01 17:14:08 crc kubenswrapper[4810]: E1201 17:14:08.212668 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6905810317c9e447908ec1a11d9c857a1073b0c334937f279a2fcd9c8bdc9fab\": container with ID starting with 6905810317c9e447908ec1a11d9c857a1073b0c334937f279a2fcd9c8bdc9fab not found: ID does not exist" containerID="6905810317c9e447908ec1a11d9c857a1073b0c334937f279a2fcd9c8bdc9fab" Dec 01 17:14:08 crc kubenswrapper[4810]: I1201 17:14:08.212708 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6905810317c9e447908ec1a11d9c857a1073b0c334937f279a2fcd9c8bdc9fab"} err="failed to get container status \"6905810317c9e447908ec1a11d9c857a1073b0c334937f279a2fcd9c8bdc9fab\": rpc error: code = NotFound desc = could not find container \"6905810317c9e447908ec1a11d9c857a1073b0c334937f279a2fcd9c8bdc9fab\": container with ID starting with 6905810317c9e447908ec1a11d9c857a1073b0c334937f279a2fcd9c8bdc9fab not found: ID does not exist" Dec 01 17:14:08 crc kubenswrapper[4810]: I1201 17:14:08.212734 4810 scope.go:117] "RemoveContainer" containerID="7bc72316cd4e8379f85514128156260e727ba94cda3abbe90b95abafd81d0012" Dec 01 17:14:08 crc kubenswrapper[4810]: E1201 17:14:08.213667 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bc72316cd4e8379f85514128156260e727ba94cda3abbe90b95abafd81d0012\": container with ID starting with 7bc72316cd4e8379f85514128156260e727ba94cda3abbe90b95abafd81d0012 not found: ID does not exist" containerID="7bc72316cd4e8379f85514128156260e727ba94cda3abbe90b95abafd81d0012" Dec 01 17:14:08 crc kubenswrapper[4810]: I1201 17:14:08.213696 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bc72316cd4e8379f85514128156260e727ba94cda3abbe90b95abafd81d0012"} err="failed to get container status \"7bc72316cd4e8379f85514128156260e727ba94cda3abbe90b95abafd81d0012\": rpc error: code = NotFound desc = could not find container \"7bc72316cd4e8379f85514128156260e727ba94cda3abbe90b95abafd81d0012\": container with ID starting with 7bc72316cd4e8379f85514128156260e727ba94cda3abbe90b95abafd81d0012 not found: ID does not exist" Dec 01 17:14:08 crc kubenswrapper[4810]: I1201 17:14:08.213713 4810 scope.go:117] "RemoveContainer" containerID="f4f84f4bfbaf711ff6860d90f8836a7e3401c2fd2e76055fbdad4cb35c38477e" Dec 01 17:14:08 crc kubenswrapper[4810]: E1201 17:14:08.213994 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4f84f4bfbaf711ff6860d90f8836a7e3401c2fd2e76055fbdad4cb35c38477e\": container with ID starting with f4f84f4bfbaf711ff6860d90f8836a7e3401c2fd2e76055fbdad4cb35c38477e not found: ID does not exist" containerID="f4f84f4bfbaf711ff6860d90f8836a7e3401c2fd2e76055fbdad4cb35c38477e" Dec 01 17:14:08 crc kubenswrapper[4810]: I1201 17:14:08.214038 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4f84f4bfbaf711ff6860d90f8836a7e3401c2fd2e76055fbdad4cb35c38477e"} err="failed to get container status \"f4f84f4bfbaf711ff6860d90f8836a7e3401c2fd2e76055fbdad4cb35c38477e\": rpc error: code = NotFound desc = could not find container \"f4f84f4bfbaf711ff6860d90f8836a7e3401c2fd2e76055fbdad4cb35c38477e\": container with ID starting with f4f84f4bfbaf711ff6860d90f8836a7e3401c2fd2e76055fbdad4cb35c38477e not found: ID does not exist" Dec 01 17:14:08 crc kubenswrapper[4810]: I1201 17:14:08.507800 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4" path="/var/lib/kubelet/pods/c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4/volumes" Dec 01 17:14:14 crc kubenswrapper[4810]: I1201 17:14:14.498635 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:14:14 crc kubenswrapper[4810]: E1201 17:14:14.499542 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:14:15 crc kubenswrapper[4810]: I1201 17:14:15.149943 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 17:14:15 crc kubenswrapper[4810]: I1201 17:14:15.150633 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="578dc44d-25f5-4ef1-bbe7-eb2328954767" containerName="nova-cell0-conductor-conductor" containerID="cri-o://2f4d53685fe68630eb2427db02a9e6c2864c578f0eee81039835517195889ab1" gracePeriod=30 Dec 01 17:14:15 crc kubenswrapper[4810]: I1201 17:14:15.216644 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 17:14:15 crc kubenswrapper[4810]: I1201 17:14:15.216915 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="ca965e9d-7a9d-4351-bfa1-36d77a9af2a3" containerName="nova-cell1-conductor-conductor" containerID="cri-o://13f99bc5259f34633b3dc899ea67dcc6b4654616360f83abebaf1504118d2bb4" gracePeriod=30 Dec 01 17:14:15 crc kubenswrapper[4810]: E1201 17:14:15.707636 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="13f99bc5259f34633b3dc899ea67dcc6b4654616360f83abebaf1504118d2bb4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 01 17:14:15 crc kubenswrapper[4810]: E1201 17:14:15.710392 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="13f99bc5259f34633b3dc899ea67dcc6b4654616360f83abebaf1504118d2bb4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 01 17:14:15 crc kubenswrapper[4810]: E1201 17:14:15.711938 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="13f99bc5259f34633b3dc899ea67dcc6b4654616360f83abebaf1504118d2bb4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 01 17:14:15 crc kubenswrapper[4810]: E1201 17:14:15.712080 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="ca965e9d-7a9d-4351-bfa1-36d77a9af2a3" containerName="nova-cell1-conductor-conductor" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.013052 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n"] Dec 01 17:14:16 crc kubenswrapper[4810]: E1201 17:14:16.013600 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4" containerName="extract-utilities" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.013617 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4" containerName="extract-utilities" Dec 01 17:14:16 crc kubenswrapper[4810]: E1201 17:14:16.013631 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b5689ce-d989-4c61-bd56-bfb326e4f400" containerName="extract-utilities" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.013638 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b5689ce-d989-4c61-bd56-bfb326e4f400" containerName="extract-utilities" Dec 01 17:14:16 crc kubenswrapper[4810]: E1201 17:14:16.013652 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4" containerName="extract-content" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.013659 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4" containerName="extract-content" Dec 01 17:14:16 crc kubenswrapper[4810]: E1201 17:14:16.013670 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f81e6c36-a33b-4fe5-b8b5-8dac22d2b294" containerName="extract-utilities" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.013676 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f81e6c36-a33b-4fe5-b8b5-8dac22d2b294" containerName="extract-utilities" Dec 01 17:14:16 crc kubenswrapper[4810]: E1201 17:14:16.013688 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f81e6c36-a33b-4fe5-b8b5-8dac22d2b294" containerName="extract-content" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.013695 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f81e6c36-a33b-4fe5-b8b5-8dac22d2b294" containerName="extract-content" Dec 01 17:14:16 crc kubenswrapper[4810]: E1201 17:14:16.013714 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db74fde5-b56b-45ca-bac8-23a2adac45f5" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.013721 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="db74fde5-b56b-45ca-bac8-23a2adac45f5" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 01 17:14:16 crc kubenswrapper[4810]: E1201 17:14:16.013739 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4" containerName="registry-server" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.013745 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4" containerName="registry-server" Dec 01 17:14:16 crc kubenswrapper[4810]: E1201 17:14:16.013756 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b5689ce-d989-4c61-bd56-bfb326e4f400" containerName="registry-server" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.013761 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b5689ce-d989-4c61-bd56-bfb326e4f400" containerName="registry-server" Dec 01 17:14:16 crc kubenswrapper[4810]: E1201 17:14:16.013779 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b5689ce-d989-4c61-bd56-bfb326e4f400" containerName="extract-content" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.013784 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b5689ce-d989-4c61-bd56-bfb326e4f400" containerName="extract-content" Dec 01 17:14:16 crc kubenswrapper[4810]: E1201 17:14:16.013800 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f81e6c36-a33b-4fe5-b8b5-8dac22d2b294" containerName="registry-server" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.013819 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f81e6c36-a33b-4fe5-b8b5-8dac22d2b294" containerName="registry-server" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.014023 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="db74fde5-b56b-45ca-bac8-23a2adac45f5" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.014038 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b5689ce-d989-4c61-bd56-bfb326e4f400" containerName="registry-server" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.014058 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7e5bcc3-ecf1-4481-a8fa-2ba4069289e4" containerName="registry-server" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.014068 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f81e6c36-a33b-4fe5-b8b5-8dac22d2b294" containerName="registry-server" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.014803 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.017905 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.017960 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-8cr6b" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.017988 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.017904 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.019164 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.019984 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.022950 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.027343 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n"] Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.179059 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.179099 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.179129 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.179150 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.179166 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.179233 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vxxl\" (UniqueName: \"kubernetes.io/projected/cd2bd342-3a99-4d38-976b-3f213999f103-kube-api-access-5vxxl\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.179255 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.179311 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.179341 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.185682 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.185977 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="90143656-4ff7-4511-8824-88d158f05993" containerName="nova-scheduler-scheduler" containerID="cri-o://6ac9d4f430ea9f350983bbe17059146649a35ad2ed281b9a2c05e082a9ecfdd6" gracePeriod=30 Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.196427 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.196714 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="30cf5aff-9c39-4531-8ee6-1284597c7d00" containerName="nova-api-log" containerID="cri-o://06fab8219fdf4668dcf89d9f07d815244dc4097c711665bc071ec2661de80ad7" gracePeriod=30 Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.196779 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="30cf5aff-9c39-4531-8ee6-1284597c7d00" containerName="nova-api-api" containerID="cri-o://46a37a18b93213b4d6649471f8576ba09563847420101eb829523635836fd3b1" gracePeriod=30 Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.226556 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.227233 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5cd84ceb-8494-4328-8c92-0e818910f06e" containerName="nova-metadata-metadata" containerID="cri-o://23e5a07093be2eca9b799a768274021963eb0c7a728239a891950c901ac444d8" gracePeriod=30 Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.227097 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5cd84ceb-8494-4328-8c92-0e818910f06e" containerName="nova-metadata-log" containerID="cri-o://3db908e2f008cd7cbe619bad66f45cf79b61c7c1c2e6812ee0b8632d24bbf459" gracePeriod=30 Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.308581 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vxxl\" (UniqueName: \"kubernetes.io/projected/cd2bd342-3a99-4d38-976b-3f213999f103-kube-api-access-5vxxl\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.308652 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.308778 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.308834 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.309065 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.309089 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.309131 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.309157 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.309183 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.312749 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.317813 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.318588 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.323641 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.332842 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.333241 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.335439 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vxxl\" (UniqueName: \"kubernetes.io/projected/cd2bd342-3a99-4d38-976b-3f213999f103-kube-api-access-5vxxl\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.336851 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.360904 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: E1201 17:14:16.409193 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2f4d53685fe68630eb2427db02a9e6c2864c578f0eee81039835517195889ab1" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 01 17:14:16 crc kubenswrapper[4810]: E1201 17:14:16.410417 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2f4d53685fe68630eb2427db02a9e6c2864c578f0eee81039835517195889ab1" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 01 17:14:16 crc kubenswrapper[4810]: E1201 17:14:16.412872 4810 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2f4d53685fe68630eb2427db02a9e6c2864c578f0eee81039835517195889ab1" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 01 17:14:16 crc kubenswrapper[4810]: E1201 17:14:16.413042 4810 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="578dc44d-25f5-4ef1-bbe7-eb2328954767" containerName="nova-cell0-conductor-conductor" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.638039 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.758599 4810 generic.go:334] "Generic (PLEG): container finished" podID="5cd84ceb-8494-4328-8c92-0e818910f06e" containerID="3db908e2f008cd7cbe619bad66f45cf79b61c7c1c2e6812ee0b8632d24bbf459" exitCode=143 Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.758961 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5cd84ceb-8494-4328-8c92-0e818910f06e","Type":"ContainerDied","Data":"3db908e2f008cd7cbe619bad66f45cf79b61c7c1c2e6812ee0b8632d24bbf459"} Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.763686 4810 generic.go:334] "Generic (PLEG): container finished" podID="30cf5aff-9c39-4531-8ee6-1284597c7d00" containerID="06fab8219fdf4668dcf89d9f07d815244dc4097c711665bc071ec2661de80ad7" exitCode=143 Dec 01 17:14:16 crc kubenswrapper[4810]: I1201 17:14:16.763741 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30cf5aff-9c39-4531-8ee6-1284597c7d00","Type":"ContainerDied","Data":"06fab8219fdf4668dcf89d9f07d815244dc4097c711665bc071ec2661de80ad7"} Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.200496 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n"] Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.542090 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.639728 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3-config-data\") pod \"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3\" (UID: \"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3\") " Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.640555 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9cqm\" (UniqueName: \"kubernetes.io/projected/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3-kube-api-access-d9cqm\") pod \"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3\" (UID: \"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3\") " Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.640783 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3-combined-ca-bundle\") pod \"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3\" (UID: \"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3\") " Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.654962 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3-kube-api-access-d9cqm" (OuterVolumeSpecName: "kube-api-access-d9cqm") pod "ca965e9d-7a9d-4351-bfa1-36d77a9af2a3" (UID: "ca965e9d-7a9d-4351-bfa1-36d77a9af2a3"). InnerVolumeSpecName "kube-api-access-d9cqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.673782 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3-config-data" (OuterVolumeSpecName: "config-data") pod "ca965e9d-7a9d-4351-bfa1-36d77a9af2a3" (UID: "ca965e9d-7a9d-4351-bfa1-36d77a9af2a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.680817 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca965e9d-7a9d-4351-bfa1-36d77a9af2a3" (UID: "ca965e9d-7a9d-4351-bfa1-36d77a9af2a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.745836 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9cqm\" (UniqueName: \"kubernetes.io/projected/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3-kube-api-access-d9cqm\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.745868 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.745879 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.778053 4810 generic.go:334] "Generic (PLEG): container finished" podID="ca965e9d-7a9d-4351-bfa1-36d77a9af2a3" containerID="13f99bc5259f34633b3dc899ea67dcc6b4654616360f83abebaf1504118d2bb4" exitCode=0 Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.778149 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3","Type":"ContainerDied","Data":"13f99bc5259f34633b3dc899ea67dcc6b4654616360f83abebaf1504118d2bb4"} Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.778177 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ca965e9d-7a9d-4351-bfa1-36d77a9af2a3","Type":"ContainerDied","Data":"85e8754a6969f138363e81f01c3e488808ce08dacd0739daf10c8f079869e0ba"} Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.778175 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.778588 4810 scope.go:117] "RemoveContainer" containerID="13f99bc5259f34633b3dc899ea67dcc6b4654616360f83abebaf1504118d2bb4" Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.780168 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" event={"ID":"cd2bd342-3a99-4d38-976b-3f213999f103","Type":"ContainerStarted","Data":"41b90709b8dabbccab1fe936cfccacc94f0ae01b94a55c928877b3ed43298f88"} Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.803390 4810 scope.go:117] "RemoveContainer" containerID="13f99bc5259f34633b3dc899ea67dcc6b4654616360f83abebaf1504118d2bb4" Dec 01 17:14:17 crc kubenswrapper[4810]: E1201 17:14:17.803979 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13f99bc5259f34633b3dc899ea67dcc6b4654616360f83abebaf1504118d2bb4\": container with ID starting with 13f99bc5259f34633b3dc899ea67dcc6b4654616360f83abebaf1504118d2bb4 not found: ID does not exist" containerID="13f99bc5259f34633b3dc899ea67dcc6b4654616360f83abebaf1504118d2bb4" Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.804020 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13f99bc5259f34633b3dc899ea67dcc6b4654616360f83abebaf1504118d2bb4"} err="failed to get container status \"13f99bc5259f34633b3dc899ea67dcc6b4654616360f83abebaf1504118d2bb4\": rpc error: code = NotFound desc = could not find container \"13f99bc5259f34633b3dc899ea67dcc6b4654616360f83abebaf1504118d2bb4\": container with ID starting with 13f99bc5259f34633b3dc899ea67dcc6b4654616360f83abebaf1504118d2bb4 not found: ID does not exist" Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.814096 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.825951 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.839457 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 17:14:17 crc kubenswrapper[4810]: E1201 17:14:17.840008 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca965e9d-7a9d-4351-bfa1-36d77a9af2a3" containerName="nova-cell1-conductor-conductor" Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.840024 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca965e9d-7a9d-4351-bfa1-36d77a9af2a3" containerName="nova-cell1-conductor-conductor" Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.840271 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca965e9d-7a9d-4351-bfa1-36d77a9af2a3" containerName="nova-cell1-conductor-conductor" Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.840963 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.844621 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.849600 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.950459 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c3b82f-7a29-4e75-b942-befc510f892f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"56c3b82f-7a29-4e75-b942-befc510f892f\") " pod="openstack/nova-cell1-conductor-0" Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.950571 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmqkd\" (UniqueName: \"kubernetes.io/projected/56c3b82f-7a29-4e75-b942-befc510f892f-kube-api-access-cmqkd\") pod \"nova-cell1-conductor-0\" (UID: \"56c3b82f-7a29-4e75-b942-befc510f892f\") " pod="openstack/nova-cell1-conductor-0" Dec 01 17:14:17 crc kubenswrapper[4810]: I1201 17:14:17.950660 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56c3b82f-7a29-4e75-b942-befc510f892f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"56c3b82f-7a29-4e75-b942-befc510f892f\") " pod="openstack/nova-cell1-conductor-0" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.052209 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56c3b82f-7a29-4e75-b942-befc510f892f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"56c3b82f-7a29-4e75-b942-befc510f892f\") " pod="openstack/nova-cell1-conductor-0" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.052358 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c3b82f-7a29-4e75-b942-befc510f892f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"56c3b82f-7a29-4e75-b942-befc510f892f\") " pod="openstack/nova-cell1-conductor-0" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.052397 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmqkd\" (UniqueName: \"kubernetes.io/projected/56c3b82f-7a29-4e75-b942-befc510f892f-kube-api-access-cmqkd\") pod \"nova-cell1-conductor-0\" (UID: \"56c3b82f-7a29-4e75-b942-befc510f892f\") " pod="openstack/nova-cell1-conductor-0" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.056953 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56c3b82f-7a29-4e75-b942-befc510f892f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"56c3b82f-7a29-4e75-b942-befc510f892f\") " pod="openstack/nova-cell1-conductor-0" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.057096 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c3b82f-7a29-4e75-b942-befc510f892f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"56c3b82f-7a29-4e75-b942-befc510f892f\") " pod="openstack/nova-cell1-conductor-0" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.074488 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmqkd\" (UniqueName: \"kubernetes.io/projected/56c3b82f-7a29-4e75-b942-befc510f892f-kube-api-access-cmqkd\") pod \"nova-cell1-conductor-0\" (UID: \"56c3b82f-7a29-4e75-b942-befc510f892f\") " pod="openstack/nova-cell1-conductor-0" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.163208 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.505432 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca965e9d-7a9d-4351-bfa1-36d77a9af2a3" path="/var/lib/kubelet/pods/ca965e9d-7a9d-4351-bfa1-36d77a9af2a3/volumes" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.662873 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.667740 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90143656-4ff7-4511-8824-88d158f05993-config-data\") pod \"90143656-4ff7-4511-8824-88d158f05993\" (UID: \"90143656-4ff7-4511-8824-88d158f05993\") " Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.667949 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz6fj\" (UniqueName: \"kubernetes.io/projected/90143656-4ff7-4511-8824-88d158f05993-kube-api-access-nz6fj\") pod \"90143656-4ff7-4511-8824-88d158f05993\" (UID: \"90143656-4ff7-4511-8824-88d158f05993\") " Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.668224 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90143656-4ff7-4511-8824-88d158f05993-combined-ca-bundle\") pod \"90143656-4ff7-4511-8824-88d158f05993\" (UID: \"90143656-4ff7-4511-8824-88d158f05993\") " Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.673366 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90143656-4ff7-4511-8824-88d158f05993-kube-api-access-nz6fj" (OuterVolumeSpecName: "kube-api-access-nz6fj") pod "90143656-4ff7-4511-8824-88d158f05993" (UID: "90143656-4ff7-4511-8824-88d158f05993"). InnerVolumeSpecName "kube-api-access-nz6fj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.712723 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90143656-4ff7-4511-8824-88d158f05993-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "90143656-4ff7-4511-8824-88d158f05993" (UID: "90143656-4ff7-4511-8824-88d158f05993"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.750676 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 17:14:18 crc kubenswrapper[4810]: W1201 17:14:18.758841 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56c3b82f_7a29_4e75_b942_befc510f892f.slice/crio-5605659c2b41459b85ad621324ace7885e1d1b7af1daf67eb3ef454727c1ea65 WatchSource:0}: Error finding container 5605659c2b41459b85ad621324ace7885e1d1b7af1daf67eb3ef454727c1ea65: Status 404 returned error can't find the container with id 5605659c2b41459b85ad621324ace7885e1d1b7af1daf67eb3ef454727c1ea65 Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.769946 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz6fj\" (UniqueName: \"kubernetes.io/projected/90143656-4ff7-4511-8824-88d158f05993-kube-api-access-nz6fj\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.769984 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90143656-4ff7-4511-8824-88d158f05993-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.778649 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90143656-4ff7-4511-8824-88d158f05993-config-data" (OuterVolumeSpecName: "config-data") pod "90143656-4ff7-4511-8824-88d158f05993" (UID: "90143656-4ff7-4511-8824-88d158f05993"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.799135 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"56c3b82f-7a29-4e75-b942-befc510f892f","Type":"ContainerStarted","Data":"5605659c2b41459b85ad621324ace7885e1d1b7af1daf67eb3ef454727c1ea65"} Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.804182 4810 generic.go:334] "Generic (PLEG): container finished" podID="90143656-4ff7-4511-8824-88d158f05993" containerID="6ac9d4f430ea9f350983bbe17059146649a35ad2ed281b9a2c05e082a9ecfdd6" exitCode=0 Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.804243 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"90143656-4ff7-4511-8824-88d158f05993","Type":"ContainerDied","Data":"6ac9d4f430ea9f350983bbe17059146649a35ad2ed281b9a2c05e082a9ecfdd6"} Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.804267 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.804608 4810 scope.go:117] "RemoveContainer" containerID="6ac9d4f430ea9f350983bbe17059146649a35ad2ed281b9a2c05e082a9ecfdd6" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.804535 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"90143656-4ff7-4511-8824-88d158f05993","Type":"ContainerDied","Data":"d3add81d4db38a45e441e110f9e1a43a9361e4f71ff23d464d24dfe826c72bd0"} Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.831701 4810 scope.go:117] "RemoveContainer" containerID="6ac9d4f430ea9f350983bbe17059146649a35ad2ed281b9a2c05e082a9ecfdd6" Dec 01 17:14:18 crc kubenswrapper[4810]: E1201 17:14:18.832142 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ac9d4f430ea9f350983bbe17059146649a35ad2ed281b9a2c05e082a9ecfdd6\": container with ID starting with 6ac9d4f430ea9f350983bbe17059146649a35ad2ed281b9a2c05e082a9ecfdd6 not found: ID does not exist" containerID="6ac9d4f430ea9f350983bbe17059146649a35ad2ed281b9a2c05e082a9ecfdd6" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.832176 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ac9d4f430ea9f350983bbe17059146649a35ad2ed281b9a2c05e082a9ecfdd6"} err="failed to get container status \"6ac9d4f430ea9f350983bbe17059146649a35ad2ed281b9a2c05e082a9ecfdd6\": rpc error: code = NotFound desc = could not find container \"6ac9d4f430ea9f350983bbe17059146649a35ad2ed281b9a2c05e082a9ecfdd6\": container with ID starting with 6ac9d4f430ea9f350983bbe17059146649a35ad2ed281b9a2c05e082a9ecfdd6 not found: ID does not exist" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.860249 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.881854 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90143656-4ff7-4511-8824-88d158f05993-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.885704 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.900236 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:14:18 crc kubenswrapper[4810]: E1201 17:14:18.901147 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90143656-4ff7-4511-8824-88d158f05993" containerName="nova-scheduler-scheduler" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.901176 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="90143656-4ff7-4511-8824-88d158f05993" containerName="nova-scheduler-scheduler" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.902553 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="90143656-4ff7-4511-8824-88d158f05993" containerName="nova-scheduler-scheduler" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.903777 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.905941 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 17:14:18 crc kubenswrapper[4810]: I1201 17:14:18.913222 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.086751 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82f24148-c70d-4110-934e-f10d5b32eb1e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"82f24148-c70d-4110-934e-f10d5b32eb1e\") " pod="openstack/nova-scheduler-0" Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.086810 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cwft\" (UniqueName: \"kubernetes.io/projected/82f24148-c70d-4110-934e-f10d5b32eb1e-kube-api-access-5cwft\") pod \"nova-scheduler-0\" (UID: \"82f24148-c70d-4110-934e-f10d5b32eb1e\") " pod="openstack/nova-scheduler-0" Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.086897 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82f24148-c70d-4110-934e-f10d5b32eb1e-config-data\") pod \"nova-scheduler-0\" (UID: \"82f24148-c70d-4110-934e-f10d5b32eb1e\") " pod="openstack/nova-scheduler-0" Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.188584 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82f24148-c70d-4110-934e-f10d5b32eb1e-config-data\") pod \"nova-scheduler-0\" (UID: \"82f24148-c70d-4110-934e-f10d5b32eb1e\") " pod="openstack/nova-scheduler-0" Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.188854 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82f24148-c70d-4110-934e-f10d5b32eb1e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"82f24148-c70d-4110-934e-f10d5b32eb1e\") " pod="openstack/nova-scheduler-0" Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.188889 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cwft\" (UniqueName: \"kubernetes.io/projected/82f24148-c70d-4110-934e-f10d5b32eb1e-kube-api-access-5cwft\") pod \"nova-scheduler-0\" (UID: \"82f24148-c70d-4110-934e-f10d5b32eb1e\") " pod="openstack/nova-scheduler-0" Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.193529 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82f24148-c70d-4110-934e-f10d5b32eb1e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"82f24148-c70d-4110-934e-f10d5b32eb1e\") " pod="openstack/nova-scheduler-0" Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.195169 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82f24148-c70d-4110-934e-f10d5b32eb1e-config-data\") pod \"nova-scheduler-0\" (UID: \"82f24148-c70d-4110-934e-f10d5b32eb1e\") " pod="openstack/nova-scheduler-0" Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.215666 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cwft\" (UniqueName: \"kubernetes.io/projected/82f24148-c70d-4110-934e-f10d5b32eb1e-kube-api-access-5cwft\") pod \"nova-scheduler-0\" (UID: \"82f24148-c70d-4110-934e-f10d5b32eb1e\") " pod="openstack/nova-scheduler-0" Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.229216 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.391236 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="30cf5aff-9c39-4531-8ee6-1284597c7d00" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.95:8774/\": read tcp 10.217.0.2:39022->10.217.1.95:8774: read: connection reset by peer" Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.392072 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="30cf5aff-9c39-4531-8ee6-1284597c7d00" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.95:8774/\": read tcp 10.217.0.2:39028->10.217.1.95:8774: read: connection reset by peer" Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.548057 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="5cd84ceb-8494-4328-8c92-0e818910f06e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.92:8775/\": dial tcp 10.217.1.92:8775: connect: connection refused" Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.548057 4810 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="5cd84ceb-8494-4328-8c92-0e818910f06e" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.92:8775/\": dial tcp 10.217.1.92:8775: connect: connection refused" Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.810585 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.821234 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"56c3b82f-7a29-4e75-b942-befc510f892f","Type":"ContainerStarted","Data":"31df946412952c9303370612cdb2fbbc4baa924f3d5847fbfc78cf1fc109f9f1"} Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.821395 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.824094 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" event={"ID":"cd2bd342-3a99-4d38-976b-3f213999f103","Type":"ContainerStarted","Data":"48a8b8c372a009c3c38018c2a3556217a6d13a8614a677422be23703f6aab6fa"} Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.842389 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.842361469 podStartE2EDuration="2.842361469s" podCreationTimestamp="2025-12-01 17:14:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:14:19.839377558 +0000 UTC m=+9625.602887161" watchObservedRunningTime="2025-12-01 17:14:19.842361469 +0000 UTC m=+9625.605871092" Dec 01 17:14:19 crc kubenswrapper[4810]: I1201 17:14:19.866679 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" podStartSLOduration=3.674122638 podStartE2EDuration="4.866659943s" podCreationTimestamp="2025-12-01 17:14:15 +0000 UTC" firstStartedPulling="2025-12-01 17:14:17.208698391 +0000 UTC m=+9622.972207994" lastFinishedPulling="2025-12-01 17:14:18.401235696 +0000 UTC m=+9624.164745299" observedRunningTime="2025-12-01 17:14:19.861182326 +0000 UTC m=+9625.624691949" watchObservedRunningTime="2025-12-01 17:14:19.866659943 +0000 UTC m=+9625.630169546" Dec 01 17:14:20 crc kubenswrapper[4810]: W1201 17:14:20.073867 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82f24148_c70d_4110_934e_f10d5b32eb1e.slice/crio-e69ab7ec8fc645da2a7cf0f0c468811a1b21ba263378a607564e8e73182ae0a0 WatchSource:0}: Error finding container e69ab7ec8fc645da2a7cf0f0c468811a1b21ba263378a607564e8e73182ae0a0: Status 404 returned error can't find the container with id e69ab7ec8fc645da2a7cf0f0c468811a1b21ba263378a607564e8e73182ae0a0 Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.503313 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90143656-4ff7-4511-8824-88d158f05993" path="/var/lib/kubelet/pods/90143656-4ff7-4511-8824-88d158f05993/volumes" Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.577867 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.724488 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt2nv\" (UniqueName: \"kubernetes.io/projected/578dc44d-25f5-4ef1-bbe7-eb2328954767-kube-api-access-mt2nv\") pod \"578dc44d-25f5-4ef1-bbe7-eb2328954767\" (UID: \"578dc44d-25f5-4ef1-bbe7-eb2328954767\") " Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.724666 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/578dc44d-25f5-4ef1-bbe7-eb2328954767-config-data\") pod \"578dc44d-25f5-4ef1-bbe7-eb2328954767\" (UID: \"578dc44d-25f5-4ef1-bbe7-eb2328954767\") " Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.724829 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/578dc44d-25f5-4ef1-bbe7-eb2328954767-combined-ca-bundle\") pod \"578dc44d-25f5-4ef1-bbe7-eb2328954767\" (UID: \"578dc44d-25f5-4ef1-bbe7-eb2328954767\") " Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.752656 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/578dc44d-25f5-4ef1-bbe7-eb2328954767-kube-api-access-mt2nv" (OuterVolumeSpecName: "kube-api-access-mt2nv") pod "578dc44d-25f5-4ef1-bbe7-eb2328954767" (UID: "578dc44d-25f5-4ef1-bbe7-eb2328954767"). InnerVolumeSpecName "kube-api-access-mt2nv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.799665 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/578dc44d-25f5-4ef1-bbe7-eb2328954767-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "578dc44d-25f5-4ef1-bbe7-eb2328954767" (UID: "578dc44d-25f5-4ef1-bbe7-eb2328954767"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.799969 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/578dc44d-25f5-4ef1-bbe7-eb2328954767-config-data" (OuterVolumeSpecName: "config-data") pod "578dc44d-25f5-4ef1-bbe7-eb2328954767" (UID: "578dc44d-25f5-4ef1-bbe7-eb2328954767"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.827494 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/578dc44d-25f5-4ef1-bbe7-eb2328954767-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.827530 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt2nv\" (UniqueName: \"kubernetes.io/projected/578dc44d-25f5-4ef1-bbe7-eb2328954767-kube-api-access-mt2nv\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.827540 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/578dc44d-25f5-4ef1-bbe7-eb2328954767-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.844344 4810 generic.go:334] "Generic (PLEG): container finished" podID="5cd84ceb-8494-4328-8c92-0e818910f06e" containerID="23e5a07093be2eca9b799a768274021963eb0c7a728239a891950c901ac444d8" exitCode=0 Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.844409 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5cd84ceb-8494-4328-8c92-0e818910f06e","Type":"ContainerDied","Data":"23e5a07093be2eca9b799a768274021963eb0c7a728239a891950c901ac444d8"} Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.845661 4810 generic.go:334] "Generic (PLEG): container finished" podID="578dc44d-25f5-4ef1-bbe7-eb2328954767" containerID="2f4d53685fe68630eb2427db02a9e6c2864c578f0eee81039835517195889ab1" exitCode=0 Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.845713 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"578dc44d-25f5-4ef1-bbe7-eb2328954767","Type":"ContainerDied","Data":"2f4d53685fe68630eb2427db02a9e6c2864c578f0eee81039835517195889ab1"} Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.845733 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"578dc44d-25f5-4ef1-bbe7-eb2328954767","Type":"ContainerDied","Data":"a35bd522cf8087246aea95321c33263df1b1816e1052396c74e09e5eee743665"} Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.845753 4810 scope.go:117] "RemoveContainer" containerID="2f4d53685fe68630eb2427db02a9e6c2864c578f0eee81039835517195889ab1" Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.845985 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.857262 4810 generic.go:334] "Generic (PLEG): container finished" podID="30cf5aff-9c39-4531-8ee6-1284597c7d00" containerID="46a37a18b93213b4d6649471f8576ba09563847420101eb829523635836fd3b1" exitCode=0 Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.857662 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30cf5aff-9c39-4531-8ee6-1284597c7d00","Type":"ContainerDied","Data":"46a37a18b93213b4d6649471f8576ba09563847420101eb829523635836fd3b1"} Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.859549 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"82f24148-c70d-4110-934e-f10d5b32eb1e","Type":"ContainerStarted","Data":"4e7c88746c3cf59de36143a8dbb3f4ca31b1514b0a57424dad78435925cf55a9"} Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.859853 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"82f24148-c70d-4110-934e-f10d5b32eb1e","Type":"ContainerStarted","Data":"e69ab7ec8fc645da2a7cf0f0c468811a1b21ba263378a607564e8e73182ae0a0"} Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.885627 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.885608197 podStartE2EDuration="2.885608197s" podCreationTimestamp="2025-12-01 17:14:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:14:20.876267076 +0000 UTC m=+9626.639776689" watchObservedRunningTime="2025-12-01 17:14:20.885608197 +0000 UTC m=+9626.649117800" Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.952820 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.971071 4810 scope.go:117] "RemoveContainer" containerID="2f4d53685fe68630eb2427db02a9e6c2864c578f0eee81039835517195889ab1" Dec 01 17:14:20 crc kubenswrapper[4810]: E1201 17:14:20.972008 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f4d53685fe68630eb2427db02a9e6c2864c578f0eee81039835517195889ab1\": container with ID starting with 2f4d53685fe68630eb2427db02a9e6c2864c578f0eee81039835517195889ab1 not found: ID does not exist" containerID="2f4d53685fe68630eb2427db02a9e6c2864c578f0eee81039835517195889ab1" Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.972072 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f4d53685fe68630eb2427db02a9e6c2864c578f0eee81039835517195889ab1"} err="failed to get container status \"2f4d53685fe68630eb2427db02a9e6c2864c578f0eee81039835517195889ab1\": rpc error: code = NotFound desc = could not find container \"2f4d53685fe68630eb2427db02a9e6c2864c578f0eee81039835517195889ab1\": container with ID starting with 2f4d53685fe68630eb2427db02a9e6c2864c578f0eee81039835517195889ab1 not found: ID does not exist" Dec 01 17:14:20 crc kubenswrapper[4810]: I1201 17:14:20.984991 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.004079 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 17:14:21 crc kubenswrapper[4810]: E1201 17:14:21.004617 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="578dc44d-25f5-4ef1-bbe7-eb2328954767" containerName="nova-cell0-conductor-conductor" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.004630 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="578dc44d-25f5-4ef1-bbe7-eb2328954767" containerName="nova-cell0-conductor-conductor" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.005092 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="578dc44d-25f5-4ef1-bbe7-eb2328954767" containerName="nova-cell0-conductor-conductor" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.005921 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.009311 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.024693 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.046642 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.099077 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.137052 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-internal-tls-certs\") pod \"30cf5aff-9c39-4531-8ee6-1284597c7d00\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.137132 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-config-data\") pod \"30cf5aff-9c39-4531-8ee6-1284597c7d00\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.137163 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-public-tls-certs\") pod \"30cf5aff-9c39-4531-8ee6-1284597c7d00\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.137590 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hj4hq\" (UniqueName: \"kubernetes.io/projected/30cf5aff-9c39-4531-8ee6-1284597c7d00-kube-api-access-hj4hq\") pod \"30cf5aff-9c39-4531-8ee6-1284597c7d00\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.137648 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30cf5aff-9c39-4531-8ee6-1284597c7d00-logs\") pod \"30cf5aff-9c39-4531-8ee6-1284597c7d00\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.137682 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-combined-ca-bundle\") pod \"30cf5aff-9c39-4531-8ee6-1284597c7d00\" (UID: \"30cf5aff-9c39-4531-8ee6-1284597c7d00\") " Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.138037 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7feb5c3-af78-44b4-a415-2a77e1433182-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e7feb5c3-af78-44b4-a415-2a77e1433182\") " pod="openstack/nova-cell0-conductor-0" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.138085 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqfdf\" (UniqueName: \"kubernetes.io/projected/e7feb5c3-af78-44b4-a415-2a77e1433182-kube-api-access-nqfdf\") pod \"nova-cell0-conductor-0\" (UID: \"e7feb5c3-af78-44b4-a415-2a77e1433182\") " pod="openstack/nova-cell0-conductor-0" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.138090 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30cf5aff-9c39-4531-8ee6-1284597c7d00-logs" (OuterVolumeSpecName: "logs") pod "30cf5aff-9c39-4531-8ee6-1284597c7d00" (UID: "30cf5aff-9c39-4531-8ee6-1284597c7d00"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.138247 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7feb5c3-af78-44b4-a415-2a77e1433182-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e7feb5c3-af78-44b4-a415-2a77e1433182\") " pod="openstack/nova-cell0-conductor-0" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.138421 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30cf5aff-9c39-4531-8ee6-1284597c7d00-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.150277 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30cf5aff-9c39-4531-8ee6-1284597c7d00-kube-api-access-hj4hq" (OuterVolumeSpecName: "kube-api-access-hj4hq") pod "30cf5aff-9c39-4531-8ee6-1284597c7d00" (UID: "30cf5aff-9c39-4531-8ee6-1284597c7d00"). InnerVolumeSpecName "kube-api-access-hj4hq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.240322 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cd84ceb-8494-4328-8c92-0e818910f06e-nova-metadata-tls-certs\") pod \"5cd84ceb-8494-4328-8c92-0e818910f06e\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.240411 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cd84ceb-8494-4328-8c92-0e818910f06e-config-data\") pod \"5cd84ceb-8494-4328-8c92-0e818910f06e\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.240532 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cd84ceb-8494-4328-8c92-0e818910f06e-logs\") pod \"5cd84ceb-8494-4328-8c92-0e818910f06e\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.240649 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrf2n\" (UniqueName: \"kubernetes.io/projected/5cd84ceb-8494-4328-8c92-0e818910f06e-kube-api-access-qrf2n\") pod \"5cd84ceb-8494-4328-8c92-0e818910f06e\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.240747 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cd84ceb-8494-4328-8c92-0e818910f06e-combined-ca-bundle\") pod \"5cd84ceb-8494-4328-8c92-0e818910f06e\" (UID: \"5cd84ceb-8494-4328-8c92-0e818910f06e\") " Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.241222 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7feb5c3-af78-44b4-a415-2a77e1433182-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e7feb5c3-af78-44b4-a415-2a77e1433182\") " pod="openstack/nova-cell0-conductor-0" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.241296 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqfdf\" (UniqueName: \"kubernetes.io/projected/e7feb5c3-af78-44b4-a415-2a77e1433182-kube-api-access-nqfdf\") pod \"nova-cell0-conductor-0\" (UID: \"e7feb5c3-af78-44b4-a415-2a77e1433182\") " pod="openstack/nova-cell0-conductor-0" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.241355 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cd84ceb-8494-4328-8c92-0e818910f06e-logs" (OuterVolumeSpecName: "logs") pod "5cd84ceb-8494-4328-8c92-0e818910f06e" (UID: "5cd84ceb-8494-4328-8c92-0e818910f06e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.241423 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7feb5c3-af78-44b4-a415-2a77e1433182-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e7feb5c3-af78-44b4-a415-2a77e1433182\") " pod="openstack/nova-cell0-conductor-0" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.241705 4810 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cd84ceb-8494-4328-8c92-0e818910f06e-logs\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.241722 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hj4hq\" (UniqueName: \"kubernetes.io/projected/30cf5aff-9c39-4531-8ee6-1284597c7d00-kube-api-access-hj4hq\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.244694 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7feb5c3-af78-44b4-a415-2a77e1433182-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e7feb5c3-af78-44b4-a415-2a77e1433182\") " pod="openstack/nova-cell0-conductor-0" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.250739 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cd84ceb-8494-4328-8c92-0e818910f06e-kube-api-access-qrf2n" (OuterVolumeSpecName: "kube-api-access-qrf2n") pod "5cd84ceb-8494-4328-8c92-0e818910f06e" (UID: "5cd84ceb-8494-4328-8c92-0e818910f06e"). InnerVolumeSpecName "kube-api-access-qrf2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.251163 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7feb5c3-af78-44b4-a415-2a77e1433182-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e7feb5c3-af78-44b4-a415-2a77e1433182\") " pod="openstack/nova-cell0-conductor-0" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.344118 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrf2n\" (UniqueName: \"kubernetes.io/projected/5cd84ceb-8494-4328-8c92-0e818910f06e-kube-api-access-qrf2n\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.852693 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqfdf\" (UniqueName: \"kubernetes.io/projected/e7feb5c3-af78-44b4-a415-2a77e1433182-kube-api-access-nqfdf\") pod \"nova-cell0-conductor-0\" (UID: \"e7feb5c3-af78-44b4-a415-2a77e1433182\") " pod="openstack/nova-cell0-conductor-0" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.892316 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30cf5aff-9c39-4531-8ee6-1284597c7d00" (UID: "30cf5aff-9c39-4531-8ee6-1284597c7d00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.895368 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.895368 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30cf5aff-9c39-4531-8ee6-1284597c7d00","Type":"ContainerDied","Data":"19f62766348c357265510cc3791251351783c7de7e96e035ed79b683c0340489"} Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.895642 4810 scope.go:117] "RemoveContainer" containerID="46a37a18b93213b4d6649471f8576ba09563847420101eb829523635836fd3b1" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.903986 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5cd84ceb-8494-4328-8c92-0e818910f06e","Type":"ContainerDied","Data":"f72abb675295eaf87b10cde1dedea6029b101469ceaaf0e22524854850ad6ca2"} Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.904009 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.969133 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:21 crc kubenswrapper[4810]: I1201 17:14:21.971788 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.001555 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cd84ceb-8494-4328-8c92-0e818910f06e-config-data" (OuterVolumeSpecName: "config-data") pod "5cd84ceb-8494-4328-8c92-0e818910f06e" (UID: "5cd84ceb-8494-4328-8c92-0e818910f06e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.021007 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "30cf5aff-9c39-4531-8ee6-1284597c7d00" (UID: "30cf5aff-9c39-4531-8ee6-1284597c7d00"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.034870 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-config-data" (OuterVolumeSpecName: "config-data") pod "30cf5aff-9c39-4531-8ee6-1284597c7d00" (UID: "30cf5aff-9c39-4531-8ee6-1284597c7d00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.042896 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cd84ceb-8494-4328-8c92-0e818910f06e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5cd84ceb-8494-4328-8c92-0e818910f06e" (UID: "5cd84ceb-8494-4328-8c92-0e818910f06e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.050205 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cd84ceb-8494-4328-8c92-0e818910f06e-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "5cd84ceb-8494-4328-8c92-0e818910f06e" (UID: "5cd84ceb-8494-4328-8c92-0e818910f06e"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.071319 4810 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cd84ceb-8494-4328-8c92-0e818910f06e-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.071347 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cd84ceb-8494-4328-8c92-0e818910f06e-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.071358 4810 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.071366 4810 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.071376 4810 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cd84ceb-8494-4328-8c92-0e818910f06e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.080333 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "30cf5aff-9c39-4531-8ee6-1284597c7d00" (UID: "30cf5aff-9c39-4531-8ee6-1284597c7d00"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.180310 4810 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30cf5aff-9c39-4531-8ee6-1284597c7d00-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.205214 4810 scope.go:117] "RemoveContainer" containerID="06fab8219fdf4668dcf89d9f07d815244dc4097c711665bc071ec2661de80ad7" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.232265 4810 scope.go:117] "RemoveContainer" containerID="23e5a07093be2eca9b799a768274021963eb0c7a728239a891950c901ac444d8" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.263276 4810 scope.go:117] "RemoveContainer" containerID="3db908e2f008cd7cbe619bad66f45cf79b61c7c1c2e6812ee0b8632d24bbf459" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.291042 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.322633 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.342517 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.355269 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.368143 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 17:14:22 crc kubenswrapper[4810]: E1201 17:14:22.368628 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30cf5aff-9c39-4531-8ee6-1284597c7d00" containerName="nova-api-log" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.368645 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="30cf5aff-9c39-4531-8ee6-1284597c7d00" containerName="nova-api-log" Dec 01 17:14:22 crc kubenswrapper[4810]: E1201 17:14:22.368663 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd84ceb-8494-4328-8c92-0e818910f06e" containerName="nova-metadata-metadata" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.368669 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd84ceb-8494-4328-8c92-0e818910f06e" containerName="nova-metadata-metadata" Dec 01 17:14:22 crc kubenswrapper[4810]: E1201 17:14:22.368689 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd84ceb-8494-4328-8c92-0e818910f06e" containerName="nova-metadata-log" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.368696 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd84ceb-8494-4328-8c92-0e818910f06e" containerName="nova-metadata-log" Dec 01 17:14:22 crc kubenswrapper[4810]: E1201 17:14:22.368717 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30cf5aff-9c39-4531-8ee6-1284597c7d00" containerName="nova-api-api" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.368723 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="30cf5aff-9c39-4531-8ee6-1284597c7d00" containerName="nova-api-api" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.368910 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cd84ceb-8494-4328-8c92-0e818910f06e" containerName="nova-metadata-log" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.369026 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="30cf5aff-9c39-4531-8ee6-1284597c7d00" containerName="nova-api-api" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.369045 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="30cf5aff-9c39-4531-8ee6-1284597c7d00" containerName="nova-api-log" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.369060 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cd84ceb-8494-4328-8c92-0e818910f06e" containerName="nova-metadata-metadata" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.370353 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.371954 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.372295 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.372807 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.378652 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.381540 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.390166 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.390725 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.390964 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.401325 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.484646 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 17:14:22 crc kubenswrapper[4810]: W1201 17:14:22.485950 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7feb5c3_af78_44b4_a415_2a77e1433182.slice/crio-6e31372542915a7896938cff04005d612164142b9047f8db9d60e7f38900d4d9 WatchSource:0}: Error finding container 6e31372542915a7896938cff04005d612164142b9047f8db9d60e7f38900d4d9: Status 404 returned error can't find the container with id 6e31372542915a7896938cff04005d612164142b9047f8db9d60e7f38900d4d9 Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.490549 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f44e8294-c07f-4c82-8957-82b096b3d87b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f44e8294-c07f-4c82-8957-82b096b3d87b\") " pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.490588 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f44e8294-c07f-4c82-8957-82b096b3d87b-logs\") pod \"nova-api-0\" (UID: \"f44e8294-c07f-4c82-8957-82b096b3d87b\") " pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.490620 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f44e8294-c07f-4c82-8957-82b096b3d87b-config-data\") pod \"nova-api-0\" (UID: \"f44e8294-c07f-4c82-8957-82b096b3d87b\") " pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.490656 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmrsx\" (UniqueName: \"kubernetes.io/projected/03fe61c9-4219-44e3-b659-6c47a5a6ee42-kube-api-access-fmrsx\") pod \"nova-metadata-0\" (UID: \"03fe61c9-4219-44e3-b659-6c47a5a6ee42\") " pod="openstack/nova-metadata-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.490712 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/03fe61c9-4219-44e3-b659-6c47a5a6ee42-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"03fe61c9-4219-44e3-b659-6c47a5a6ee42\") " pod="openstack/nova-metadata-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.490761 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f44e8294-c07f-4c82-8957-82b096b3d87b-public-tls-certs\") pod \"nova-api-0\" (UID: \"f44e8294-c07f-4c82-8957-82b096b3d87b\") " pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.490777 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zfzx\" (UniqueName: \"kubernetes.io/projected/f44e8294-c07f-4c82-8957-82b096b3d87b-kube-api-access-4zfzx\") pod \"nova-api-0\" (UID: \"f44e8294-c07f-4c82-8957-82b096b3d87b\") " pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.490808 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03fe61c9-4219-44e3-b659-6c47a5a6ee42-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"03fe61c9-4219-44e3-b659-6c47a5a6ee42\") " pod="openstack/nova-metadata-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.490828 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f44e8294-c07f-4c82-8957-82b096b3d87b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f44e8294-c07f-4c82-8957-82b096b3d87b\") " pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.490853 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03fe61c9-4219-44e3-b659-6c47a5a6ee42-logs\") pod \"nova-metadata-0\" (UID: \"03fe61c9-4219-44e3-b659-6c47a5a6ee42\") " pod="openstack/nova-metadata-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.490877 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03fe61c9-4219-44e3-b659-6c47a5a6ee42-config-data\") pod \"nova-metadata-0\" (UID: \"03fe61c9-4219-44e3-b659-6c47a5a6ee42\") " pod="openstack/nova-metadata-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.503971 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30cf5aff-9c39-4531-8ee6-1284597c7d00" path="/var/lib/kubelet/pods/30cf5aff-9c39-4531-8ee6-1284597c7d00/volumes" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.504723 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="578dc44d-25f5-4ef1-bbe7-eb2328954767" path="/var/lib/kubelet/pods/578dc44d-25f5-4ef1-bbe7-eb2328954767/volumes" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.505242 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cd84ceb-8494-4328-8c92-0e818910f06e" path="/var/lib/kubelet/pods/5cd84ceb-8494-4328-8c92-0e818910f06e/volumes" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.593012 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f44e8294-c07f-4c82-8957-82b096b3d87b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f44e8294-c07f-4c82-8957-82b096b3d87b\") " pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.593066 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f44e8294-c07f-4c82-8957-82b096b3d87b-logs\") pod \"nova-api-0\" (UID: \"f44e8294-c07f-4c82-8957-82b096b3d87b\") " pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.593118 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f44e8294-c07f-4c82-8957-82b096b3d87b-config-data\") pod \"nova-api-0\" (UID: \"f44e8294-c07f-4c82-8957-82b096b3d87b\") " pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.593562 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f44e8294-c07f-4c82-8957-82b096b3d87b-logs\") pod \"nova-api-0\" (UID: \"f44e8294-c07f-4c82-8957-82b096b3d87b\") " pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.593793 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmrsx\" (UniqueName: \"kubernetes.io/projected/03fe61c9-4219-44e3-b659-6c47a5a6ee42-kube-api-access-fmrsx\") pod \"nova-metadata-0\" (UID: \"03fe61c9-4219-44e3-b659-6c47a5a6ee42\") " pod="openstack/nova-metadata-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.594004 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/03fe61c9-4219-44e3-b659-6c47a5a6ee42-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"03fe61c9-4219-44e3-b659-6c47a5a6ee42\") " pod="openstack/nova-metadata-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.594112 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f44e8294-c07f-4c82-8957-82b096b3d87b-public-tls-certs\") pod \"nova-api-0\" (UID: \"f44e8294-c07f-4c82-8957-82b096b3d87b\") " pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.594148 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zfzx\" (UniqueName: \"kubernetes.io/projected/f44e8294-c07f-4c82-8957-82b096b3d87b-kube-api-access-4zfzx\") pod \"nova-api-0\" (UID: \"f44e8294-c07f-4c82-8957-82b096b3d87b\") " pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.594248 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03fe61c9-4219-44e3-b659-6c47a5a6ee42-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"03fe61c9-4219-44e3-b659-6c47a5a6ee42\") " pod="openstack/nova-metadata-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.594281 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f44e8294-c07f-4c82-8957-82b096b3d87b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f44e8294-c07f-4c82-8957-82b096b3d87b\") " pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.594319 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03fe61c9-4219-44e3-b659-6c47a5a6ee42-logs\") pod \"nova-metadata-0\" (UID: \"03fe61c9-4219-44e3-b659-6c47a5a6ee42\") " pod="openstack/nova-metadata-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.594381 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03fe61c9-4219-44e3-b659-6c47a5a6ee42-config-data\") pod \"nova-metadata-0\" (UID: \"03fe61c9-4219-44e3-b659-6c47a5a6ee42\") " pod="openstack/nova-metadata-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.595095 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03fe61c9-4219-44e3-b659-6c47a5a6ee42-logs\") pod \"nova-metadata-0\" (UID: \"03fe61c9-4219-44e3-b659-6c47a5a6ee42\") " pod="openstack/nova-metadata-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.597328 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/03fe61c9-4219-44e3-b659-6c47a5a6ee42-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"03fe61c9-4219-44e3-b659-6c47a5a6ee42\") " pod="openstack/nova-metadata-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.597788 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f44e8294-c07f-4c82-8957-82b096b3d87b-config-data\") pod \"nova-api-0\" (UID: \"f44e8294-c07f-4c82-8957-82b096b3d87b\") " pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.597974 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03fe61c9-4219-44e3-b659-6c47a5a6ee42-config-data\") pod \"nova-metadata-0\" (UID: \"03fe61c9-4219-44e3-b659-6c47a5a6ee42\") " pod="openstack/nova-metadata-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.599407 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03fe61c9-4219-44e3-b659-6c47a5a6ee42-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"03fe61c9-4219-44e3-b659-6c47a5a6ee42\") " pod="openstack/nova-metadata-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.599989 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f44e8294-c07f-4c82-8957-82b096b3d87b-public-tls-certs\") pod \"nova-api-0\" (UID: \"f44e8294-c07f-4c82-8957-82b096b3d87b\") " pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.600554 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f44e8294-c07f-4c82-8957-82b096b3d87b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f44e8294-c07f-4c82-8957-82b096b3d87b\") " pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.603167 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f44e8294-c07f-4c82-8957-82b096b3d87b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f44e8294-c07f-4c82-8957-82b096b3d87b\") " pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.615617 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zfzx\" (UniqueName: \"kubernetes.io/projected/f44e8294-c07f-4c82-8957-82b096b3d87b-kube-api-access-4zfzx\") pod \"nova-api-0\" (UID: \"f44e8294-c07f-4c82-8957-82b096b3d87b\") " pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.623488 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmrsx\" (UniqueName: \"kubernetes.io/projected/03fe61c9-4219-44e3-b659-6c47a5a6ee42-kube-api-access-fmrsx\") pod \"nova-metadata-0\" (UID: \"03fe61c9-4219-44e3-b659-6c47a5a6ee42\") " pod="openstack/nova-metadata-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.697030 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.704298 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.928295 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e7feb5c3-af78-44b4-a415-2a77e1433182","Type":"ContainerStarted","Data":"fdd9ca399878f811fc26b4274e83a10af65d4c51c51ebc774d05f74b1adb8be1"} Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.928628 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e7feb5c3-af78-44b4-a415-2a77e1433182","Type":"ContainerStarted","Data":"6e31372542915a7896938cff04005d612164142b9047f8db9d60e7f38900d4d9"} Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.929640 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 01 17:14:22 crc kubenswrapper[4810]: I1201 17:14:22.965632 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.965612999 podStartE2EDuration="2.965612999s" podCreationTimestamp="2025-12-01 17:14:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:14:22.962605168 +0000 UTC m=+9628.726114781" watchObservedRunningTime="2025-12-01 17:14:22.965612999 +0000 UTC m=+9628.729122602" Dec 01 17:14:23 crc kubenswrapper[4810]: I1201 17:14:23.205666 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 01 17:14:23 crc kubenswrapper[4810]: I1201 17:14:23.215010 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 17:14:23 crc kubenswrapper[4810]: W1201 17:14:23.234615 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf44e8294_c07f_4c82_8957_82b096b3d87b.slice/crio-8d4d8f2cbcb61d4d1bc071138797d2267e6874a9bb97bf8af1033e30388473ad WatchSource:0}: Error finding container 8d4d8f2cbcb61d4d1bc071138797d2267e6874a9bb97bf8af1033e30388473ad: Status 404 returned error can't find the container with id 8d4d8f2cbcb61d4d1bc071138797d2267e6874a9bb97bf8af1033e30388473ad Dec 01 17:14:23 crc kubenswrapper[4810]: I1201 17:14:23.322988 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 17:14:23 crc kubenswrapper[4810]: I1201 17:14:23.941799 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"03fe61c9-4219-44e3-b659-6c47a5a6ee42","Type":"ContainerStarted","Data":"b1f8994b98238924e86cea1852221812d634ff8d7e7329172bc6a419c3b841cf"} Dec 01 17:14:23 crc kubenswrapper[4810]: I1201 17:14:23.942227 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"03fe61c9-4219-44e3-b659-6c47a5a6ee42","Type":"ContainerStarted","Data":"0ceaac0197d8b7d8deb7afe4a85ee648c86107c05a5c84f2791b4059e45ad002"} Dec 01 17:14:23 crc kubenswrapper[4810]: I1201 17:14:23.942244 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"03fe61c9-4219-44e3-b659-6c47a5a6ee42","Type":"ContainerStarted","Data":"11ad4ca78b4250d58b04ac5331344da7d73ff7dfe884d53a41fab27b37c7bf3a"} Dec 01 17:14:23 crc kubenswrapper[4810]: I1201 17:14:23.944209 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f44e8294-c07f-4c82-8957-82b096b3d87b","Type":"ContainerStarted","Data":"7dbb794762fa2538d0bff822361e36444486154b8bc967eb2090645e445eceb0"} Dec 01 17:14:23 crc kubenswrapper[4810]: I1201 17:14:23.944243 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f44e8294-c07f-4c82-8957-82b096b3d87b","Type":"ContainerStarted","Data":"c8d719257d2e9ebdb8cf08865d973003e879d260366c30b19bb9165653032426"} Dec 01 17:14:23 crc kubenswrapper[4810]: I1201 17:14:23.944257 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f44e8294-c07f-4c82-8957-82b096b3d87b","Type":"ContainerStarted","Data":"8d4d8f2cbcb61d4d1bc071138797d2267e6874a9bb97bf8af1033e30388473ad"} Dec 01 17:14:23 crc kubenswrapper[4810]: I1201 17:14:23.973865 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.973844305 podStartE2EDuration="1.973844305s" podCreationTimestamp="2025-12-01 17:14:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:14:23.963500587 +0000 UTC m=+9629.727010190" watchObservedRunningTime="2025-12-01 17:14:23.973844305 +0000 UTC m=+9629.737353908" Dec 01 17:14:24 crc kubenswrapper[4810]: I1201 17:14:24.000980 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.000963564 podStartE2EDuration="2.000963564s" podCreationTimestamp="2025-12-01 17:14:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:14:23.989211928 +0000 UTC m=+9629.752721531" watchObservedRunningTime="2025-12-01 17:14:24.000963564 +0000 UTC m=+9629.764473167" Dec 01 17:14:24 crc kubenswrapper[4810]: I1201 17:14:24.231455 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 17:14:26 crc kubenswrapper[4810]: I1201 17:14:26.492093 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:14:26 crc kubenswrapper[4810]: E1201 17:14:26.492762 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:14:27 crc kubenswrapper[4810]: I1201 17:14:27.704538 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 17:14:27 crc kubenswrapper[4810]: I1201 17:14:27.705068 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 17:14:29 crc kubenswrapper[4810]: I1201 17:14:29.231264 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 17:14:29 crc kubenswrapper[4810]: I1201 17:14:29.267298 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 17:14:30 crc kubenswrapper[4810]: I1201 17:14:30.040376 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 17:14:32 crc kubenswrapper[4810]: I1201 17:14:32.018637 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 01 17:14:32 crc kubenswrapper[4810]: I1201 17:14:32.697665 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 17:14:32 crc kubenswrapper[4810]: I1201 17:14:32.697712 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 17:14:32 crc kubenswrapper[4810]: I1201 17:14:32.705212 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 17:14:32 crc kubenswrapper[4810]: I1201 17:14:32.705257 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 17:14:33 crc kubenswrapper[4810]: I1201 17:14:33.715755 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f44e8294-c07f-4c82-8957-82b096b3d87b" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 17:14:33 crc kubenswrapper[4810]: I1201 17:14:33.715755 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f44e8294-c07f-4c82-8957-82b096b3d87b" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 17:14:33 crc kubenswrapper[4810]: I1201 17:14:33.728634 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="03fe61c9-4219-44e3-b659-6c47a5a6ee42" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 17:14:33 crc kubenswrapper[4810]: I1201 17:14:33.728635 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="03fe61c9-4219-44e3-b659-6c47a5a6ee42" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 17:14:37 crc kubenswrapper[4810]: I1201 17:14:37.491804 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:14:37 crc kubenswrapper[4810]: E1201 17:14:37.492741 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:14:42 crc kubenswrapper[4810]: I1201 17:14:42.708308 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 17:14:42 crc kubenswrapper[4810]: I1201 17:14:42.711039 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 17:14:42 crc kubenswrapper[4810]: I1201 17:14:42.711522 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 17:14:42 crc kubenswrapper[4810]: I1201 17:14:42.714415 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 17:14:42 crc kubenswrapper[4810]: I1201 17:14:42.714957 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 17:14:42 crc kubenswrapper[4810]: I1201 17:14:42.720497 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 17:14:42 crc kubenswrapper[4810]: I1201 17:14:42.721747 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 17:14:43 crc kubenswrapper[4810]: I1201 17:14:43.220242 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 17:14:43 crc kubenswrapper[4810]: I1201 17:14:43.225532 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 17:14:43 crc kubenswrapper[4810]: I1201 17:14:43.234062 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 17:14:51 crc kubenswrapper[4810]: I1201 17:14:51.491931 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:14:51 crc kubenswrapper[4810]: E1201 17:14:51.493111 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:15:00 crc kubenswrapper[4810]: I1201 17:15:00.160791 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm"] Dec 01 17:15:00 crc kubenswrapper[4810]: I1201 17:15:00.162734 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm" Dec 01 17:15:00 crc kubenswrapper[4810]: I1201 17:15:00.169993 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 17:15:00 crc kubenswrapper[4810]: I1201 17:15:00.170280 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 17:15:00 crc kubenswrapper[4810]: I1201 17:15:00.185041 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm"] Dec 01 17:15:00 crc kubenswrapper[4810]: I1201 17:15:00.274590 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9c08bef-52bd-45b2-b69e-ebf76cb5d32f-config-volume\") pod \"collect-profiles-29410155-dnjvm\" (UID: \"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm" Dec 01 17:15:00 crc kubenswrapper[4810]: I1201 17:15:00.274733 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9c08bef-52bd-45b2-b69e-ebf76cb5d32f-secret-volume\") pod \"collect-profiles-29410155-dnjvm\" (UID: \"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm" Dec 01 17:15:00 crc kubenswrapper[4810]: I1201 17:15:00.276047 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzkwv\" (UniqueName: \"kubernetes.io/projected/e9c08bef-52bd-45b2-b69e-ebf76cb5d32f-kube-api-access-qzkwv\") pod \"collect-profiles-29410155-dnjvm\" (UID: \"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm" Dec 01 17:15:00 crc kubenswrapper[4810]: I1201 17:15:00.377641 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9c08bef-52bd-45b2-b69e-ebf76cb5d32f-config-volume\") pod \"collect-profiles-29410155-dnjvm\" (UID: \"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm" Dec 01 17:15:00 crc kubenswrapper[4810]: I1201 17:15:00.377778 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9c08bef-52bd-45b2-b69e-ebf76cb5d32f-secret-volume\") pod \"collect-profiles-29410155-dnjvm\" (UID: \"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm" Dec 01 17:15:00 crc kubenswrapper[4810]: I1201 17:15:00.377900 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzkwv\" (UniqueName: \"kubernetes.io/projected/e9c08bef-52bd-45b2-b69e-ebf76cb5d32f-kube-api-access-qzkwv\") pod \"collect-profiles-29410155-dnjvm\" (UID: \"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm" Dec 01 17:15:00 crc kubenswrapper[4810]: I1201 17:15:00.379039 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9c08bef-52bd-45b2-b69e-ebf76cb5d32f-config-volume\") pod \"collect-profiles-29410155-dnjvm\" (UID: \"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm" Dec 01 17:15:00 crc kubenswrapper[4810]: I1201 17:15:00.391500 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9c08bef-52bd-45b2-b69e-ebf76cb5d32f-secret-volume\") pod \"collect-profiles-29410155-dnjvm\" (UID: \"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm" Dec 01 17:15:00 crc kubenswrapper[4810]: I1201 17:15:00.398024 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzkwv\" (UniqueName: \"kubernetes.io/projected/e9c08bef-52bd-45b2-b69e-ebf76cb5d32f-kube-api-access-qzkwv\") pod \"collect-profiles-29410155-dnjvm\" (UID: \"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm" Dec 01 17:15:00 crc kubenswrapper[4810]: I1201 17:15:00.501386 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm" Dec 01 17:15:01 crc kubenswrapper[4810]: I1201 17:15:01.116858 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm"] Dec 01 17:15:01 crc kubenswrapper[4810]: I1201 17:15:01.450804 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm" event={"ID":"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f","Type":"ContainerStarted","Data":"6eb07abffdec0de77a12c1ff9c7aa2551ce5b26c48ba67319afcdfe8cc529d48"} Dec 01 17:15:01 crc kubenswrapper[4810]: I1201 17:15:01.450851 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm" event={"ID":"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f","Type":"ContainerStarted","Data":"075f13e5a1a0d94e7ec72c9580506c34ea5d053d5a10c68b3fe937c1b0bb16b1"} Dec 01 17:15:01 crc kubenswrapper[4810]: I1201 17:15:01.467793 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm" podStartSLOduration=1.467778983 podStartE2EDuration="1.467778983s" podCreationTimestamp="2025-12-01 17:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 17:15:01.464932206 +0000 UTC m=+9667.228441799" watchObservedRunningTime="2025-12-01 17:15:01.467778983 +0000 UTC m=+9667.231288576" Dec 01 17:15:02 crc kubenswrapper[4810]: I1201 17:15:02.463768 4810 generic.go:334] "Generic (PLEG): container finished" podID="e9c08bef-52bd-45b2-b69e-ebf76cb5d32f" containerID="6eb07abffdec0de77a12c1ff9c7aa2551ce5b26c48ba67319afcdfe8cc529d48" exitCode=0 Dec 01 17:15:02 crc kubenswrapper[4810]: I1201 17:15:02.463811 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm" event={"ID":"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f","Type":"ContainerDied","Data":"6eb07abffdec0de77a12c1ff9c7aa2551ce5b26c48ba67319afcdfe8cc529d48"} Dec 01 17:15:03 crc kubenswrapper[4810]: I1201 17:15:03.892574 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm" Dec 01 17:15:03 crc kubenswrapper[4810]: I1201 17:15:03.975328 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9c08bef-52bd-45b2-b69e-ebf76cb5d32f-config-volume\") pod \"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f\" (UID: \"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f\") " Dec 01 17:15:03 crc kubenswrapper[4810]: I1201 17:15:03.976558 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9c08bef-52bd-45b2-b69e-ebf76cb5d32f-secret-volume\") pod \"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f\" (UID: \"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f\") " Dec 01 17:15:03 crc kubenswrapper[4810]: I1201 17:15:03.976685 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzkwv\" (UniqueName: \"kubernetes.io/projected/e9c08bef-52bd-45b2-b69e-ebf76cb5d32f-kube-api-access-qzkwv\") pod \"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f\" (UID: \"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f\") " Dec 01 17:15:03 crc kubenswrapper[4810]: I1201 17:15:03.976405 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9c08bef-52bd-45b2-b69e-ebf76cb5d32f-config-volume" (OuterVolumeSpecName: "config-volume") pod "e9c08bef-52bd-45b2-b69e-ebf76cb5d32f" (UID: "e9c08bef-52bd-45b2-b69e-ebf76cb5d32f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:15:03 crc kubenswrapper[4810]: I1201 17:15:03.977402 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9c08bef-52bd-45b2-b69e-ebf76cb5d32f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 17:15:03 crc kubenswrapper[4810]: I1201 17:15:03.983233 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9c08bef-52bd-45b2-b69e-ebf76cb5d32f-kube-api-access-qzkwv" (OuterVolumeSpecName: "kube-api-access-qzkwv") pod "e9c08bef-52bd-45b2-b69e-ebf76cb5d32f" (UID: "e9c08bef-52bd-45b2-b69e-ebf76cb5d32f"). InnerVolumeSpecName "kube-api-access-qzkwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:15:03 crc kubenswrapper[4810]: I1201 17:15:03.983875 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9c08bef-52bd-45b2-b69e-ebf76cb5d32f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e9c08bef-52bd-45b2-b69e-ebf76cb5d32f" (UID: "e9c08bef-52bd-45b2-b69e-ebf76cb5d32f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:15:04 crc kubenswrapper[4810]: I1201 17:15:04.081990 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9c08bef-52bd-45b2-b69e-ebf76cb5d32f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 17:15:04 crc kubenswrapper[4810]: I1201 17:15:04.082035 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzkwv\" (UniqueName: \"kubernetes.io/projected/e9c08bef-52bd-45b2-b69e-ebf76cb5d32f-kube-api-access-qzkwv\") on node \"crc\" DevicePath \"\"" Dec 01 17:15:04 crc kubenswrapper[4810]: I1201 17:15:04.492203 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm" Dec 01 17:15:04 crc kubenswrapper[4810]: I1201 17:15:04.518635 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410155-dnjvm" event={"ID":"e9c08bef-52bd-45b2-b69e-ebf76cb5d32f","Type":"ContainerDied","Data":"075f13e5a1a0d94e7ec72c9580506c34ea5d053d5a10c68b3fe937c1b0bb16b1"} Dec 01 17:15:04 crc kubenswrapper[4810]: I1201 17:15:04.518676 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="075f13e5a1a0d94e7ec72c9580506c34ea5d053d5a10c68b3fe937c1b0bb16b1" Dec 01 17:15:04 crc kubenswrapper[4810]: I1201 17:15:04.552402 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw"] Dec 01 17:15:04 crc kubenswrapper[4810]: I1201 17:15:04.563062 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410110-47pfw"] Dec 01 17:15:05 crc kubenswrapper[4810]: I1201 17:15:05.492533 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:15:05 crc kubenswrapper[4810]: E1201 17:15:05.493012 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:15:06 crc kubenswrapper[4810]: I1201 17:15:06.511891 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5af27677-6662-4102-882b-38d36c928031" path="/var/lib/kubelet/pods/5af27677-6662-4102-882b-38d36c928031/volumes" Dec 01 17:15:18 crc kubenswrapper[4810]: I1201 17:15:18.629973 4810 scope.go:117] "RemoveContainer" containerID="9c2b483d543b590039596544bbe7e243cff494dd2fd6f4d6d63290643ab73bb0" Dec 01 17:15:20 crc kubenswrapper[4810]: I1201 17:15:20.491942 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:15:20 crc kubenswrapper[4810]: E1201 17:15:20.492717 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:15:34 crc kubenswrapper[4810]: I1201 17:15:34.512747 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:15:34 crc kubenswrapper[4810]: E1201 17:15:34.515602 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:15:47 crc kubenswrapper[4810]: I1201 17:15:47.492291 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:15:47 crc kubenswrapper[4810]: E1201 17:15:47.493726 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:16:02 crc kubenswrapper[4810]: I1201 17:16:02.492221 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:16:02 crc kubenswrapper[4810]: E1201 17:16:02.495842 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:16:14 crc kubenswrapper[4810]: I1201 17:16:14.499367 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:16:14 crc kubenswrapper[4810]: E1201 17:16:14.500630 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:16:28 crc kubenswrapper[4810]: I1201 17:16:28.500111 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:16:28 crc kubenswrapper[4810]: E1201 17:16:28.500869 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:16:42 crc kubenswrapper[4810]: I1201 17:16:42.491030 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:16:42 crc kubenswrapper[4810]: E1201 17:16:42.491931 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:16:55 crc kubenswrapper[4810]: I1201 17:16:55.491812 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:16:55 crc kubenswrapper[4810]: E1201 17:16:55.492885 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:17:09 crc kubenswrapper[4810]: I1201 17:17:09.492098 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:17:09 crc kubenswrapper[4810]: E1201 17:17:09.492966 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:17:20 crc kubenswrapper[4810]: I1201 17:17:20.492240 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:17:20 crc kubenswrapper[4810]: E1201 17:17:20.493677 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:17:32 crc kubenswrapper[4810]: I1201 17:17:32.491551 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:17:32 crc kubenswrapper[4810]: E1201 17:17:32.492374 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:17:46 crc kubenswrapper[4810]: I1201 17:17:46.492230 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:17:46 crc kubenswrapper[4810]: E1201 17:17:46.493436 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:17:58 crc kubenswrapper[4810]: I1201 17:17:58.491025 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:17:58 crc kubenswrapper[4810]: E1201 17:17:58.491956 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:18:09 crc kubenswrapper[4810]: I1201 17:18:09.491170 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:18:09 crc kubenswrapper[4810]: E1201 17:18:09.492120 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:18:21 crc kubenswrapper[4810]: I1201 17:18:21.491352 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:18:21 crc kubenswrapper[4810]: E1201 17:18:21.492309 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:18:33 crc kubenswrapper[4810]: I1201 17:18:33.491211 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:18:33 crc kubenswrapper[4810]: E1201 17:18:33.492284 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:18:46 crc kubenswrapper[4810]: I1201 17:18:46.491425 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:18:46 crc kubenswrapper[4810]: E1201 17:18:46.493186 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:19:01 crc kubenswrapper[4810]: I1201 17:19:01.491286 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:19:01 crc kubenswrapper[4810]: E1201 17:19:01.492216 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:19:14 crc kubenswrapper[4810]: I1201 17:19:14.500718 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:19:15 crc kubenswrapper[4810]: I1201 17:19:15.502862 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"ac4b6e55100e42785c4ade7daada4a9f96712f083cb05f6b3c5e6cea2dcf7296"} Dec 01 17:20:55 crc kubenswrapper[4810]: I1201 17:20:55.525956 4810 generic.go:334] "Generic (PLEG): container finished" podID="cd2bd342-3a99-4d38-976b-3f213999f103" containerID="48a8b8c372a009c3c38018c2a3556217a6d13a8614a677422be23703f6aab6fa" exitCode=0 Dec 01 17:20:55 crc kubenswrapper[4810]: I1201 17:20:55.526083 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" event={"ID":"cd2bd342-3a99-4d38-976b-3f213999f103","Type":"ContainerDied","Data":"48a8b8c372a009c3c38018c2a3556217a6d13a8614a677422be23703f6aab6fa"} Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.023020 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.038386 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cell1-compute-config-1\") pod \"cd2bd342-3a99-4d38-976b-3f213999f103\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.044717 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-inventory\") pod \"cd2bd342-3a99-4d38-976b-3f213999f103\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.044805 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cells-global-config-0\") pod \"cd2bd342-3a99-4d38-976b-3f213999f103\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.044888 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cell1-combined-ca-bundle\") pod \"cd2bd342-3a99-4d38-976b-3f213999f103\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.044920 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-ssh-key\") pod \"cd2bd342-3a99-4d38-976b-3f213999f103\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.045013 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-migration-ssh-key-1\") pod \"cd2bd342-3a99-4d38-976b-3f213999f103\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.045089 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-migration-ssh-key-0\") pod \"cd2bd342-3a99-4d38-976b-3f213999f103\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.045159 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vxxl\" (UniqueName: \"kubernetes.io/projected/cd2bd342-3a99-4d38-976b-3f213999f103-kube-api-access-5vxxl\") pod \"cd2bd342-3a99-4d38-976b-3f213999f103\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.045195 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cell1-compute-config-0\") pod \"cd2bd342-3a99-4d38-976b-3f213999f103\" (UID: \"cd2bd342-3a99-4d38-976b-3f213999f103\") " Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.051091 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd2bd342-3a99-4d38-976b-3f213999f103-kube-api-access-5vxxl" (OuterVolumeSpecName: "kube-api-access-5vxxl") pod "cd2bd342-3a99-4d38-976b-3f213999f103" (UID: "cd2bd342-3a99-4d38-976b-3f213999f103"). InnerVolumeSpecName "kube-api-access-5vxxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.148258 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vxxl\" (UniqueName: \"kubernetes.io/projected/cd2bd342-3a99-4d38-976b-3f213999f103-kube-api-access-5vxxl\") on node \"crc\" DevicePath \"\"" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.553412 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" event={"ID":"cd2bd342-3a99-4d38-976b-3f213999f103","Type":"ContainerDied","Data":"41b90709b8dabbccab1fe936cfccacc94f0ae01b94a55c928877b3ed43298f88"} Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.553575 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41b90709b8dabbccab1fe936cfccacc94f0ae01b94a55c928877b3ed43298f88" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.553682 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.776196 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "cd2bd342-3a99-4d38-976b-3f213999f103" (UID: "cd2bd342-3a99-4d38-976b-3f213999f103"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.789666 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cd2bd342-3a99-4d38-976b-3f213999f103" (UID: "cd2bd342-3a99-4d38-976b-3f213999f103"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.790815 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "cd2bd342-3a99-4d38-976b-3f213999f103" (UID: "cd2bd342-3a99-4d38-976b-3f213999f103"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.792088 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-inventory" (OuterVolumeSpecName: "inventory") pod "cd2bd342-3a99-4d38-976b-3f213999f103" (UID: "cd2bd342-3a99-4d38-976b-3f213999f103"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.795590 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "cd2bd342-3a99-4d38-976b-3f213999f103" (UID: "cd2bd342-3a99-4d38-976b-3f213999f103"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.806335 4810 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.806377 4810 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.806399 4810 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.806414 4810 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.806428 4810 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.810595 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "cd2bd342-3a99-4d38-976b-3f213999f103" (UID: "cd2bd342-3a99-4d38-976b-3f213999f103"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.830645 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "cd2bd342-3a99-4d38-976b-3f213999f103" (UID: "cd2bd342-3a99-4d38-976b-3f213999f103"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.841831 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "cd2bd342-3a99-4d38-976b-3f213999f103" (UID: "cd2bd342-3a99-4d38-976b-3f213999f103"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.908916 4810 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.908955 4810 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:20:57 crc kubenswrapper[4810]: I1201 17:20:57.908967 4810 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cd2bd342-3a99-4d38-976b-3f213999f103-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 17:21:32 crc kubenswrapper[4810]: I1201 17:21:32.972728 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:21:32 crc kubenswrapper[4810]: I1201 17:21:32.973229 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:22:02 crc kubenswrapper[4810]: I1201 17:22:02.972367 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:22:02 crc kubenswrapper[4810]: I1201 17:22:02.973449 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:22:32 crc kubenswrapper[4810]: I1201 17:22:32.972500 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:22:32 crc kubenswrapper[4810]: I1201 17:22:32.973038 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:22:32 crc kubenswrapper[4810]: I1201 17:22:32.973084 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 17:22:32 crc kubenswrapper[4810]: I1201 17:22:32.973879 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ac4b6e55100e42785c4ade7daada4a9f96712f083cb05f6b3c5e6cea2dcf7296"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 17:22:32 crc kubenswrapper[4810]: I1201 17:22:32.973926 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://ac4b6e55100e42785c4ade7daada4a9f96712f083cb05f6b3c5e6cea2dcf7296" gracePeriod=600 Dec 01 17:22:33 crc kubenswrapper[4810]: I1201 17:22:33.569770 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="ac4b6e55100e42785c4ade7daada4a9f96712f083cb05f6b3c5e6cea2dcf7296" exitCode=0 Dec 01 17:22:33 crc kubenswrapper[4810]: I1201 17:22:33.569856 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"ac4b6e55100e42785c4ade7daada4a9f96712f083cb05f6b3c5e6cea2dcf7296"} Dec 01 17:22:33 crc kubenswrapper[4810]: I1201 17:22:33.570050 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd"} Dec 01 17:22:33 crc kubenswrapper[4810]: I1201 17:22:33.570138 4810 scope.go:117] "RemoveContainer" containerID="da526cf7c56f46c67de440581426357a649406042cd27ace94a0314cdc5a37fb" Dec 01 17:22:43 crc kubenswrapper[4810]: I1201 17:22:43.995306 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Dec 01 17:22:43 crc kubenswrapper[4810]: I1201 17:22:43.996221 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="f8d6ddae-90bb-43a8-8b28-fb05dd494ba3" containerName="adoption" containerID="cri-o://053a9cd297dff6f3bd2ddc9ee7d252d1f236607a51ac59962c81e9455c5a0f2a" gracePeriod=30 Dec 01 17:23:15 crc kubenswrapper[4810]: I1201 17:23:15.040953 4810 generic.go:334] "Generic (PLEG): container finished" podID="f8d6ddae-90bb-43a8-8b28-fb05dd494ba3" containerID="053a9cd297dff6f3bd2ddc9ee7d252d1f236607a51ac59962c81e9455c5a0f2a" exitCode=137 Dec 01 17:23:15 crc kubenswrapper[4810]: I1201 17:23:15.041519 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"f8d6ddae-90bb-43a8-8b28-fb05dd494ba3","Type":"ContainerDied","Data":"053a9cd297dff6f3bd2ddc9ee7d252d1f236607a51ac59962c81e9455c5a0f2a"} Dec 01 17:23:15 crc kubenswrapper[4810]: I1201 17:23:15.041674 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"f8d6ddae-90bb-43a8-8b28-fb05dd494ba3","Type":"ContainerDied","Data":"cb990b29946c7c4e3745ee064ef5f50627ee656980823b6c5cf70db855f18c3c"} Dec 01 17:23:15 crc kubenswrapper[4810]: I1201 17:23:15.041711 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb990b29946c7c4e3745ee064ef5f50627ee656980823b6c5cf70db855f18c3c" Dec 01 17:23:15 crc kubenswrapper[4810]: I1201 17:23:15.312017 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 01 17:23:15 crc kubenswrapper[4810]: I1201 17:23:15.418441 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-42fa2d68-0280-474c-b021-0e7adb73eaf0\") pod \"f8d6ddae-90bb-43a8-8b28-fb05dd494ba3\" (UID: \"f8d6ddae-90bb-43a8-8b28-fb05dd494ba3\") " Dec 01 17:23:15 crc kubenswrapper[4810]: I1201 17:23:15.418566 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmn6h\" (UniqueName: \"kubernetes.io/projected/f8d6ddae-90bb-43a8-8b28-fb05dd494ba3-kube-api-access-jmn6h\") pod \"f8d6ddae-90bb-43a8-8b28-fb05dd494ba3\" (UID: \"f8d6ddae-90bb-43a8-8b28-fb05dd494ba3\") " Dec 01 17:23:15 crc kubenswrapper[4810]: I1201 17:23:15.431876 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8d6ddae-90bb-43a8-8b28-fb05dd494ba3-kube-api-access-jmn6h" (OuterVolumeSpecName: "kube-api-access-jmn6h") pod "f8d6ddae-90bb-43a8-8b28-fb05dd494ba3" (UID: "f8d6ddae-90bb-43a8-8b28-fb05dd494ba3"). InnerVolumeSpecName "kube-api-access-jmn6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:23:15 crc kubenswrapper[4810]: I1201 17:23:15.453367 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-42fa2d68-0280-474c-b021-0e7adb73eaf0" (OuterVolumeSpecName: "mariadb-data") pod "f8d6ddae-90bb-43a8-8b28-fb05dd494ba3" (UID: "f8d6ddae-90bb-43a8-8b28-fb05dd494ba3"). InnerVolumeSpecName "pvc-42fa2d68-0280-474c-b021-0e7adb73eaf0". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 17:23:15 crc kubenswrapper[4810]: I1201 17:23:15.523097 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-42fa2d68-0280-474c-b021-0e7adb73eaf0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-42fa2d68-0280-474c-b021-0e7adb73eaf0\") on node \"crc\" " Dec 01 17:23:15 crc kubenswrapper[4810]: I1201 17:23:15.537775 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmn6h\" (UniqueName: \"kubernetes.io/projected/f8d6ddae-90bb-43a8-8b28-fb05dd494ba3-kube-api-access-jmn6h\") on node \"crc\" DevicePath \"\"" Dec 01 17:23:15 crc kubenswrapper[4810]: I1201 17:23:15.573638 4810 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 01 17:23:15 crc kubenswrapper[4810]: I1201 17:23:15.573875 4810 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-42fa2d68-0280-474c-b021-0e7adb73eaf0" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-42fa2d68-0280-474c-b021-0e7adb73eaf0") on node "crc" Dec 01 17:23:15 crc kubenswrapper[4810]: I1201 17:23:15.640163 4810 reconciler_common.go:293] "Volume detached for volume \"pvc-42fa2d68-0280-474c-b021-0e7adb73eaf0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-42fa2d68-0280-474c-b021-0e7adb73eaf0\") on node \"crc\" DevicePath \"\"" Dec 01 17:23:16 crc kubenswrapper[4810]: I1201 17:23:16.053836 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 01 17:23:16 crc kubenswrapper[4810]: I1201 17:23:16.104464 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Dec 01 17:23:16 crc kubenswrapper[4810]: I1201 17:23:16.116622 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Dec 01 17:23:16 crc kubenswrapper[4810]: I1201 17:23:16.504944 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8d6ddae-90bb-43a8-8b28-fb05dd494ba3" path="/var/lib/kubelet/pods/f8d6ddae-90bb-43a8-8b28-fb05dd494ba3/volumes" Dec 01 17:23:16 crc kubenswrapper[4810]: I1201 17:23:16.815084 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Dec 01 17:23:16 crc kubenswrapper[4810]: I1201 17:23:16.815512 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="00f95a04-d522-4fd8-ab09-ed6dd151d2c3" containerName="adoption" containerID="cri-o://bf278a2f671ea4e4f6c001b1f357946db1ebd7816507ed7aae899253c8dac572" gracePeriod=30 Dec 01 17:23:18 crc kubenswrapper[4810]: I1201 17:23:18.869236 4810 scope.go:117] "RemoveContainer" containerID="053a9cd297dff6f3bd2ddc9ee7d252d1f236607a51ac59962c81e9455c5a0f2a" Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.355755 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.442564 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/00f95a04-d522-4fd8-ab09-ed6dd151d2c3-ovn-data-cert\") pod \"00f95a04-d522-4fd8-ab09-ed6dd151d2c3\" (UID: \"00f95a04-d522-4fd8-ab09-ed6dd151d2c3\") " Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.443586 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-47574dfa-681b-43d9-a251-554a92055dd4\") pod \"00f95a04-d522-4fd8-ab09-ed6dd151d2c3\" (UID: \"00f95a04-d522-4fd8-ab09-ed6dd151d2c3\") " Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.443675 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vh2m\" (UniqueName: \"kubernetes.io/projected/00f95a04-d522-4fd8-ab09-ed6dd151d2c3-kube-api-access-7vh2m\") pod \"00f95a04-d522-4fd8-ab09-ed6dd151d2c3\" (UID: \"00f95a04-d522-4fd8-ab09-ed6dd151d2c3\") " Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.444388 4810 generic.go:334] "Generic (PLEG): container finished" podID="00f95a04-d522-4fd8-ab09-ed6dd151d2c3" containerID="bf278a2f671ea4e4f6c001b1f357946db1ebd7816507ed7aae899253c8dac572" exitCode=137 Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.444428 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"00f95a04-d522-4fd8-ab09-ed6dd151d2c3","Type":"ContainerDied","Data":"bf278a2f671ea4e4f6c001b1f357946db1ebd7816507ed7aae899253c8dac572"} Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.444458 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"00f95a04-d522-4fd8-ab09-ed6dd151d2c3","Type":"ContainerDied","Data":"e95a4f9ea8e60dda9356ea03e859f39102c6c65d31447f7b73a5af43b4bc7c32"} Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.444515 4810 scope.go:117] "RemoveContainer" containerID="bf278a2f671ea4e4f6c001b1f357946db1ebd7816507ed7aae899253c8dac572" Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.444514 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.458560 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00f95a04-d522-4fd8-ab09-ed6dd151d2c3-kube-api-access-7vh2m" (OuterVolumeSpecName: "kube-api-access-7vh2m") pod "00f95a04-d522-4fd8-ab09-ed6dd151d2c3" (UID: "00f95a04-d522-4fd8-ab09-ed6dd151d2c3"). InnerVolumeSpecName "kube-api-access-7vh2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.458566 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00f95a04-d522-4fd8-ab09-ed6dd151d2c3-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "00f95a04-d522-4fd8-ab09-ed6dd151d2c3" (UID: "00f95a04-d522-4fd8-ab09-ed6dd151d2c3"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.465571 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-47574dfa-681b-43d9-a251-554a92055dd4" (OuterVolumeSpecName: "ovn-data") pod "00f95a04-d522-4fd8-ab09-ed6dd151d2c3" (UID: "00f95a04-d522-4fd8-ab09-ed6dd151d2c3"). InnerVolumeSpecName "pvc-47574dfa-681b-43d9-a251-554a92055dd4". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.547010 4810 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-47574dfa-681b-43d9-a251-554a92055dd4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-47574dfa-681b-43d9-a251-554a92055dd4\") on node \"crc\" " Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.547358 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vh2m\" (UniqueName: \"kubernetes.io/projected/00f95a04-d522-4fd8-ab09-ed6dd151d2c3-kube-api-access-7vh2m\") on node \"crc\" DevicePath \"\"" Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.547500 4810 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/00f95a04-d522-4fd8-ab09-ed6dd151d2c3-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.579134 4810 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.579295 4810 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-47574dfa-681b-43d9-a251-554a92055dd4" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-47574dfa-681b-43d9-a251-554a92055dd4") on node "crc" Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.580862 4810 scope.go:117] "RemoveContainer" containerID="bf278a2f671ea4e4f6c001b1f357946db1ebd7816507ed7aae899253c8dac572" Dec 01 17:23:47 crc kubenswrapper[4810]: E1201 17:23:47.581655 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf278a2f671ea4e4f6c001b1f357946db1ebd7816507ed7aae899253c8dac572\": container with ID starting with bf278a2f671ea4e4f6c001b1f357946db1ebd7816507ed7aae899253c8dac572 not found: ID does not exist" containerID="bf278a2f671ea4e4f6c001b1f357946db1ebd7816507ed7aae899253c8dac572" Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.581689 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf278a2f671ea4e4f6c001b1f357946db1ebd7816507ed7aae899253c8dac572"} err="failed to get container status \"bf278a2f671ea4e4f6c001b1f357946db1ebd7816507ed7aae899253c8dac572\": rpc error: code = NotFound desc = could not find container \"bf278a2f671ea4e4f6c001b1f357946db1ebd7816507ed7aae899253c8dac572\": container with ID starting with bf278a2f671ea4e4f6c001b1f357946db1ebd7816507ed7aae899253c8dac572 not found: ID does not exist" Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.650295 4810 reconciler_common.go:293] "Volume detached for volume \"pvc-47574dfa-681b-43d9-a251-554a92055dd4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-47574dfa-681b-43d9-a251-554a92055dd4\") on node \"crc\" DevicePath \"\"" Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.779399 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Dec 01 17:23:47 crc kubenswrapper[4810]: I1201 17:23:47.789979 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Dec 01 17:23:48 crc kubenswrapper[4810]: I1201 17:23:48.502642 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00f95a04-d522-4fd8-ab09-ed6dd151d2c3" path="/var/lib/kubelet/pods/00f95a04-d522-4fd8-ab09-ed6dd151d2c3/volumes" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.265295 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-24f67"] Dec 01 17:24:54 crc kubenswrapper[4810]: E1201 17:24:54.266497 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8d6ddae-90bb-43a8-8b28-fb05dd494ba3" containerName="adoption" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.266514 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8d6ddae-90bb-43a8-8b28-fb05dd494ba3" containerName="adoption" Dec 01 17:24:54 crc kubenswrapper[4810]: E1201 17:24:54.266537 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd2bd342-3a99-4d38-976b-3f213999f103" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.266545 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd2bd342-3a99-4d38-976b-3f213999f103" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 01 17:24:54 crc kubenswrapper[4810]: E1201 17:24:54.266573 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9c08bef-52bd-45b2-b69e-ebf76cb5d32f" containerName="collect-profiles" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.266579 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9c08bef-52bd-45b2-b69e-ebf76cb5d32f" containerName="collect-profiles" Dec 01 17:24:54 crc kubenswrapper[4810]: E1201 17:24:54.266595 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00f95a04-d522-4fd8-ab09-ed6dd151d2c3" containerName="adoption" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.266602 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="00f95a04-d522-4fd8-ab09-ed6dd151d2c3" containerName="adoption" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.266849 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="00f95a04-d522-4fd8-ab09-ed6dd151d2c3" containerName="adoption" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.267043 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9c08bef-52bd-45b2-b69e-ebf76cb5d32f" containerName="collect-profiles" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.267060 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8d6ddae-90bb-43a8-8b28-fb05dd494ba3" containerName="adoption" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.267089 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd2bd342-3a99-4d38-976b-3f213999f103" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.271139 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24f67" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.284334 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-24f67"] Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.465969 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1249c0fe-6560-408d-a7f4-3fced6c32e55-catalog-content\") pod \"certified-operators-24f67\" (UID: \"1249c0fe-6560-408d-a7f4-3fced6c32e55\") " pod="openshift-marketplace/certified-operators-24f67" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.466090 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1249c0fe-6560-408d-a7f4-3fced6c32e55-utilities\") pod \"certified-operators-24f67\" (UID: \"1249c0fe-6560-408d-a7f4-3fced6c32e55\") " pod="openshift-marketplace/certified-operators-24f67" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.466188 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b8k9\" (UniqueName: \"kubernetes.io/projected/1249c0fe-6560-408d-a7f4-3fced6c32e55-kube-api-access-7b8k9\") pod \"certified-operators-24f67\" (UID: \"1249c0fe-6560-408d-a7f4-3fced6c32e55\") " pod="openshift-marketplace/certified-operators-24f67" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.469866 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xdmnf/must-gather-tcrzd"] Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.472231 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xdmnf/must-gather-tcrzd" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.474213 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xdmnf"/"kube-root-ca.crt" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.475800 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xdmnf"/"openshift-service-ca.crt" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.511473 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xdmnf/must-gather-tcrzd"] Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.573659 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87jvt\" (UniqueName: \"kubernetes.io/projected/58292b71-6004-474e-aacf-6b8c7349ab13-kube-api-access-87jvt\") pod \"must-gather-tcrzd\" (UID: \"58292b71-6004-474e-aacf-6b8c7349ab13\") " pod="openshift-must-gather-xdmnf/must-gather-tcrzd" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.573727 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1249c0fe-6560-408d-a7f4-3fced6c32e55-catalog-content\") pod \"certified-operators-24f67\" (UID: \"1249c0fe-6560-408d-a7f4-3fced6c32e55\") " pod="openshift-marketplace/certified-operators-24f67" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.573778 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1249c0fe-6560-408d-a7f4-3fced6c32e55-utilities\") pod \"certified-operators-24f67\" (UID: \"1249c0fe-6560-408d-a7f4-3fced6c32e55\") " pod="openshift-marketplace/certified-operators-24f67" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.573829 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b8k9\" (UniqueName: \"kubernetes.io/projected/1249c0fe-6560-408d-a7f4-3fced6c32e55-kube-api-access-7b8k9\") pod \"certified-operators-24f67\" (UID: \"1249c0fe-6560-408d-a7f4-3fced6c32e55\") " pod="openshift-marketplace/certified-operators-24f67" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.573861 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/58292b71-6004-474e-aacf-6b8c7349ab13-must-gather-output\") pod \"must-gather-tcrzd\" (UID: \"58292b71-6004-474e-aacf-6b8c7349ab13\") " pod="openshift-must-gather-xdmnf/must-gather-tcrzd" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.574243 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1249c0fe-6560-408d-a7f4-3fced6c32e55-catalog-content\") pod \"certified-operators-24f67\" (UID: \"1249c0fe-6560-408d-a7f4-3fced6c32e55\") " pod="openshift-marketplace/certified-operators-24f67" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.574556 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1249c0fe-6560-408d-a7f4-3fced6c32e55-utilities\") pod \"certified-operators-24f67\" (UID: \"1249c0fe-6560-408d-a7f4-3fced6c32e55\") " pod="openshift-marketplace/certified-operators-24f67" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.676041 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/58292b71-6004-474e-aacf-6b8c7349ab13-must-gather-output\") pod \"must-gather-tcrzd\" (UID: \"58292b71-6004-474e-aacf-6b8c7349ab13\") " pod="openshift-must-gather-xdmnf/must-gather-tcrzd" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.676519 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/58292b71-6004-474e-aacf-6b8c7349ab13-must-gather-output\") pod \"must-gather-tcrzd\" (UID: \"58292b71-6004-474e-aacf-6b8c7349ab13\") " pod="openshift-must-gather-xdmnf/must-gather-tcrzd" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.676831 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87jvt\" (UniqueName: \"kubernetes.io/projected/58292b71-6004-474e-aacf-6b8c7349ab13-kube-api-access-87jvt\") pod \"must-gather-tcrzd\" (UID: \"58292b71-6004-474e-aacf-6b8c7349ab13\") " pod="openshift-must-gather-xdmnf/must-gather-tcrzd" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.690065 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xdmnf"/"kube-root-ca.crt" Dec 01 17:24:54 crc kubenswrapper[4810]: I1201 17:24:54.702021 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xdmnf"/"openshift-service-ca.crt" Dec 01 17:24:55 crc kubenswrapper[4810]: I1201 17:24:55.051554 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87jvt\" (UniqueName: \"kubernetes.io/projected/58292b71-6004-474e-aacf-6b8c7349ab13-kube-api-access-87jvt\") pod \"must-gather-tcrzd\" (UID: \"58292b71-6004-474e-aacf-6b8c7349ab13\") " pod="openshift-must-gather-xdmnf/must-gather-tcrzd" Dec 01 17:24:55 crc kubenswrapper[4810]: I1201 17:24:55.051602 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b8k9\" (UniqueName: \"kubernetes.io/projected/1249c0fe-6560-408d-a7f4-3fced6c32e55-kube-api-access-7b8k9\") pod \"certified-operators-24f67\" (UID: \"1249c0fe-6560-408d-a7f4-3fced6c32e55\") " pod="openshift-marketplace/certified-operators-24f67" Dec 01 17:24:55 crc kubenswrapper[4810]: I1201 17:24:55.100149 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xdmnf/must-gather-tcrzd" Dec 01 17:24:55 crc kubenswrapper[4810]: I1201 17:24:55.199604 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24f67" Dec 01 17:24:55 crc kubenswrapper[4810]: I1201 17:24:55.812783 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xdmnf/must-gather-tcrzd"] Dec 01 17:24:55 crc kubenswrapper[4810]: I1201 17:24:55.828641 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 17:24:55 crc kubenswrapper[4810]: W1201 17:24:55.901304 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1249c0fe_6560_408d_a7f4_3fced6c32e55.slice/crio-e0a1cf1223a02775586ff951aa8d928898b97eb99fe85209d902c9895e0ed86b WatchSource:0}: Error finding container e0a1cf1223a02775586ff951aa8d928898b97eb99fe85209d902c9895e0ed86b: Status 404 returned error can't find the container with id e0a1cf1223a02775586ff951aa8d928898b97eb99fe85209d902c9895e0ed86b Dec 01 17:24:55 crc kubenswrapper[4810]: I1201 17:24:55.906792 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-24f67"] Dec 01 17:24:56 crc kubenswrapper[4810]: E1201 17:24:56.222374 4810 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1249c0fe_6560_408d_a7f4_3fced6c32e55.slice/crio-conmon-a6cc460a100afeb3c6bd7f11c454e7c746eab74827294b0b6ed622836f611f51.scope\": RecentStats: unable to find data in memory cache]" Dec 01 17:24:56 crc kubenswrapper[4810]: I1201 17:24:56.222948 4810 generic.go:334] "Generic (PLEG): container finished" podID="1249c0fe-6560-408d-a7f4-3fced6c32e55" containerID="a6cc460a100afeb3c6bd7f11c454e7c746eab74827294b0b6ed622836f611f51" exitCode=0 Dec 01 17:24:56 crc kubenswrapper[4810]: I1201 17:24:56.224036 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24f67" event={"ID":"1249c0fe-6560-408d-a7f4-3fced6c32e55","Type":"ContainerDied","Data":"a6cc460a100afeb3c6bd7f11c454e7c746eab74827294b0b6ed622836f611f51"} Dec 01 17:24:56 crc kubenswrapper[4810]: I1201 17:24:56.224066 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24f67" event={"ID":"1249c0fe-6560-408d-a7f4-3fced6c32e55","Type":"ContainerStarted","Data":"e0a1cf1223a02775586ff951aa8d928898b97eb99fe85209d902c9895e0ed86b"} Dec 01 17:24:56 crc kubenswrapper[4810]: I1201 17:24:56.225353 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xdmnf/must-gather-tcrzd" event={"ID":"58292b71-6004-474e-aacf-6b8c7349ab13","Type":"ContainerStarted","Data":"8d56fb36b1e641e3ce8484ae761267b34b71dca0509a32b473cb848a9bf65f26"} Dec 01 17:24:58 crc kubenswrapper[4810]: I1201 17:24:58.250310 4810 generic.go:334] "Generic (PLEG): container finished" podID="1249c0fe-6560-408d-a7f4-3fced6c32e55" containerID="f790563d337418401c18b9af8ba84736b327a0619f348035df66c29592fea2d3" exitCode=0 Dec 01 17:24:58 crc kubenswrapper[4810]: I1201 17:24:58.250545 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24f67" event={"ID":"1249c0fe-6560-408d-a7f4-3fced6c32e55","Type":"ContainerDied","Data":"f790563d337418401c18b9af8ba84736b327a0619f348035df66c29592fea2d3"} Dec 01 17:25:02 crc kubenswrapper[4810]: I1201 17:25:02.299507 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24f67" event={"ID":"1249c0fe-6560-408d-a7f4-3fced6c32e55","Type":"ContainerStarted","Data":"a3c07d5732bc6ad91f5eb2358c9e01bf934ca773129004b7d562e13b26c2a7b7"} Dec 01 17:25:02 crc kubenswrapper[4810]: I1201 17:25:02.302101 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xdmnf/must-gather-tcrzd" event={"ID":"58292b71-6004-474e-aacf-6b8c7349ab13","Type":"ContainerStarted","Data":"d1e8f6aa8ffa77af42e3972468d9704bc6945998d53e8e8b15871ea23457fad2"} Dec 01 17:25:02 crc kubenswrapper[4810]: I1201 17:25:02.302135 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xdmnf/must-gather-tcrzd" event={"ID":"58292b71-6004-474e-aacf-6b8c7349ab13","Type":"ContainerStarted","Data":"dbdccde095f3fc50cffdc9f2bf7cf984033ebde7ce9452363bf8db0845f2c38e"} Dec 01 17:25:02 crc kubenswrapper[4810]: I1201 17:25:02.326985 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-24f67" podStartSLOduration=3.606420544 podStartE2EDuration="8.32696656s" podCreationTimestamp="2025-12-01 17:24:54 +0000 UTC" firstStartedPulling="2025-12-01 17:24:56.224862653 +0000 UTC m=+10261.988372256" lastFinishedPulling="2025-12-01 17:25:00.945408659 +0000 UTC m=+10266.708918272" observedRunningTime="2025-12-01 17:25:02.317694191 +0000 UTC m=+10268.081203834" watchObservedRunningTime="2025-12-01 17:25:02.32696656 +0000 UTC m=+10268.090476163" Dec 01 17:25:02 crc kubenswrapper[4810]: I1201 17:25:02.341296 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xdmnf/must-gather-tcrzd" podStartSLOduration=3.164160705 podStartE2EDuration="8.341278405s" podCreationTimestamp="2025-12-01 17:24:54 +0000 UTC" firstStartedPulling="2025-12-01 17:24:55.82859881 +0000 UTC m=+10261.592108413" lastFinishedPulling="2025-12-01 17:25:01.0057165 +0000 UTC m=+10266.769226113" observedRunningTime="2025-12-01 17:25:02.335368786 +0000 UTC m=+10268.098878399" watchObservedRunningTime="2025-12-01 17:25:02.341278405 +0000 UTC m=+10268.104788008" Dec 01 17:25:02 crc kubenswrapper[4810]: I1201 17:25:02.972562 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:25:02 crc kubenswrapper[4810]: I1201 17:25:02.972643 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:25:05 crc kubenswrapper[4810]: I1201 17:25:05.156113 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xdmnf/crc-debug-g9lw9"] Dec 01 17:25:05 crc kubenswrapper[4810]: I1201 17:25:05.158512 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xdmnf/crc-debug-g9lw9" Dec 01 17:25:05 crc kubenswrapper[4810]: I1201 17:25:05.164929 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-xdmnf"/"default-dockercfg-sh947" Dec 01 17:25:05 crc kubenswrapper[4810]: I1201 17:25:05.200619 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-24f67" Dec 01 17:25:05 crc kubenswrapper[4810]: I1201 17:25:05.201035 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-24f67" Dec 01 17:25:05 crc kubenswrapper[4810]: I1201 17:25:05.326981 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c9e882bf-3bf3-4317-a24c-a67fc7b4061e-host\") pod \"crc-debug-g9lw9\" (UID: \"c9e882bf-3bf3-4317-a24c-a67fc7b4061e\") " pod="openshift-must-gather-xdmnf/crc-debug-g9lw9" Dec 01 17:25:05 crc kubenswrapper[4810]: I1201 17:25:05.327203 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fqf8\" (UniqueName: \"kubernetes.io/projected/c9e882bf-3bf3-4317-a24c-a67fc7b4061e-kube-api-access-2fqf8\") pod \"crc-debug-g9lw9\" (UID: \"c9e882bf-3bf3-4317-a24c-a67fc7b4061e\") " pod="openshift-must-gather-xdmnf/crc-debug-g9lw9" Dec 01 17:25:05 crc kubenswrapper[4810]: I1201 17:25:05.466036 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fqf8\" (UniqueName: \"kubernetes.io/projected/c9e882bf-3bf3-4317-a24c-a67fc7b4061e-kube-api-access-2fqf8\") pod \"crc-debug-g9lw9\" (UID: \"c9e882bf-3bf3-4317-a24c-a67fc7b4061e\") " pod="openshift-must-gather-xdmnf/crc-debug-g9lw9" Dec 01 17:25:05 crc kubenswrapper[4810]: I1201 17:25:05.467073 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c9e882bf-3bf3-4317-a24c-a67fc7b4061e-host\") pod \"crc-debug-g9lw9\" (UID: \"c9e882bf-3bf3-4317-a24c-a67fc7b4061e\") " pod="openshift-must-gather-xdmnf/crc-debug-g9lw9" Dec 01 17:25:05 crc kubenswrapper[4810]: I1201 17:25:05.467218 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c9e882bf-3bf3-4317-a24c-a67fc7b4061e-host\") pod \"crc-debug-g9lw9\" (UID: \"c9e882bf-3bf3-4317-a24c-a67fc7b4061e\") " pod="openshift-must-gather-xdmnf/crc-debug-g9lw9" Dec 01 17:25:05 crc kubenswrapper[4810]: I1201 17:25:05.484216 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fqf8\" (UniqueName: \"kubernetes.io/projected/c9e882bf-3bf3-4317-a24c-a67fc7b4061e-kube-api-access-2fqf8\") pod \"crc-debug-g9lw9\" (UID: \"c9e882bf-3bf3-4317-a24c-a67fc7b4061e\") " pod="openshift-must-gather-xdmnf/crc-debug-g9lw9" Dec 01 17:25:05 crc kubenswrapper[4810]: I1201 17:25:05.782502 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xdmnf/crc-debug-g9lw9" Dec 01 17:25:05 crc kubenswrapper[4810]: W1201 17:25:05.821284 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9e882bf_3bf3_4317_a24c_a67fc7b4061e.slice/crio-45f8c3abe27ad59b9648fc34f702b8582cc9c00c217eaa9b6b8dadbba4292e79 WatchSource:0}: Error finding container 45f8c3abe27ad59b9648fc34f702b8582cc9c00c217eaa9b6b8dadbba4292e79: Status 404 returned error can't find the container with id 45f8c3abe27ad59b9648fc34f702b8582cc9c00c217eaa9b6b8dadbba4292e79 Dec 01 17:25:06 crc kubenswrapper[4810]: I1201 17:25:06.259562 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-24f67" podUID="1249c0fe-6560-408d-a7f4-3fced6c32e55" containerName="registry-server" probeResult="failure" output=< Dec 01 17:25:06 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Dec 01 17:25:06 crc kubenswrapper[4810]: > Dec 01 17:25:06 crc kubenswrapper[4810]: I1201 17:25:06.360146 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xdmnf/crc-debug-g9lw9" event={"ID":"c9e882bf-3bf3-4317-a24c-a67fc7b4061e","Type":"ContainerStarted","Data":"45f8c3abe27ad59b9648fc34f702b8582cc9c00c217eaa9b6b8dadbba4292e79"} Dec 01 17:25:10 crc kubenswrapper[4810]: I1201 17:25:10.608777 4810 trace.go:236] Trace[1400559150]: "Calculate volume metrics of ovndbcluster-nb-etc-ovn for pod openstack/ovsdbserver-nb-1" (01-Dec-2025 17:25:09.238) (total time: 1370ms): Dec 01 17:25:10 crc kubenswrapper[4810]: Trace[1400559150]: [1.370745251s] [1.370745251s] END Dec 01 17:25:15 crc kubenswrapper[4810]: I1201 17:25:15.256026 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-24f67" Dec 01 17:25:15 crc kubenswrapper[4810]: I1201 17:25:15.316791 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-24f67" Dec 01 17:25:15 crc kubenswrapper[4810]: I1201 17:25:15.490776 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-24f67"] Dec 01 17:25:16 crc kubenswrapper[4810]: I1201 17:25:16.535859 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-24f67" podUID="1249c0fe-6560-408d-a7f4-3fced6c32e55" containerName="registry-server" containerID="cri-o://a3c07d5732bc6ad91f5eb2358c9e01bf934ca773129004b7d562e13b26c2a7b7" gracePeriod=2 Dec 01 17:25:17 crc kubenswrapper[4810]: I1201 17:25:17.551721 4810 generic.go:334] "Generic (PLEG): container finished" podID="1249c0fe-6560-408d-a7f4-3fced6c32e55" containerID="a3c07d5732bc6ad91f5eb2358c9e01bf934ca773129004b7d562e13b26c2a7b7" exitCode=0 Dec 01 17:25:17 crc kubenswrapper[4810]: I1201 17:25:17.551762 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24f67" event={"ID":"1249c0fe-6560-408d-a7f4-3fced6c32e55","Type":"ContainerDied","Data":"a3c07d5732bc6ad91f5eb2358c9e01bf934ca773129004b7d562e13b26c2a7b7"} Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.338290 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24f67" Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.482466 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1249c0fe-6560-408d-a7f4-3fced6c32e55-utilities\") pod \"1249c0fe-6560-408d-a7f4-3fced6c32e55\" (UID: \"1249c0fe-6560-408d-a7f4-3fced6c32e55\") " Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.482551 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1249c0fe-6560-408d-a7f4-3fced6c32e55-catalog-content\") pod \"1249c0fe-6560-408d-a7f4-3fced6c32e55\" (UID: \"1249c0fe-6560-408d-a7f4-3fced6c32e55\") " Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.482726 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7b8k9\" (UniqueName: \"kubernetes.io/projected/1249c0fe-6560-408d-a7f4-3fced6c32e55-kube-api-access-7b8k9\") pod \"1249c0fe-6560-408d-a7f4-3fced6c32e55\" (UID: \"1249c0fe-6560-408d-a7f4-3fced6c32e55\") " Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.483747 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1249c0fe-6560-408d-a7f4-3fced6c32e55-utilities" (OuterVolumeSpecName: "utilities") pod "1249c0fe-6560-408d-a7f4-3fced6c32e55" (UID: "1249c0fe-6560-408d-a7f4-3fced6c32e55"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.489013 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1249c0fe-6560-408d-a7f4-3fced6c32e55-kube-api-access-7b8k9" (OuterVolumeSpecName: "kube-api-access-7b8k9") pod "1249c0fe-6560-408d-a7f4-3fced6c32e55" (UID: "1249c0fe-6560-408d-a7f4-3fced6c32e55"). InnerVolumeSpecName "kube-api-access-7b8k9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.532605 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1249c0fe-6560-408d-a7f4-3fced6c32e55-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1249c0fe-6560-408d-a7f4-3fced6c32e55" (UID: "1249c0fe-6560-408d-a7f4-3fced6c32e55"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.570078 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-24f67" event={"ID":"1249c0fe-6560-408d-a7f4-3fced6c32e55","Type":"ContainerDied","Data":"e0a1cf1223a02775586ff951aa8d928898b97eb99fe85209d902c9895e0ed86b"} Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.570161 4810 scope.go:117] "RemoveContainer" containerID="a3c07d5732bc6ad91f5eb2358c9e01bf934ca773129004b7d562e13b26c2a7b7" Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.570291 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-24f67" Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.576894 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xdmnf/crc-debug-g9lw9" event={"ID":"c9e882bf-3bf3-4317-a24c-a67fc7b4061e","Type":"ContainerStarted","Data":"0a77b268b170923c7b15850a6967412f781775ecfed08b1fc44f54a6e617a40c"} Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.584775 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7b8k9\" (UniqueName: \"kubernetes.io/projected/1249c0fe-6560-408d-a7f4-3fced6c32e55-kube-api-access-7b8k9\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.584807 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1249c0fe-6560-408d-a7f4-3fced6c32e55-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.584818 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1249c0fe-6560-408d-a7f4-3fced6c32e55-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.592213 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xdmnf/crc-debug-g9lw9" podStartSLOduration=1.496675589 podStartE2EDuration="13.592195022s" podCreationTimestamp="2025-12-01 17:25:05 +0000 UTC" firstStartedPulling="2025-12-01 17:25:05.823302266 +0000 UTC m=+10271.586811869" lastFinishedPulling="2025-12-01 17:25:17.918821699 +0000 UTC m=+10283.682331302" observedRunningTime="2025-12-01 17:25:18.588637707 +0000 UTC m=+10284.352147310" watchObservedRunningTime="2025-12-01 17:25:18.592195022 +0000 UTC m=+10284.355704625" Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.605773 4810 scope.go:117] "RemoveContainer" containerID="f790563d337418401c18b9af8ba84736b327a0619f348035df66c29592fea2d3" Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.622239 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-24f67"] Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.631389 4810 scope.go:117] "RemoveContainer" containerID="a6cc460a100afeb3c6bd7f11c454e7c746eab74827294b0b6ed622836f611f51" Dec 01 17:25:18 crc kubenswrapper[4810]: I1201 17:25:18.634825 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-24f67"] Dec 01 17:25:20 crc kubenswrapper[4810]: I1201 17:25:20.524946 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1249c0fe-6560-408d-a7f4-3fced6c32e55" path="/var/lib/kubelet/pods/1249c0fe-6560-408d-a7f4-3fced6c32e55/volumes" Dec 01 17:25:32 crc kubenswrapper[4810]: I1201 17:25:32.972236 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:25:32 crc kubenswrapper[4810]: I1201 17:25:32.972804 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:25:57 crc kubenswrapper[4810]: I1201 17:25:57.065131 4810 generic.go:334] "Generic (PLEG): container finished" podID="c9e882bf-3bf3-4317-a24c-a67fc7b4061e" containerID="0a77b268b170923c7b15850a6967412f781775ecfed08b1fc44f54a6e617a40c" exitCode=0 Dec 01 17:25:57 crc kubenswrapper[4810]: I1201 17:25:57.065217 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xdmnf/crc-debug-g9lw9" event={"ID":"c9e882bf-3bf3-4317-a24c-a67fc7b4061e","Type":"ContainerDied","Data":"0a77b268b170923c7b15850a6967412f781775ecfed08b1fc44f54a6e617a40c"} Dec 01 17:25:58 crc kubenswrapper[4810]: I1201 17:25:58.207505 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xdmnf/crc-debug-g9lw9" Dec 01 17:25:58 crc kubenswrapper[4810]: I1201 17:25:58.257924 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xdmnf/crc-debug-g9lw9"] Dec 01 17:25:58 crc kubenswrapper[4810]: I1201 17:25:58.266118 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xdmnf/crc-debug-g9lw9"] Dec 01 17:25:58 crc kubenswrapper[4810]: I1201 17:25:58.403681 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c9e882bf-3bf3-4317-a24c-a67fc7b4061e-host\") pod \"c9e882bf-3bf3-4317-a24c-a67fc7b4061e\" (UID: \"c9e882bf-3bf3-4317-a24c-a67fc7b4061e\") " Dec 01 17:25:58 crc kubenswrapper[4810]: I1201 17:25:58.403773 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c9e882bf-3bf3-4317-a24c-a67fc7b4061e-host" (OuterVolumeSpecName: "host") pod "c9e882bf-3bf3-4317-a24c-a67fc7b4061e" (UID: "c9e882bf-3bf3-4317-a24c-a67fc7b4061e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:25:58 crc kubenswrapper[4810]: I1201 17:25:58.404216 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fqf8\" (UniqueName: \"kubernetes.io/projected/c9e882bf-3bf3-4317-a24c-a67fc7b4061e-kube-api-access-2fqf8\") pod \"c9e882bf-3bf3-4317-a24c-a67fc7b4061e\" (UID: \"c9e882bf-3bf3-4317-a24c-a67fc7b4061e\") " Dec 01 17:25:58 crc kubenswrapper[4810]: I1201 17:25:58.404725 4810 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c9e882bf-3bf3-4317-a24c-a67fc7b4061e-host\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:58 crc kubenswrapper[4810]: I1201 17:25:58.412268 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9e882bf-3bf3-4317-a24c-a67fc7b4061e-kube-api-access-2fqf8" (OuterVolumeSpecName: "kube-api-access-2fqf8") pod "c9e882bf-3bf3-4317-a24c-a67fc7b4061e" (UID: "c9e882bf-3bf3-4317-a24c-a67fc7b4061e"). InnerVolumeSpecName "kube-api-access-2fqf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:25:58 crc kubenswrapper[4810]: I1201 17:25:58.502604 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9e882bf-3bf3-4317-a24c-a67fc7b4061e" path="/var/lib/kubelet/pods/c9e882bf-3bf3-4317-a24c-a67fc7b4061e/volumes" Dec 01 17:25:58 crc kubenswrapper[4810]: I1201 17:25:58.506411 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fqf8\" (UniqueName: \"kubernetes.io/projected/c9e882bf-3bf3-4317-a24c-a67fc7b4061e-kube-api-access-2fqf8\") on node \"crc\" DevicePath \"\"" Dec 01 17:25:59 crc kubenswrapper[4810]: I1201 17:25:59.082969 4810 scope.go:117] "RemoveContainer" containerID="0a77b268b170923c7b15850a6967412f781775ecfed08b1fc44f54a6e617a40c" Dec 01 17:25:59 crc kubenswrapper[4810]: I1201 17:25:59.083133 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xdmnf/crc-debug-g9lw9" Dec 01 17:25:59 crc kubenswrapper[4810]: I1201 17:25:59.490851 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xdmnf/crc-debug-rscgb"] Dec 01 17:25:59 crc kubenswrapper[4810]: E1201 17:25:59.491761 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1249c0fe-6560-408d-a7f4-3fced6c32e55" containerName="extract-utilities" Dec 01 17:25:59 crc kubenswrapper[4810]: I1201 17:25:59.491776 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1249c0fe-6560-408d-a7f4-3fced6c32e55" containerName="extract-utilities" Dec 01 17:25:59 crc kubenswrapper[4810]: E1201 17:25:59.491798 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1249c0fe-6560-408d-a7f4-3fced6c32e55" containerName="extract-content" Dec 01 17:25:59 crc kubenswrapper[4810]: I1201 17:25:59.491807 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1249c0fe-6560-408d-a7f4-3fced6c32e55" containerName="extract-content" Dec 01 17:25:59 crc kubenswrapper[4810]: E1201 17:25:59.491828 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1249c0fe-6560-408d-a7f4-3fced6c32e55" containerName="registry-server" Dec 01 17:25:59 crc kubenswrapper[4810]: I1201 17:25:59.491836 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="1249c0fe-6560-408d-a7f4-3fced6c32e55" containerName="registry-server" Dec 01 17:25:59 crc kubenswrapper[4810]: E1201 17:25:59.491849 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e882bf-3bf3-4317-a24c-a67fc7b4061e" containerName="container-00" Dec 01 17:25:59 crc kubenswrapper[4810]: I1201 17:25:59.491856 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e882bf-3bf3-4317-a24c-a67fc7b4061e" containerName="container-00" Dec 01 17:25:59 crc kubenswrapper[4810]: I1201 17:25:59.492131 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="1249c0fe-6560-408d-a7f4-3fced6c32e55" containerName="registry-server" Dec 01 17:25:59 crc kubenswrapper[4810]: I1201 17:25:59.492151 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9e882bf-3bf3-4317-a24c-a67fc7b4061e" containerName="container-00" Dec 01 17:25:59 crc kubenswrapper[4810]: I1201 17:25:59.493153 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xdmnf/crc-debug-rscgb" Dec 01 17:25:59 crc kubenswrapper[4810]: I1201 17:25:59.494969 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-xdmnf"/"default-dockercfg-sh947" Dec 01 17:25:59 crc kubenswrapper[4810]: I1201 17:25:59.628999 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckfdx\" (UniqueName: \"kubernetes.io/projected/28ede625-26f9-4802-8d47-0b3cfc0899ba-kube-api-access-ckfdx\") pod \"crc-debug-rscgb\" (UID: \"28ede625-26f9-4802-8d47-0b3cfc0899ba\") " pod="openshift-must-gather-xdmnf/crc-debug-rscgb" Dec 01 17:25:59 crc kubenswrapper[4810]: I1201 17:25:59.629175 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28ede625-26f9-4802-8d47-0b3cfc0899ba-host\") pod \"crc-debug-rscgb\" (UID: \"28ede625-26f9-4802-8d47-0b3cfc0899ba\") " pod="openshift-must-gather-xdmnf/crc-debug-rscgb" Dec 01 17:25:59 crc kubenswrapper[4810]: I1201 17:25:59.731390 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckfdx\" (UniqueName: \"kubernetes.io/projected/28ede625-26f9-4802-8d47-0b3cfc0899ba-kube-api-access-ckfdx\") pod \"crc-debug-rscgb\" (UID: \"28ede625-26f9-4802-8d47-0b3cfc0899ba\") " pod="openshift-must-gather-xdmnf/crc-debug-rscgb" Dec 01 17:25:59 crc kubenswrapper[4810]: I1201 17:25:59.731595 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28ede625-26f9-4802-8d47-0b3cfc0899ba-host\") pod \"crc-debug-rscgb\" (UID: \"28ede625-26f9-4802-8d47-0b3cfc0899ba\") " pod="openshift-must-gather-xdmnf/crc-debug-rscgb" Dec 01 17:25:59 crc kubenswrapper[4810]: I1201 17:25:59.731724 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28ede625-26f9-4802-8d47-0b3cfc0899ba-host\") pod \"crc-debug-rscgb\" (UID: \"28ede625-26f9-4802-8d47-0b3cfc0899ba\") " pod="openshift-must-gather-xdmnf/crc-debug-rscgb" Dec 01 17:25:59 crc kubenswrapper[4810]: I1201 17:25:59.758799 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckfdx\" (UniqueName: \"kubernetes.io/projected/28ede625-26f9-4802-8d47-0b3cfc0899ba-kube-api-access-ckfdx\") pod \"crc-debug-rscgb\" (UID: \"28ede625-26f9-4802-8d47-0b3cfc0899ba\") " pod="openshift-must-gather-xdmnf/crc-debug-rscgb" Dec 01 17:25:59 crc kubenswrapper[4810]: I1201 17:25:59.814450 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xdmnf/crc-debug-rscgb" Dec 01 17:25:59 crc kubenswrapper[4810]: W1201 17:25:59.866977 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28ede625_26f9_4802_8d47_0b3cfc0899ba.slice/crio-f3ad778378486edae37fbbaecf95ed996bd9b1c8d8f878dba18b17cb4e924408 WatchSource:0}: Error finding container f3ad778378486edae37fbbaecf95ed996bd9b1c8d8f878dba18b17cb4e924408: Status 404 returned error can't find the container with id f3ad778378486edae37fbbaecf95ed996bd9b1c8d8f878dba18b17cb4e924408 Dec 01 17:26:00 crc kubenswrapper[4810]: I1201 17:26:00.094918 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xdmnf/crc-debug-rscgb" event={"ID":"28ede625-26f9-4802-8d47-0b3cfc0899ba","Type":"ContainerStarted","Data":"f3ad778378486edae37fbbaecf95ed996bd9b1c8d8f878dba18b17cb4e924408"} Dec 01 17:26:01 crc kubenswrapper[4810]: I1201 17:26:01.113730 4810 generic.go:334] "Generic (PLEG): container finished" podID="28ede625-26f9-4802-8d47-0b3cfc0899ba" containerID="ba322f74151ed15529d2e68dff0e68ddf9ccc5d53d9aa7c7b65dedc56abf6a37" exitCode=0 Dec 01 17:26:01 crc kubenswrapper[4810]: I1201 17:26:01.114111 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xdmnf/crc-debug-rscgb" event={"ID":"28ede625-26f9-4802-8d47-0b3cfc0899ba","Type":"ContainerDied","Data":"ba322f74151ed15529d2e68dff0e68ddf9ccc5d53d9aa7c7b65dedc56abf6a37"} Dec 01 17:26:01 crc kubenswrapper[4810]: I1201 17:26:01.484215 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xdmnf/crc-debug-rscgb"] Dec 01 17:26:01 crc kubenswrapper[4810]: I1201 17:26:01.493260 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xdmnf/crc-debug-rscgb"] Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.254456 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xdmnf/crc-debug-rscgb" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.287039 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28ede625-26f9-4802-8d47-0b3cfc0899ba-host\") pod \"28ede625-26f9-4802-8d47-0b3cfc0899ba\" (UID: \"28ede625-26f9-4802-8d47-0b3cfc0899ba\") " Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.287203 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28ede625-26f9-4802-8d47-0b3cfc0899ba-host" (OuterVolumeSpecName: "host") pod "28ede625-26f9-4802-8d47-0b3cfc0899ba" (UID: "28ede625-26f9-4802-8d47-0b3cfc0899ba"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.287942 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckfdx\" (UniqueName: \"kubernetes.io/projected/28ede625-26f9-4802-8d47-0b3cfc0899ba-kube-api-access-ckfdx\") pod \"28ede625-26f9-4802-8d47-0b3cfc0899ba\" (UID: \"28ede625-26f9-4802-8d47-0b3cfc0899ba\") " Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.288851 4810 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28ede625-26f9-4802-8d47-0b3cfc0899ba-host\") on node \"crc\" DevicePath \"\"" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.296870 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28ede625-26f9-4802-8d47-0b3cfc0899ba-kube-api-access-ckfdx" (OuterVolumeSpecName: "kube-api-access-ckfdx") pod "28ede625-26f9-4802-8d47-0b3cfc0899ba" (UID: "28ede625-26f9-4802-8d47-0b3cfc0899ba"). InnerVolumeSpecName "kube-api-access-ckfdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.390270 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckfdx\" (UniqueName: \"kubernetes.io/projected/28ede625-26f9-4802-8d47-0b3cfc0899ba-kube-api-access-ckfdx\") on node \"crc\" DevicePath \"\"" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.512650 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28ede625-26f9-4802-8d47-0b3cfc0899ba" path="/var/lib/kubelet/pods/28ede625-26f9-4802-8d47-0b3cfc0899ba/volumes" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.712866 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xdmnf/crc-debug-l4tn6"] Dec 01 17:26:02 crc kubenswrapper[4810]: E1201 17:26:02.713453 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28ede625-26f9-4802-8d47-0b3cfc0899ba" containerName="container-00" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.713502 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="28ede625-26f9-4802-8d47-0b3cfc0899ba" containerName="container-00" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.713752 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="28ede625-26f9-4802-8d47-0b3cfc0899ba" containerName="container-00" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.714654 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xdmnf/crc-debug-l4tn6" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.800096 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bqkc\" (UniqueName: \"kubernetes.io/projected/b587d444-217c-4ef1-b48e-b411f2c238a4-kube-api-access-7bqkc\") pod \"crc-debug-l4tn6\" (UID: \"b587d444-217c-4ef1-b48e-b411f2c238a4\") " pod="openshift-must-gather-xdmnf/crc-debug-l4tn6" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.800300 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b587d444-217c-4ef1-b48e-b411f2c238a4-host\") pod \"crc-debug-l4tn6\" (UID: \"b587d444-217c-4ef1-b48e-b411f2c238a4\") " pod="openshift-must-gather-xdmnf/crc-debug-l4tn6" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.900960 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b587d444-217c-4ef1-b48e-b411f2c238a4-host\") pod \"crc-debug-l4tn6\" (UID: \"b587d444-217c-4ef1-b48e-b411f2c238a4\") " pod="openshift-must-gather-xdmnf/crc-debug-l4tn6" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.901078 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bqkc\" (UniqueName: \"kubernetes.io/projected/b587d444-217c-4ef1-b48e-b411f2c238a4-kube-api-access-7bqkc\") pod \"crc-debug-l4tn6\" (UID: \"b587d444-217c-4ef1-b48e-b411f2c238a4\") " pod="openshift-must-gather-xdmnf/crc-debug-l4tn6" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.901532 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b587d444-217c-4ef1-b48e-b411f2c238a4-host\") pod \"crc-debug-l4tn6\" (UID: \"b587d444-217c-4ef1-b48e-b411f2c238a4\") " pod="openshift-must-gather-xdmnf/crc-debug-l4tn6" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.922088 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bqkc\" (UniqueName: \"kubernetes.io/projected/b587d444-217c-4ef1-b48e-b411f2c238a4-kube-api-access-7bqkc\") pod \"crc-debug-l4tn6\" (UID: \"b587d444-217c-4ef1-b48e-b411f2c238a4\") " pod="openshift-must-gather-xdmnf/crc-debug-l4tn6" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.972632 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.972690 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.972735 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.973498 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 17:26:02 crc kubenswrapper[4810]: I1201 17:26:02.973547 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" gracePeriod=600 Dec 01 17:26:03 crc kubenswrapper[4810]: I1201 17:26:03.041398 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xdmnf/crc-debug-l4tn6" Dec 01 17:26:03 crc kubenswrapper[4810]: W1201 17:26:03.076198 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb587d444_217c_4ef1_b48e_b411f2c238a4.slice/crio-792803b0144b8bb0b553049feb1a3170595dfc7a22d3695bd170e31ffa61c431 WatchSource:0}: Error finding container 792803b0144b8bb0b553049feb1a3170595dfc7a22d3695bd170e31ffa61c431: Status 404 returned error can't find the container with id 792803b0144b8bb0b553049feb1a3170595dfc7a22d3695bd170e31ffa61c431 Dec 01 17:26:03 crc kubenswrapper[4810]: E1201 17:26:03.109214 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:26:03 crc kubenswrapper[4810]: I1201 17:26:03.142236 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" exitCode=0 Dec 01 17:26:03 crc kubenswrapper[4810]: I1201 17:26:03.142289 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd"} Dec 01 17:26:03 crc kubenswrapper[4810]: I1201 17:26:03.142323 4810 scope.go:117] "RemoveContainer" containerID="ac4b6e55100e42785c4ade7daada4a9f96712f083cb05f6b3c5e6cea2dcf7296" Dec 01 17:26:03 crc kubenswrapper[4810]: I1201 17:26:03.145195 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:26:03 crc kubenswrapper[4810]: E1201 17:26:03.145785 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:26:03 crc kubenswrapper[4810]: I1201 17:26:03.147984 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xdmnf/crc-debug-l4tn6" event={"ID":"b587d444-217c-4ef1-b48e-b411f2c238a4","Type":"ContainerStarted","Data":"792803b0144b8bb0b553049feb1a3170595dfc7a22d3695bd170e31ffa61c431"} Dec 01 17:26:03 crc kubenswrapper[4810]: I1201 17:26:03.157489 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xdmnf/crc-debug-rscgb" Dec 01 17:26:03 crc kubenswrapper[4810]: I1201 17:26:03.242247 4810 scope.go:117] "RemoveContainer" containerID="ba322f74151ed15529d2e68dff0e68ddf9ccc5d53d9aa7c7b65dedc56abf6a37" Dec 01 17:26:04 crc kubenswrapper[4810]: I1201 17:26:04.168329 4810 generic.go:334] "Generic (PLEG): container finished" podID="b587d444-217c-4ef1-b48e-b411f2c238a4" containerID="36342b54b5e1fea5a042995bb08c7b17c28000a31bf56b440069715b1ffccdb4" exitCode=0 Dec 01 17:26:04 crc kubenswrapper[4810]: I1201 17:26:04.168381 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xdmnf/crc-debug-l4tn6" event={"ID":"b587d444-217c-4ef1-b48e-b411f2c238a4","Type":"ContainerDied","Data":"36342b54b5e1fea5a042995bb08c7b17c28000a31bf56b440069715b1ffccdb4"} Dec 01 17:26:04 crc kubenswrapper[4810]: I1201 17:26:04.214872 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xdmnf/crc-debug-l4tn6"] Dec 01 17:26:04 crc kubenswrapper[4810]: I1201 17:26:04.224197 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xdmnf/crc-debug-l4tn6"] Dec 01 17:26:05 crc kubenswrapper[4810]: I1201 17:26:05.306820 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xdmnf/crc-debug-l4tn6" Dec 01 17:26:05 crc kubenswrapper[4810]: I1201 17:26:05.357947 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b587d444-217c-4ef1-b48e-b411f2c238a4-host\") pod \"b587d444-217c-4ef1-b48e-b411f2c238a4\" (UID: \"b587d444-217c-4ef1-b48e-b411f2c238a4\") " Dec 01 17:26:05 crc kubenswrapper[4810]: I1201 17:26:05.358093 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b587d444-217c-4ef1-b48e-b411f2c238a4-host" (OuterVolumeSpecName: "host") pod "b587d444-217c-4ef1-b48e-b411f2c238a4" (UID: "b587d444-217c-4ef1-b48e-b411f2c238a4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 17:26:05 crc kubenswrapper[4810]: I1201 17:26:05.358309 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bqkc\" (UniqueName: \"kubernetes.io/projected/b587d444-217c-4ef1-b48e-b411f2c238a4-kube-api-access-7bqkc\") pod \"b587d444-217c-4ef1-b48e-b411f2c238a4\" (UID: \"b587d444-217c-4ef1-b48e-b411f2c238a4\") " Dec 01 17:26:05 crc kubenswrapper[4810]: I1201 17:26:05.358946 4810 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b587d444-217c-4ef1-b48e-b411f2c238a4-host\") on node \"crc\" DevicePath \"\"" Dec 01 17:26:05 crc kubenswrapper[4810]: I1201 17:26:05.364128 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b587d444-217c-4ef1-b48e-b411f2c238a4-kube-api-access-7bqkc" (OuterVolumeSpecName: "kube-api-access-7bqkc") pod "b587d444-217c-4ef1-b48e-b411f2c238a4" (UID: "b587d444-217c-4ef1-b48e-b411f2c238a4"). InnerVolumeSpecName "kube-api-access-7bqkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:26:05 crc kubenswrapper[4810]: I1201 17:26:05.461100 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bqkc\" (UniqueName: \"kubernetes.io/projected/b587d444-217c-4ef1-b48e-b411f2c238a4-kube-api-access-7bqkc\") on node \"crc\" DevicePath \"\"" Dec 01 17:26:06 crc kubenswrapper[4810]: I1201 17:26:06.195319 4810 scope.go:117] "RemoveContainer" containerID="36342b54b5e1fea5a042995bb08c7b17c28000a31bf56b440069715b1ffccdb4" Dec 01 17:26:06 crc kubenswrapper[4810]: I1201 17:26:06.195369 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xdmnf/crc-debug-l4tn6" Dec 01 17:26:06 crc kubenswrapper[4810]: I1201 17:26:06.512967 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b587d444-217c-4ef1-b48e-b411f2c238a4" path="/var/lib/kubelet/pods/b587d444-217c-4ef1-b48e-b411f2c238a4/volumes" Dec 01 17:26:15 crc kubenswrapper[4810]: I1201 17:26:15.492229 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:26:15 crc kubenswrapper[4810]: E1201 17:26:15.492855 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:26:28 crc kubenswrapper[4810]: I1201 17:26:28.494843 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:26:28 crc kubenswrapper[4810]: E1201 17:26:28.495999 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:26:42 crc kubenswrapper[4810]: I1201 17:26:42.491285 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:26:42 crc kubenswrapper[4810]: E1201 17:26:42.491922 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:26:53 crc kubenswrapper[4810]: I1201 17:26:53.491919 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:26:53 crc kubenswrapper[4810]: E1201 17:26:53.492963 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:27:05 crc kubenswrapper[4810]: I1201 17:27:05.491948 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:27:05 crc kubenswrapper[4810]: E1201 17:27:05.492955 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:27:16 crc kubenswrapper[4810]: I1201 17:27:16.491489 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:27:16 crc kubenswrapper[4810]: E1201 17:27:16.492366 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:27:28 crc kubenswrapper[4810]: I1201 17:27:28.492241 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:27:28 crc kubenswrapper[4810]: E1201 17:27:28.494039 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:27:43 crc kubenswrapper[4810]: I1201 17:27:43.491463 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:27:43 crc kubenswrapper[4810]: E1201 17:27:43.492300 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:27:58 crc kubenswrapper[4810]: I1201 17:27:58.492322 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:27:58 crc kubenswrapper[4810]: E1201 17:27:58.493441 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:28:11 crc kubenswrapper[4810]: I1201 17:28:11.491589 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:28:11 crc kubenswrapper[4810]: E1201 17:28:11.492519 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:28:23 crc kubenswrapper[4810]: I1201 17:28:23.491191 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:28:23 crc kubenswrapper[4810]: E1201 17:28:23.491906 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:28:25 crc kubenswrapper[4810]: I1201 17:28:25.655809 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7/init-config-reloader/0.log" Dec 01 17:28:25 crc kubenswrapper[4810]: I1201 17:28:25.904486 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7/init-config-reloader/0.log" Dec 01 17:28:25 crc kubenswrapper[4810]: I1201 17:28:25.971162 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7/alertmanager/0.log" Dec 01 17:28:26 crc kubenswrapper[4810]: I1201 17:28:26.057952 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_8c0a871f-d2bc-4d0e-8758-4fb9dce01bc7/config-reloader/0.log" Dec 01 17:28:26 crc kubenswrapper[4810]: I1201 17:28:26.161955 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_52d88aa8-f7c3-4775-81f4-7d2ce2c4832a/aodh-api/0.log" Dec 01 17:28:26 crc kubenswrapper[4810]: I1201 17:28:26.305147 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_52d88aa8-f7c3-4775-81f4-7d2ce2c4832a/aodh-evaluator/0.log" Dec 01 17:28:26 crc kubenswrapper[4810]: I1201 17:28:26.391050 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_52d88aa8-f7c3-4775-81f4-7d2ce2c4832a/aodh-listener/0.log" Dec 01 17:28:26 crc kubenswrapper[4810]: I1201 17:28:26.554850 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5fdf9c97c6-8mwkg_9b61ad6b-ee0b-4236-883c-139c4a499987/barbican-api/0.log" Dec 01 17:28:26 crc kubenswrapper[4810]: I1201 17:28:26.579140 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_52d88aa8-f7c3-4775-81f4-7d2ce2c4832a/aodh-notifier/0.log" Dec 01 17:28:26 crc kubenswrapper[4810]: I1201 17:28:26.650345 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5fdf9c97c6-8mwkg_9b61ad6b-ee0b-4236-883c-139c4a499987/barbican-api-log/0.log" Dec 01 17:28:26 crc kubenswrapper[4810]: I1201 17:28:26.873882 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-855f768b54-b52dt_67805ea2-2459-4c00-9b18-54d3e60d1281/barbican-keystone-listener/0.log" Dec 01 17:28:27 crc kubenswrapper[4810]: I1201 17:28:27.171712 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-855f768b54-b52dt_67805ea2-2459-4c00-9b18-54d3e60d1281/barbican-keystone-listener-log/0.log" Dec 01 17:28:27 crc kubenswrapper[4810]: I1201 17:28:27.335597 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-679666fcbc-r7dpm_835bf15c-547d-4bb3-89bd-767b9c7e8f94/barbican-worker/0.log" Dec 01 17:28:27 crc kubenswrapper[4810]: I1201 17:28:27.339566 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-679666fcbc-r7dpm_835bf15c-547d-4bb3-89bd-767b9c7e8f94/barbican-worker-log/0.log" Dec 01 17:28:27 crc kubenswrapper[4810]: I1201 17:28:27.518548 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-2fqfj_2f0d06d5-7eea-4c8d-8700-439ea72af7cc/bootstrap-openstack-openstack-cell1/0.log" Dec 01 17:28:27 crc kubenswrapper[4810]: I1201 17:28:27.661577 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_40aec98f-dc77-482d-b562-c5e5c9eed98d/ceilometer-central-agent/0.log" Dec 01 17:28:27 crc kubenswrapper[4810]: I1201 17:28:27.854651 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_40aec98f-dc77-482d-b562-c5e5c9eed98d/ceilometer-notification-agent/0.log" Dec 01 17:28:27 crc kubenswrapper[4810]: I1201 17:28:27.875936 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_40aec98f-dc77-482d-b562-c5e5c9eed98d/proxy-httpd/0.log" Dec 01 17:28:27 crc kubenswrapper[4810]: I1201 17:28:27.939711 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_40aec98f-dc77-482d-b562-c5e5c9eed98d/sg-core/0.log" Dec 01 17:28:28 crc kubenswrapper[4810]: I1201 17:28:28.116097 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4412031a-3547-49c1-907b-450db04bc075/cinder-api-log/0.log" Dec 01 17:28:28 crc kubenswrapper[4810]: I1201 17:28:28.166337 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4412031a-3547-49c1-907b-450db04bc075/cinder-api/0.log" Dec 01 17:28:28 crc kubenswrapper[4810]: I1201 17:28:28.360442 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_82892053-06d4-414e-9c79-4fd70eea44b3/cinder-scheduler/0.log" Dec 01 17:28:28 crc kubenswrapper[4810]: I1201 17:28:28.373967 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_82892053-06d4-414e-9c79-4fd70eea44b3/probe/0.log" Dec 01 17:28:28 crc kubenswrapper[4810]: I1201 17:28:28.429368 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-kk9r8_e050d22d-6f4f-449a-99dc-8dbf9b5603a3/configure-network-openstack-openstack-cell1/0.log" Dec 01 17:28:28 crc kubenswrapper[4810]: I1201 17:28:28.769403 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5555dfd759-gjw5c_cca49c71-6d05-4439-86a0-1e920aa27fc8/init/0.log" Dec 01 17:28:28 crc kubenswrapper[4810]: I1201 17:28:28.780525 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-5sbhr_ba759ad7-3f5c-4d82-a406-54b887ede26a/configure-os-openstack-openstack-cell1/0.log" Dec 01 17:28:28 crc kubenswrapper[4810]: I1201 17:28:28.985874 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5555dfd759-gjw5c_cca49c71-6d05-4439-86a0-1e920aa27fc8/init/0.log" Dec 01 17:28:29 crc kubenswrapper[4810]: I1201 17:28:29.025124 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5555dfd759-gjw5c_cca49c71-6d05-4439-86a0-1e920aa27fc8/dnsmasq-dns/0.log" Dec 01 17:28:29 crc kubenswrapper[4810]: I1201 17:28:29.038267 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-ptllz_c1bcac83-4d6b-46bc-83bc-3b5328c1c021/download-cache-openstack-openstack-cell1/0.log" Dec 01 17:28:29 crc kubenswrapper[4810]: I1201 17:28:29.251035 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe/glance-log/0.log" Dec 01 17:28:29 crc kubenswrapper[4810]: I1201 17:28:29.274784 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1644e6e7-38cf-4ecc-bf2e-1c0482cf0cbe/glance-httpd/0.log" Dec 01 17:28:29 crc kubenswrapper[4810]: I1201 17:28:29.366208 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_11ef5ba9-de25-4010-8a26-a8f38b5e7e08/glance-httpd/0.log" Dec 01 17:28:29 crc kubenswrapper[4810]: I1201 17:28:29.494194 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_11ef5ba9-de25-4010-8a26-a8f38b5e7e08/glance-log/0.log" Dec 01 17:28:29 crc kubenswrapper[4810]: I1201 17:28:29.913620 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-656cb86f76-jpbk2_9890a860-9f91-4a78-b26b-23c6c97a130d/heat-engine/0.log" Dec 01 17:28:30 crc kubenswrapper[4810]: I1201 17:28:30.162978 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-574946f46-cwn7p_73a62ae2-3e32-424b-b181-da93d1c88f35/horizon/0.log" Dec 01 17:28:30 crc kubenswrapper[4810]: I1201 17:28:30.213977 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-db946669-jbbqh_0b1083d7-8870-4fbb-b4aa-a25dfaef9b14/heat-api/0.log" Dec 01 17:28:30 crc kubenswrapper[4810]: I1201 17:28:30.378248 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-7d4fd6fc6c-6wbqj_24dcd733-d9bb-4def-9325-f1dd2611d1f1/heat-cfnapi/0.log" Dec 01 17:28:30 crc kubenswrapper[4810]: I1201 17:28:30.680423 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-574946f46-cwn7p_73a62ae2-3e32-424b-b181-da93d1c88f35/horizon-log/0.log" Dec 01 17:28:30 crc kubenswrapper[4810]: I1201 17:28:30.694604 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-mfkj5_84f52a8b-4f35-4a57-93c9-05090d31a876/install-os-openstack-openstack-cell1/0.log" Dec 01 17:28:30 crc kubenswrapper[4810]: I1201 17:28:30.743772 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-sr2t8_17bffe79-728c-4500-a27e-a2b246f82517/install-certs-openstack-openstack-cell1/0.log" Dec 01 17:28:30 crc kubenswrapper[4810]: I1201 17:28:30.910852 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29410141-lkmhs_4a37de29-2b57-4f07-b0d2-cf186488012e/keystone-cron/0.log" Dec 01 17:28:31 crc kubenswrapper[4810]: I1201 17:28:31.037885 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7d4469b5c6-z5bkt_83a32a66-545a-4122-a47e-a7111a5d745f/keystone-api/0.log" Dec 01 17:28:31 crc kubenswrapper[4810]: I1201 17:28:31.073526 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_4b853756-ebf7-4dac-9e50-67524f3b8abe/kube-state-metrics/0.log" Dec 01 17:28:31 crc kubenswrapper[4810]: I1201 17:28:31.246861 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-xjz8d_051ab66c-513f-4f4f-98a9-a20426bea652/libvirt-openstack-openstack-cell1/0.log" Dec 01 17:28:31 crc kubenswrapper[4810]: I1201 17:28:31.582582 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-9776bf6cc-vk57d_3fc53f93-4c92-44c4-998b-8be33a38d18f/neutron-httpd/0.log" Dec 01 17:28:31 crc kubenswrapper[4810]: I1201 17:28:31.606428 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-9776bf6cc-vk57d_3fc53f93-4c92-44c4-998b-8be33a38d18f/neutron-api/0.log" Dec 01 17:28:31 crc kubenswrapper[4810]: I1201 17:28:31.706100 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-jjwfn_db74fde5-b56b-45ca-bac8-23a2adac45f5/neutron-dhcp-openstack-openstack-cell1/0.log" Dec 01 17:28:31 crc kubenswrapper[4810]: I1201 17:28:31.926844 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-xpffb_69ca686f-3e3b-42eb-a5b5-f1a28a950c6a/neutron-metadata-openstack-openstack-cell1/0.log" Dec 01 17:28:32 crc kubenswrapper[4810]: I1201 17:28:32.054673 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-2csms_791f52f5-96d0-4949-813d-c86ddf995d9f/neutron-sriov-openstack-openstack-cell1/0.log" Dec 01 17:28:32 crc kubenswrapper[4810]: I1201 17:28:32.805693 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_f44e8294-c07f-4c82-8957-82b096b3d87b/nova-api-log/0.log" Dec 01 17:28:32 crc kubenswrapper[4810]: I1201 17:28:32.852850 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_f44e8294-c07f-4c82-8957-82b096b3d87b/nova-api-api/0.log" Dec 01 17:28:32 crc kubenswrapper[4810]: I1201 17:28:32.969749 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_e7feb5c3-af78-44b4-a415-2a77e1433182/nova-cell0-conductor-conductor/0.log" Dec 01 17:28:33 crc kubenswrapper[4810]: I1201 17:28:33.124520 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_56c3b82f-7a29-4e75-b942-befc510f892f/nova-cell1-conductor-conductor/0.log" Dec 01 17:28:33 crc kubenswrapper[4810]: I1201 17:28:33.280545 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_cf4e50ff-bd60-4fef-a1dd-b539720329ba/nova-cell1-novncproxy-novncproxy/0.log" Dec 01 17:28:33 crc kubenswrapper[4810]: I1201 17:28:33.422062 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellnbr8n_cd2bd342-3a99-4d38-976b-3f213999f103/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Dec 01 17:28:33 crc kubenswrapper[4810]: I1201 17:28:33.801741 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-8cbjx_7bfcd0d6-c140-4c51-aa35-8acd8a96d68f/nova-cell1-openstack-openstack-cell1/0.log" Dec 01 17:28:34 crc kubenswrapper[4810]: I1201 17:28:34.640495 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_03fe61c9-4219-44e3-b659-6c47a5a6ee42/nova-metadata-log/0.log" Dec 01 17:28:34 crc kubenswrapper[4810]: I1201 17:28:34.680431 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_82f24148-c70d-4110-934e-f10d5b32eb1e/nova-scheduler-scheduler/0.log" Dec 01 17:28:34 crc kubenswrapper[4810]: I1201 17:28:34.901694 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-86b8958b56-x87pp_a6821df4-0318-4d9c-8d47-116e602817fb/init/0.log" Dec 01 17:28:34 crc kubenswrapper[4810]: I1201 17:28:34.928108 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_03fe61c9-4219-44e3-b659-6c47a5a6ee42/nova-metadata-metadata/0.log" Dec 01 17:28:35 crc kubenswrapper[4810]: I1201 17:28:35.079096 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-86b8958b56-x87pp_a6821df4-0318-4d9c-8d47-116e602817fb/init/0.log" Dec 01 17:28:35 crc kubenswrapper[4810]: I1201 17:28:35.140518 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-86b8958b56-x87pp_a6821df4-0318-4d9c-8d47-116e602817fb/octavia-api-provider-agent/0.log" Dec 01 17:28:35 crc kubenswrapper[4810]: I1201 17:28:35.283025 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-frfm6_6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b/init/0.log" Dec 01 17:28:35 crc kubenswrapper[4810]: I1201 17:28:35.327720 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-86b8958b56-x87pp_a6821df4-0318-4d9c-8d47-116e602817fb/octavia-api/0.log" Dec 01 17:28:35 crc kubenswrapper[4810]: I1201 17:28:35.477499 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-frfm6_6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b/init/0.log" Dec 01 17:28:35 crc kubenswrapper[4810]: I1201 17:28:35.631901 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-qthjz_a88e2d45-97f3-4c58-ade1-9ce92b0897a8/init/0.log" Dec 01 17:28:35 crc kubenswrapper[4810]: I1201 17:28:35.663856 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-frfm6_6d1d24bd-a0f3-4a10-89fb-e24bcd07f05b/octavia-healthmanager/0.log" Dec 01 17:28:35 crc kubenswrapper[4810]: I1201 17:28:35.880197 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-qthjz_a88e2d45-97f3-4c58-ade1-9ce92b0897a8/init/0.log" Dec 01 17:28:35 crc kubenswrapper[4810]: I1201 17:28:35.895497 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-qthjz_a88e2d45-97f3-4c58-ade1-9ce92b0897a8/octavia-housekeeping/0.log" Dec 01 17:28:35 crc kubenswrapper[4810]: I1201 17:28:35.984356 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-699f5f4fd7-ck6rq_883efbb3-cd1c-45d5-a17e-482ca1b8eb9e/init/0.log" Dec 01 17:28:36 crc kubenswrapper[4810]: I1201 17:28:36.170890 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-699f5f4fd7-ck6rq_883efbb3-cd1c-45d5-a17e-482ca1b8eb9e/octavia-amphora-httpd/0.log" Dec 01 17:28:36 crc kubenswrapper[4810]: I1201 17:28:36.182404 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-699f5f4fd7-ck6rq_883efbb3-cd1c-45d5-a17e-482ca1b8eb9e/init/0.log" Dec 01 17:28:36 crc kubenswrapper[4810]: I1201 17:28:36.216390 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-qs2z8_30053c9f-67a1-4ec1-b2bf-c61ff171a0d9/init/0.log" Dec 01 17:28:36 crc kubenswrapper[4810]: I1201 17:28:36.478582 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-qs2z8_30053c9f-67a1-4ec1-b2bf-c61ff171a0d9/init/0.log" Dec 01 17:28:36 crc kubenswrapper[4810]: I1201 17:28:36.482761 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-qs2z8_30053c9f-67a1-4ec1-b2bf-c61ff171a0d9/octavia-rsyslog/0.log" Dec 01 17:28:36 crc kubenswrapper[4810]: I1201 17:28:36.493617 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:28:36 crc kubenswrapper[4810]: E1201 17:28:36.493872 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:28:36 crc kubenswrapper[4810]: I1201 17:28:36.550453 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-4gnz2_8e768084-da29-4e95-ae6e-feacb0cbfb70/init/0.log" Dec 01 17:28:36 crc kubenswrapper[4810]: I1201 17:28:36.808349 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-4gnz2_8e768084-da29-4e95-ae6e-feacb0cbfb70/init/0.log" Dec 01 17:28:36 crc kubenswrapper[4810]: I1201 17:28:36.870431 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877/mysql-bootstrap/0.log" Dec 01 17:28:36 crc kubenswrapper[4810]: I1201 17:28:36.973138 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-4gnz2_8e768084-da29-4e95-ae6e-feacb0cbfb70/octavia-worker/0.log" Dec 01 17:28:37 crc kubenswrapper[4810]: I1201 17:28:37.071808 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877/mysql-bootstrap/0.log" Dec 01 17:28:37 crc kubenswrapper[4810]: I1201 17:28:37.112775 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bbcb3289-cf5d-4f8d-bf1f-fe30d45ef877/galera/0.log" Dec 01 17:28:37 crc kubenswrapper[4810]: I1201 17:28:37.276760 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9a264de6-cd64-441c-b935-d7c30e7ec733/mysql-bootstrap/0.log" Dec 01 17:28:37 crc kubenswrapper[4810]: I1201 17:28:37.705671 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_20828f9e-5524-4c73-8754-2e1d60bb775a/openstackclient/0.log" Dec 01 17:28:37 crc kubenswrapper[4810]: I1201 17:28:37.743133 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9a264de6-cd64-441c-b935-d7c30e7ec733/mysql-bootstrap/0.log" Dec 01 17:28:37 crc kubenswrapper[4810]: I1201 17:28:37.756375 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9a264de6-cd64-441c-b935-d7c30e7ec733/galera/0.log" Dec 01 17:28:38 crc kubenswrapper[4810]: I1201 17:28:38.057228 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4kw8j_8a092636-4ef4-4e56-91ad-cf0518f2c88d/ovn-controller/0.log" Dec 01 17:28:38 crc kubenswrapper[4810]: I1201 17:28:38.120561 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-tmvcm_59c20ba6-6bec-443f-911d-02d9139ac260/openstack-network-exporter/0.log" Dec 01 17:28:38 crc kubenswrapper[4810]: I1201 17:28:38.301170 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rq44f_6340cc9b-5129-437b-90fe-8b031224943c/ovsdb-server-init/0.log" Dec 01 17:28:38 crc kubenswrapper[4810]: I1201 17:28:38.490423 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rq44f_6340cc9b-5129-437b-90fe-8b031224943c/ovsdb-server/0.log" Dec 01 17:28:38 crc kubenswrapper[4810]: I1201 17:28:38.502931 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rq44f_6340cc9b-5129-437b-90fe-8b031224943c/ovs-vswitchd/0.log" Dec 01 17:28:38 crc kubenswrapper[4810]: I1201 17:28:38.515697 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-rq44f_6340cc9b-5129-437b-90fe-8b031224943c/ovsdb-server-init/0.log" Dec 01 17:28:38 crc kubenswrapper[4810]: I1201 17:28:38.697852 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_5e8bace8-979e-4aa2-bacd-e9c45c301391/openstack-network-exporter/0.log" Dec 01 17:28:38 crc kubenswrapper[4810]: I1201 17:28:38.793219 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_5e8bace8-979e-4aa2-bacd-e9c45c301391/ovn-northd/0.log" Dec 01 17:28:38 crc kubenswrapper[4810]: I1201 17:28:38.910161 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-zsqk6_e1252894-fad3-4959-99b1-5564dff922ec/ovn-openstack-openstack-cell1/0.log" Dec 01 17:28:39 crc kubenswrapper[4810]: I1201 17:28:39.017505 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_2edd15df-7752-4d42-b3e9-cfd79df85565/openstack-network-exporter/0.log" Dec 01 17:28:39 crc kubenswrapper[4810]: I1201 17:28:39.062682 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_2edd15df-7752-4d42-b3e9-cfd79df85565/ovsdbserver-nb/0.log" Dec 01 17:28:39 crc kubenswrapper[4810]: I1201 17:28:39.213033 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05/openstack-network-exporter/0.log" Dec 01 17:28:39 crc kubenswrapper[4810]: I1201 17:28:39.255811 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_d7d2b7f9-4a49-45e2-abfd-a55a3f5d3a05/ovsdbserver-nb/0.log" Dec 01 17:28:39 crc kubenswrapper[4810]: I1201 17:28:39.429871 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_418356d0-f609-4fa5-a31a-2b3c83a6f5d7/openstack-network-exporter/0.log" Dec 01 17:28:39 crc kubenswrapper[4810]: I1201 17:28:39.451782 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_418356d0-f609-4fa5-a31a-2b3c83a6f5d7/ovsdbserver-nb/0.log" Dec 01 17:28:39 crc kubenswrapper[4810]: I1201 17:28:39.608313 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d2db4252-6bc3-40d8-a375-03660d2aeae0/openstack-network-exporter/0.log" Dec 01 17:28:39 crc kubenswrapper[4810]: I1201 17:28:39.631705 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d2db4252-6bc3-40d8-a375-03660d2aeae0/ovsdbserver-sb/0.log" Dec 01 17:28:39 crc kubenswrapper[4810]: I1201 17:28:39.722028 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_fd18cf90-4008-4cf1-953e-d0caf3382dd2/openstack-network-exporter/0.log" Dec 01 17:28:39 crc kubenswrapper[4810]: I1201 17:28:39.852901 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_fd18cf90-4008-4cf1-953e-d0caf3382dd2/ovsdbserver-sb/0.log" Dec 01 17:28:39 crc kubenswrapper[4810]: I1201 17:28:39.934104 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_b15e71bf-2d3e-47b1-9013-a051eb580331/openstack-network-exporter/0.log" Dec 01 17:28:40 crc kubenswrapper[4810]: I1201 17:28:40.005766 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_b15e71bf-2d3e-47b1-9013-a051eb580331/ovsdbserver-sb/0.log" Dec 01 17:28:40 crc kubenswrapper[4810]: I1201 17:28:40.248783 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-668cb558f4-t5z5m_77f99599-4efe-41a7-8781-dcfd3fd4e8f4/placement-api/0.log" Dec 01 17:28:40 crc kubenswrapper[4810]: I1201 17:28:40.290943 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-668cb558f4-t5z5m_77f99599-4efe-41a7-8781-dcfd3fd4e8f4/placement-log/0.log" Dec 01 17:28:40 crc kubenswrapper[4810]: I1201 17:28:40.426578 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-c5qqn5_6c6da269-a52f-427a-a457-0e6fe712f398/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Dec 01 17:28:40 crc kubenswrapper[4810]: I1201 17:28:40.512096 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e98b441b-4ef4-45d4-8358-ab8603841e8a/init-config-reloader/0.log" Dec 01 17:28:40 crc kubenswrapper[4810]: I1201 17:28:40.718157 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e98b441b-4ef4-45d4-8358-ab8603841e8a/init-config-reloader/0.log" Dec 01 17:28:40 crc kubenswrapper[4810]: I1201 17:28:40.792902 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e98b441b-4ef4-45d4-8358-ab8603841e8a/prometheus/0.log" Dec 01 17:28:40 crc kubenswrapper[4810]: I1201 17:28:40.804632 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e98b441b-4ef4-45d4-8358-ab8603841e8a/config-reloader/0.log" Dec 01 17:28:40 crc kubenswrapper[4810]: I1201 17:28:40.854038 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e98b441b-4ef4-45d4-8358-ab8603841e8a/thanos-sidecar/0.log" Dec 01 17:28:40 crc kubenswrapper[4810]: I1201 17:28:40.988511 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8d4f1ab5-87de-41d0-877c-8ff39bdff385/setup-container/0.log" Dec 01 17:28:41 crc kubenswrapper[4810]: I1201 17:28:41.215073 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8d4f1ab5-87de-41d0-877c-8ff39bdff385/rabbitmq/0.log" Dec 01 17:28:41 crc kubenswrapper[4810]: I1201 17:28:41.261531 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_8d4f1ab5-87de-41d0-877c-8ff39bdff385/setup-container/0.log" Dec 01 17:28:41 crc kubenswrapper[4810]: I1201 17:28:41.269017 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_fc454c84-0d74-446b-a396-3b0185e716a2/setup-container/0.log" Dec 01 17:28:41 crc kubenswrapper[4810]: I1201 17:28:41.526794 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_fc454c84-0d74-446b-a396-3b0185e716a2/rabbitmq/0.log" Dec 01 17:28:41 crc kubenswrapper[4810]: I1201 17:28:41.528162 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_fc454c84-0d74-446b-a396-3b0185e716a2/setup-container/0.log" Dec 01 17:28:41 crc kubenswrapper[4810]: I1201 17:28:41.566435 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-j84h9_60cc48f6-a14e-4bb3-b979-60e2a240313c/reboot-os-openstack-openstack-cell1/0.log" Dec 01 17:28:41 crc kubenswrapper[4810]: I1201 17:28:41.727417 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-h4xb9_c2ad2c28-93e4-4521-96ab-40c186aade64/run-os-openstack-openstack-cell1/0.log" Dec 01 17:28:42 crc kubenswrapper[4810]: I1201 17:28:42.492273 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-ksq6w_6f47201d-aeaa-430b-bbc6-86b19519f233/ssh-known-hosts-openstack/0.log" Dec 01 17:28:42 crc kubenswrapper[4810]: I1201 17:28:42.711516 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-79646c549d-jxbn6_e9c91a11-9c09-49e7-b228-b37eddc05cee/proxy-server/0.log" Dec 01 17:28:42 crc kubenswrapper[4810]: I1201 17:28:42.744726 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-79646c549d-jxbn6_e9c91a11-9c09-49e7-b228-b37eddc05cee/proxy-httpd/0.log" Dec 01 17:28:42 crc kubenswrapper[4810]: I1201 17:28:42.782473 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-zcb6q_58f5e1dc-2c1b-4bee-b02f-c584327d7cbc/swift-ring-rebalance/0.log" Dec 01 17:28:42 crc kubenswrapper[4810]: I1201 17:28:42.953770 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-kp8l4_d35d5ae0-9a24-4816-80d6-b25a71f73a52/telemetry-openstack-openstack-cell1/0.log" Dec 01 17:28:43 crc kubenswrapper[4810]: I1201 17:28:43.082437 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-75v8b_229207e0-27f2-4caa-b587-8340921ac643/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Dec 01 17:28:43 crc kubenswrapper[4810]: I1201 17:28:43.180531 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-fr5h2_68a329d7-d28e-40ea-bf13-8b3209932027/validate-network-openstack-openstack-cell1/0.log" Dec 01 17:28:45 crc kubenswrapper[4810]: I1201 17:28:45.933644 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_0b48f20e-a414-4e62-a9a7-8ecfaef12dc7/memcached/0.log" Dec 01 17:28:47 crc kubenswrapper[4810]: I1201 17:28:47.491102 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:28:47 crc kubenswrapper[4810]: E1201 17:28:47.492175 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:29:02 crc kubenswrapper[4810]: I1201 17:29:02.490982 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:29:02 crc kubenswrapper[4810]: E1201 17:29:02.491741 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:29:12 crc kubenswrapper[4810]: I1201 17:29:12.558683 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6_2d8a8743-6e6d-438c-bb7a-c48047c8c7ed/util/0.log" Dec 01 17:29:12 crc kubenswrapper[4810]: I1201 17:29:12.717625 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6_2d8a8743-6e6d-438c-bb7a-c48047c8c7ed/pull/0.log" Dec 01 17:29:12 crc kubenswrapper[4810]: I1201 17:29:12.725257 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6_2d8a8743-6e6d-438c-bb7a-c48047c8c7ed/util/0.log" Dec 01 17:29:12 crc kubenswrapper[4810]: I1201 17:29:12.750872 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6_2d8a8743-6e6d-438c-bb7a-c48047c8c7ed/pull/0.log" Dec 01 17:29:12 crc kubenswrapper[4810]: I1201 17:29:12.918754 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6_2d8a8743-6e6d-438c-bb7a-c48047c8c7ed/extract/0.log" Dec 01 17:29:12 crc kubenswrapper[4810]: I1201 17:29:12.921102 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6_2d8a8743-6e6d-438c-bb7a-c48047c8c7ed/pull/0.log" Dec 01 17:29:12 crc kubenswrapper[4810]: I1201 17:29:12.940796 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_34453e0ae93d07abc4f6e497f8998de77c1bdd8f20510be6b58912cf3b6rbq6_2d8a8743-6e6d-438c-bb7a-c48047c8c7ed/util/0.log" Dec 01 17:29:13 crc kubenswrapper[4810]: I1201 17:29:13.153206 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-4zjxs_308f7f89-2d44-42cb-8533-2b82306341c6/kube-rbac-proxy/0.log" Dec 01 17:29:13 crc kubenswrapper[4810]: I1201 17:29:13.189653 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-4zjxs_308f7f89-2d44-42cb-8533-2b82306341c6/manager/0.log" Dec 01 17:29:13 crc kubenswrapper[4810]: I1201 17:29:13.194587 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-ccwrk_4b6ce85f-d265-40a2-aa92-2a2d76339acf/kube-rbac-proxy/0.log" Dec 01 17:29:13 crc kubenswrapper[4810]: I1201 17:29:13.409966 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-ccwrk_4b6ce85f-d265-40a2-aa92-2a2d76339acf/manager/0.log" Dec 01 17:29:13 crc kubenswrapper[4810]: I1201 17:29:13.417084 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-7bcjp_322588d9-4c11-4676-8eb4-15cce77ccd44/manager/0.log" Dec 01 17:29:13 crc kubenswrapper[4810]: I1201 17:29:13.448590 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-7bcjp_322588d9-4c11-4676-8eb4-15cce77ccd44/kube-rbac-proxy/0.log" Dec 01 17:29:13 crc kubenswrapper[4810]: I1201 17:29:13.577248 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-nnvmc_ae4759b1-d998-49b4-8031-f638141177a5/kube-rbac-proxy/0.log" Dec 01 17:29:13 crc kubenswrapper[4810]: I1201 17:29:13.735043 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-nnvmc_ae4759b1-d998-49b4-8031-f638141177a5/manager/0.log" Dec 01 17:29:13 crc kubenswrapper[4810]: I1201 17:29:13.772047 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-lh6jm_fc17d065-8168-486a-ba45-1fd0cd31d9f0/kube-rbac-proxy/0.log" Dec 01 17:29:13 crc kubenswrapper[4810]: I1201 17:29:13.856282 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-lh6jm_fc17d065-8168-486a-ba45-1fd0cd31d9f0/manager/0.log" Dec 01 17:29:13 crc kubenswrapper[4810]: I1201 17:29:13.926000 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-9zzgw_b6329d13-284d-48cc-866e-57744f0826bf/kube-rbac-proxy/0.log" Dec 01 17:29:14 crc kubenswrapper[4810]: I1201 17:29:14.013775 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-9zzgw_b6329d13-284d-48cc-866e-57744f0826bf/manager/0.log" Dec 01 17:29:14 crc kubenswrapper[4810]: I1201 17:29:14.115220 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-2lp78_50c4dc84-7965-4b16-8935-f557dc0c3a74/kube-rbac-proxy/0.log" Dec 01 17:29:14 crc kubenswrapper[4810]: I1201 17:29:14.309903 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-hpln8_573876c9-548b-477b-8d92-b7b7d50c7bb5/kube-rbac-proxy/0.log" Dec 01 17:29:14 crc kubenswrapper[4810]: I1201 17:29:14.330338 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-hpln8_573876c9-548b-477b-8d92-b7b7d50c7bb5/manager/0.log" Dec 01 17:29:14 crc kubenswrapper[4810]: I1201 17:29:14.493250 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-2lp78_50c4dc84-7965-4b16-8935-f557dc0c3a74/manager/0.log" Dec 01 17:29:14 crc kubenswrapper[4810]: I1201 17:29:14.497821 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:29:14 crc kubenswrapper[4810]: E1201 17:29:14.498094 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:29:14 crc kubenswrapper[4810]: I1201 17:29:14.515754 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-gvx8g_e6700584-b120-491c-8920-caa3332ba444/kube-rbac-proxy/0.log" Dec 01 17:29:14 crc kubenswrapper[4810]: I1201 17:29:14.647576 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-gvx8g_e6700584-b120-491c-8920-caa3332ba444/manager/0.log" Dec 01 17:29:14 crc kubenswrapper[4810]: I1201 17:29:14.719589 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-wss6v_b56ca08d-173a-48a4-a537-8b9e84b4b2a2/manager/0.log" Dec 01 17:29:14 crc kubenswrapper[4810]: I1201 17:29:14.730440 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-wss6v_b56ca08d-173a-48a4-a537-8b9e84b4b2a2/kube-rbac-proxy/0.log" Dec 01 17:29:14 crc kubenswrapper[4810]: I1201 17:29:14.845129 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-72tnh_e67d0431-4b18-4743-842d-d0728b23727a/kube-rbac-proxy/0.log" Dec 01 17:29:14 crc kubenswrapper[4810]: I1201 17:29:14.918874 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-72tnh_e67d0431-4b18-4743-842d-d0728b23727a/manager/0.log" Dec 01 17:29:15 crc kubenswrapper[4810]: I1201 17:29:15.063270 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-lrsc2_27e828db-c91c-49ff-88f2-eaab1ea075a0/kube-rbac-proxy/0.log" Dec 01 17:29:15 crc kubenswrapper[4810]: I1201 17:29:15.070307 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-lrsc2_27e828db-c91c-49ff-88f2-eaab1ea075a0/manager/0.log" Dec 01 17:29:15 crc kubenswrapper[4810]: I1201 17:29:15.210962 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-x2png_7798189d-af43-43aa-afd1-4257194bcb12/kube-rbac-proxy/0.log" Dec 01 17:29:15 crc kubenswrapper[4810]: I1201 17:29:15.427858 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-x2png_7798189d-af43-43aa-afd1-4257194bcb12/manager/0.log" Dec 01 17:29:15 crc kubenswrapper[4810]: I1201 17:29:15.433120 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-9pbbh_c88f9be0-7e6f-4021-86c2-a43b34012a2c/kube-rbac-proxy/0.log" Dec 01 17:29:15 crc kubenswrapper[4810]: I1201 17:29:15.506776 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-9pbbh_c88f9be0-7e6f-4021-86c2-a43b34012a2c/manager/0.log" Dec 01 17:29:15 crc kubenswrapper[4810]: I1201 17:29:15.621196 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6698bcb44667k99_68116291-3829-4a92-aba9-c1fbb755c988/kube-rbac-proxy/0.log" Dec 01 17:29:15 crc kubenswrapper[4810]: I1201 17:29:15.635710 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6698bcb44667k99_68116291-3829-4a92-aba9-c1fbb755c988/manager/0.log" Dec 01 17:29:16 crc kubenswrapper[4810]: I1201 17:29:16.038980 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6ddddd9d6f-rs8qj_666e10e7-f807-43b7-9ee4-5dabdeb0c8f7/operator/0.log" Dec 01 17:29:16 crc kubenswrapper[4810]: I1201 17:29:16.238667 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-788fn_816177c4-0c9b-4826-8844-84db106ec51d/kube-rbac-proxy/0.log" Dec 01 17:29:16 crc kubenswrapper[4810]: I1201 17:29:16.314187 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-4s9sl_15fa679c-c2b6-4b41-a18c-cf0f42d1b4cf/registry-server/0.log" Dec 01 17:29:16 crc kubenswrapper[4810]: I1201 17:29:16.515432 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-c82g8_7007dd52-d7df-4f27-81d0-95ee762c87c1/kube-rbac-proxy/0.log" Dec 01 17:29:16 crc kubenswrapper[4810]: I1201 17:29:16.604802 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-c82g8_7007dd52-d7df-4f27-81d0-95ee762c87c1/manager/0.log" Dec 01 17:29:16 crc kubenswrapper[4810]: I1201 17:29:16.615827 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-788fn_816177c4-0c9b-4826-8844-84db106ec51d/manager/0.log" Dec 01 17:29:16 crc kubenswrapper[4810]: I1201 17:29:16.742445 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-xvf22_380f8fee-59c6-445e-9a1e-47abe0eda51d/operator/0.log" Dec 01 17:29:16 crc kubenswrapper[4810]: I1201 17:29:16.883665 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-z85kg_19093fac-7ee4-43aa-ad57-7561d75de41f/kube-rbac-proxy/0.log" Dec 01 17:29:16 crc kubenswrapper[4810]: I1201 17:29:16.996338 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-z85kg_19093fac-7ee4-43aa-ad57-7561d75de41f/manager/0.log" Dec 01 17:29:17 crc kubenswrapper[4810]: I1201 17:29:17.029990 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-92m6z_58be13b4-ea8d-4ec2-8421-18d3fb2f744d/kube-rbac-proxy/0.log" Dec 01 17:29:17 crc kubenswrapper[4810]: I1201 17:29:17.224569 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-fs6gw_36f3dbad-9483-4695-95fe-7ce6520b4db3/manager/0.log" Dec 01 17:29:17 crc kubenswrapper[4810]: I1201 17:29:17.245637 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-fs6gw_36f3dbad-9483-4695-95fe-7ce6520b4db3/kube-rbac-proxy/0.log" Dec 01 17:29:17 crc kubenswrapper[4810]: I1201 17:29:17.343551 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-92m6z_58be13b4-ea8d-4ec2-8421-18d3fb2f744d/manager/0.log" Dec 01 17:29:17 crc kubenswrapper[4810]: I1201 17:29:17.459415 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-8p7kt_dc03c0f9-e5d2-4fa3-98c6-d9a1d46638e0/kube-rbac-proxy/0.log" Dec 01 17:29:17 crc kubenswrapper[4810]: I1201 17:29:17.566338 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-8p7kt_dc03c0f9-e5d2-4fa3-98c6-d9a1d46638e0/manager/0.log" Dec 01 17:29:18 crc kubenswrapper[4810]: I1201 17:29:18.257279 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-656fd97d56-9ghs2_2840e582-6c19-4473-9392-37a693d800dc/manager/0.log" Dec 01 17:29:26 crc kubenswrapper[4810]: I1201 17:29:26.491967 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:29:26 crc kubenswrapper[4810]: E1201 17:29:26.493226 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:29:36 crc kubenswrapper[4810]: I1201 17:29:36.241058 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-jnrsh_62d665ff-bba5-4ed3-b0f2-471be90260b7/control-plane-machine-set-operator/0.log" Dec 01 17:29:36 crc kubenswrapper[4810]: I1201 17:29:36.390913 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-r2vhz_7e7e3fa5-5c91-4694-95f4-67a503c3f42c/kube-rbac-proxy/0.log" Dec 01 17:29:36 crc kubenswrapper[4810]: I1201 17:29:36.443574 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-r2vhz_7e7e3fa5-5c91-4694-95f4-67a503c3f42c/machine-api-operator/0.log" Dec 01 17:29:41 crc kubenswrapper[4810]: I1201 17:29:41.491761 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:29:41 crc kubenswrapper[4810]: E1201 17:29:41.492797 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:29:49 crc kubenswrapper[4810]: I1201 17:29:49.110188 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-6smjs_7b781210-3868-4d4b-9f15-890ac7f08766/cert-manager-controller/0.log" Dec 01 17:29:49 crc kubenswrapper[4810]: I1201 17:29:49.374570 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-j8wjf_4cabe2df-a7db-4c52-bfc1-4e205d62e3f4/cert-manager-webhook/0.log" Dec 01 17:29:49 crc kubenswrapper[4810]: I1201 17:29:49.382648 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-pzdxf_73c25462-0e75-40f9-82bb-4ab8ae29d8cf/cert-manager-cainjector/0.log" Dec 01 17:29:56 crc kubenswrapper[4810]: I1201 17:29:56.491771 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:29:56 crc kubenswrapper[4810]: E1201 17:29:56.492634 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:30:00 crc kubenswrapper[4810]: I1201 17:30:00.169713 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c"] Dec 01 17:30:00 crc kubenswrapper[4810]: E1201 17:30:00.170735 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b587d444-217c-4ef1-b48e-b411f2c238a4" containerName="container-00" Dec 01 17:30:00 crc kubenswrapper[4810]: I1201 17:30:00.170752 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="b587d444-217c-4ef1-b48e-b411f2c238a4" containerName="container-00" Dec 01 17:30:00 crc kubenswrapper[4810]: I1201 17:30:00.171042 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="b587d444-217c-4ef1-b48e-b411f2c238a4" containerName="container-00" Dec 01 17:30:00 crc kubenswrapper[4810]: I1201 17:30:00.174073 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c" Dec 01 17:30:00 crc kubenswrapper[4810]: I1201 17:30:00.176337 4810 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 17:30:00 crc kubenswrapper[4810]: I1201 17:30:00.176589 4810 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 17:30:00 crc kubenswrapper[4810]: I1201 17:30:00.191940 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c"] Dec 01 17:30:00 crc kubenswrapper[4810]: I1201 17:30:00.285553 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7597c683-ad23-4d34-aaef-c1841982c7bd-secret-volume\") pod \"collect-profiles-29410170-6kg8c\" (UID: \"7597c683-ad23-4d34-aaef-c1841982c7bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c" Dec 01 17:30:00 crc kubenswrapper[4810]: I1201 17:30:00.285910 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn58s\" (UniqueName: \"kubernetes.io/projected/7597c683-ad23-4d34-aaef-c1841982c7bd-kube-api-access-sn58s\") pod \"collect-profiles-29410170-6kg8c\" (UID: \"7597c683-ad23-4d34-aaef-c1841982c7bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c" Dec 01 17:30:00 crc kubenswrapper[4810]: I1201 17:30:00.285967 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7597c683-ad23-4d34-aaef-c1841982c7bd-config-volume\") pod \"collect-profiles-29410170-6kg8c\" (UID: \"7597c683-ad23-4d34-aaef-c1841982c7bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c" Dec 01 17:30:00 crc kubenswrapper[4810]: I1201 17:30:00.387426 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn58s\" (UniqueName: \"kubernetes.io/projected/7597c683-ad23-4d34-aaef-c1841982c7bd-kube-api-access-sn58s\") pod \"collect-profiles-29410170-6kg8c\" (UID: \"7597c683-ad23-4d34-aaef-c1841982c7bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c" Dec 01 17:30:00 crc kubenswrapper[4810]: I1201 17:30:00.387567 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7597c683-ad23-4d34-aaef-c1841982c7bd-config-volume\") pod \"collect-profiles-29410170-6kg8c\" (UID: \"7597c683-ad23-4d34-aaef-c1841982c7bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c" Dec 01 17:30:00 crc kubenswrapper[4810]: I1201 17:30:00.387730 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7597c683-ad23-4d34-aaef-c1841982c7bd-secret-volume\") pod \"collect-profiles-29410170-6kg8c\" (UID: \"7597c683-ad23-4d34-aaef-c1841982c7bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c" Dec 01 17:30:00 crc kubenswrapper[4810]: I1201 17:30:00.388448 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7597c683-ad23-4d34-aaef-c1841982c7bd-config-volume\") pod \"collect-profiles-29410170-6kg8c\" (UID: \"7597c683-ad23-4d34-aaef-c1841982c7bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c" Dec 01 17:30:00 crc kubenswrapper[4810]: I1201 17:30:00.395466 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7597c683-ad23-4d34-aaef-c1841982c7bd-secret-volume\") pod \"collect-profiles-29410170-6kg8c\" (UID: \"7597c683-ad23-4d34-aaef-c1841982c7bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c" Dec 01 17:30:00 crc kubenswrapper[4810]: I1201 17:30:00.409280 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn58s\" (UniqueName: \"kubernetes.io/projected/7597c683-ad23-4d34-aaef-c1841982c7bd-kube-api-access-sn58s\") pod \"collect-profiles-29410170-6kg8c\" (UID: \"7597c683-ad23-4d34-aaef-c1841982c7bd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c" Dec 01 17:30:00 crc kubenswrapper[4810]: I1201 17:30:00.509051 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c" Dec 01 17:30:01 crc kubenswrapper[4810]: I1201 17:30:01.040488 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c"] Dec 01 17:30:01 crc kubenswrapper[4810]: I1201 17:30:01.156831 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c" event={"ID":"7597c683-ad23-4d34-aaef-c1841982c7bd","Type":"ContainerStarted","Data":"803f1431e3ef135e23fada7fc536615c08666cf5f6ea14de06f57942e6ff44fa"} Dec 01 17:30:02 crc kubenswrapper[4810]: I1201 17:30:02.171350 4810 generic.go:334] "Generic (PLEG): container finished" podID="7597c683-ad23-4d34-aaef-c1841982c7bd" containerID="aead48a932614fed026401fa278ddc54cf0da7d5d4b3e23eedd16865865d63d7" exitCode=0 Dec 01 17:30:02 crc kubenswrapper[4810]: I1201 17:30:02.171412 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c" event={"ID":"7597c683-ad23-4d34-aaef-c1841982c7bd","Type":"ContainerDied","Data":"aead48a932614fed026401fa278ddc54cf0da7d5d4b3e23eedd16865865d63d7"} Dec 01 17:30:03 crc kubenswrapper[4810]: I1201 17:30:03.638370 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c" Dec 01 17:30:03 crc kubenswrapper[4810]: I1201 17:30:03.754199 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7597c683-ad23-4d34-aaef-c1841982c7bd-secret-volume\") pod \"7597c683-ad23-4d34-aaef-c1841982c7bd\" (UID: \"7597c683-ad23-4d34-aaef-c1841982c7bd\") " Dec 01 17:30:03 crc kubenswrapper[4810]: I1201 17:30:03.754382 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sn58s\" (UniqueName: \"kubernetes.io/projected/7597c683-ad23-4d34-aaef-c1841982c7bd-kube-api-access-sn58s\") pod \"7597c683-ad23-4d34-aaef-c1841982c7bd\" (UID: \"7597c683-ad23-4d34-aaef-c1841982c7bd\") " Dec 01 17:30:03 crc kubenswrapper[4810]: I1201 17:30:03.754463 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7597c683-ad23-4d34-aaef-c1841982c7bd-config-volume\") pod \"7597c683-ad23-4d34-aaef-c1841982c7bd\" (UID: \"7597c683-ad23-4d34-aaef-c1841982c7bd\") " Dec 01 17:30:03 crc kubenswrapper[4810]: I1201 17:30:03.755313 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7597c683-ad23-4d34-aaef-c1841982c7bd-config-volume" (OuterVolumeSpecName: "config-volume") pod "7597c683-ad23-4d34-aaef-c1841982c7bd" (UID: "7597c683-ad23-4d34-aaef-c1841982c7bd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 17:30:03 crc kubenswrapper[4810]: I1201 17:30:03.767575 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7597c683-ad23-4d34-aaef-c1841982c7bd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7597c683-ad23-4d34-aaef-c1841982c7bd" (UID: "7597c683-ad23-4d34-aaef-c1841982c7bd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 17:30:03 crc kubenswrapper[4810]: I1201 17:30:03.768785 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7597c683-ad23-4d34-aaef-c1841982c7bd-kube-api-access-sn58s" (OuterVolumeSpecName: "kube-api-access-sn58s") pod "7597c683-ad23-4d34-aaef-c1841982c7bd" (UID: "7597c683-ad23-4d34-aaef-c1841982c7bd"). InnerVolumeSpecName "kube-api-access-sn58s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:30:03 crc kubenswrapper[4810]: I1201 17:30:03.857076 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sn58s\" (UniqueName: \"kubernetes.io/projected/7597c683-ad23-4d34-aaef-c1841982c7bd-kube-api-access-sn58s\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:03 crc kubenswrapper[4810]: I1201 17:30:03.857115 4810 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7597c683-ad23-4d34-aaef-c1841982c7bd-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:03 crc kubenswrapper[4810]: I1201 17:30:03.857124 4810 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7597c683-ad23-4d34-aaef-c1841982c7bd-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:04 crc kubenswrapper[4810]: I1201 17:30:04.193586 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c" event={"ID":"7597c683-ad23-4d34-aaef-c1841982c7bd","Type":"ContainerDied","Data":"803f1431e3ef135e23fada7fc536615c08666cf5f6ea14de06f57942e6ff44fa"} Dec 01 17:30:04 crc kubenswrapper[4810]: I1201 17:30:04.193628 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410170-6kg8c" Dec 01 17:30:04 crc kubenswrapper[4810]: I1201 17:30:04.193638 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="803f1431e3ef135e23fada7fc536615c08666cf5f6ea14de06f57942e6ff44fa" Dec 01 17:30:04 crc kubenswrapper[4810]: I1201 17:30:04.719030 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf"] Dec 01 17:30:04 crc kubenswrapper[4810]: I1201 17:30:04.729491 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410125-gwhhf"] Dec 01 17:30:05 crc kubenswrapper[4810]: I1201 17:30:05.069855 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-qxwtv_e16a3390-afd1-4812-b571-883e7c1515f7/nmstate-console-plugin/0.log" Dec 01 17:30:05 crc kubenswrapper[4810]: I1201 17:30:05.183864 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-n8r5h_3e4852d7-004b-462b-9b99-c9a13097292d/nmstate-handler/0.log" Dec 01 17:30:05 crc kubenswrapper[4810]: I1201 17:30:05.273367 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-hsl7k_c46e2836-e450-4375-bc15-3295c3772671/nmstate-metrics/0.log" Dec 01 17:30:05 crc kubenswrapper[4810]: I1201 17:30:05.275941 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-hsl7k_c46e2836-e450-4375-bc15-3295c3772671/kube-rbac-proxy/0.log" Dec 01 17:30:05 crc kubenswrapper[4810]: I1201 17:30:05.478538 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-6k5cn_33232e23-1ded-4abb-9df2-247452f3f6bf/nmstate-operator/0.log" Dec 01 17:30:05 crc kubenswrapper[4810]: I1201 17:30:05.562576 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-hzsts_3cda0965-f2f6-4a96-9e5a-91938cf20b0f/nmstate-webhook/0.log" Dec 01 17:30:06 crc kubenswrapper[4810]: I1201 17:30:06.515318 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00c1105f-2db1-4336-8452-93682dd15d0e" path="/var/lib/kubelet/pods/00c1105f-2db1-4336-8452-93682dd15d0e/volumes" Dec 01 17:30:07 crc kubenswrapper[4810]: I1201 17:30:07.490813 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:30:07 crc kubenswrapper[4810]: E1201 17:30:07.491630 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:30:19 crc kubenswrapper[4810]: I1201 17:30:19.102958 4810 scope.go:117] "RemoveContainer" containerID="7d28af347c9a678c5144d00b33c1dcd32ba3e0e5adf63dd1c92efbb4bc135c8f" Dec 01 17:30:20 crc kubenswrapper[4810]: I1201 17:30:20.491819 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:30:20 crc kubenswrapper[4810]: E1201 17:30:20.492424 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:30:22 crc kubenswrapper[4810]: I1201 17:30:22.884878 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-54tc5_49a2d6d2-847f-42bb-995d-54c4009c52ca/kube-rbac-proxy/0.log" Dec 01 17:30:23 crc kubenswrapper[4810]: I1201 17:30:23.116598 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-76dwm_c6232fb4-99db-4844-b364-4f1141abdbb4/cp-frr-files/0.log" Dec 01 17:30:23 crc kubenswrapper[4810]: I1201 17:30:23.241262 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-54tc5_49a2d6d2-847f-42bb-995d-54c4009c52ca/controller/0.log" Dec 01 17:30:23 crc kubenswrapper[4810]: I1201 17:30:23.356436 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-76dwm_c6232fb4-99db-4844-b364-4f1141abdbb4/cp-reloader/0.log" Dec 01 17:30:23 crc kubenswrapper[4810]: I1201 17:30:23.377548 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-76dwm_c6232fb4-99db-4844-b364-4f1141abdbb4/cp-frr-files/0.log" Dec 01 17:30:23 crc kubenswrapper[4810]: I1201 17:30:23.396957 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-76dwm_c6232fb4-99db-4844-b364-4f1141abdbb4/cp-metrics/0.log" Dec 01 17:30:23 crc kubenswrapper[4810]: I1201 17:30:23.445158 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-76dwm_c6232fb4-99db-4844-b364-4f1141abdbb4/cp-reloader/0.log" Dec 01 17:30:23 crc kubenswrapper[4810]: I1201 17:30:23.642976 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-76dwm_c6232fb4-99db-4844-b364-4f1141abdbb4/cp-reloader/0.log" Dec 01 17:30:23 crc kubenswrapper[4810]: I1201 17:30:23.665990 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-76dwm_c6232fb4-99db-4844-b364-4f1141abdbb4/cp-frr-files/0.log" Dec 01 17:30:23 crc kubenswrapper[4810]: I1201 17:30:23.666276 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-76dwm_c6232fb4-99db-4844-b364-4f1141abdbb4/cp-metrics/0.log" Dec 01 17:30:23 crc kubenswrapper[4810]: I1201 17:30:23.685038 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-76dwm_c6232fb4-99db-4844-b364-4f1141abdbb4/cp-metrics/0.log" Dec 01 17:30:23 crc kubenswrapper[4810]: I1201 17:30:23.840878 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-76dwm_c6232fb4-99db-4844-b364-4f1141abdbb4/cp-reloader/0.log" Dec 01 17:30:23 crc kubenswrapper[4810]: I1201 17:30:23.902715 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-76dwm_c6232fb4-99db-4844-b364-4f1141abdbb4/controller/0.log" Dec 01 17:30:23 crc kubenswrapper[4810]: I1201 17:30:23.923128 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-76dwm_c6232fb4-99db-4844-b364-4f1141abdbb4/cp-frr-files/0.log" Dec 01 17:30:23 crc kubenswrapper[4810]: I1201 17:30:23.933564 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-76dwm_c6232fb4-99db-4844-b364-4f1141abdbb4/cp-metrics/0.log" Dec 01 17:30:24 crc kubenswrapper[4810]: I1201 17:30:24.362716 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-76dwm_c6232fb4-99db-4844-b364-4f1141abdbb4/kube-rbac-proxy/0.log" Dec 01 17:30:24 crc kubenswrapper[4810]: I1201 17:30:24.366619 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-76dwm_c6232fb4-99db-4844-b364-4f1141abdbb4/frr-metrics/0.log" Dec 01 17:30:24 crc kubenswrapper[4810]: I1201 17:30:24.386235 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-76dwm_c6232fb4-99db-4844-b364-4f1141abdbb4/kube-rbac-proxy-frr/0.log" Dec 01 17:30:24 crc kubenswrapper[4810]: I1201 17:30:24.602914 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-sth2l_d774165e-4724-45f6-994e-979f8b8a5fdd/frr-k8s-webhook-server/0.log" Dec 01 17:30:24 crc kubenswrapper[4810]: I1201 17:30:24.608185 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-76dwm_c6232fb4-99db-4844-b364-4f1141abdbb4/reloader/0.log" Dec 01 17:30:24 crc kubenswrapper[4810]: I1201 17:30:24.851506 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-855bb66d65-ldd2w_6e8d7459-6378-4fa0-90c2-0533b2407764/manager/0.log" Dec 01 17:30:25 crc kubenswrapper[4810]: I1201 17:30:25.063897 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7c9798ff88-mt9tt_55a90415-842f-4a4f-b9ee-7932c3baca33/webhook-server/0.log" Dec 01 17:30:25 crc kubenswrapper[4810]: I1201 17:30:25.127161 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-4gcwv_d7514ada-83b0-48a3-a03b-07b2f346f607/kube-rbac-proxy/0.log" Dec 01 17:30:26 crc kubenswrapper[4810]: I1201 17:30:26.268676 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-4gcwv_d7514ada-83b0-48a3-a03b-07b2f346f607/speaker/0.log" Dec 01 17:30:27 crc kubenswrapper[4810]: I1201 17:30:27.580685 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-76dwm_c6232fb4-99db-4844-b364-4f1141abdbb4/frr/0.log" Dec 01 17:30:34 crc kubenswrapper[4810]: I1201 17:30:34.498544 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:30:34 crc kubenswrapper[4810]: E1201 17:30:34.499346 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:30:38 crc kubenswrapper[4810]: I1201 17:30:38.308558 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jbj4l"] Dec 01 17:30:38 crc kubenswrapper[4810]: E1201 17:30:38.309922 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7597c683-ad23-4d34-aaef-c1841982c7bd" containerName="collect-profiles" Dec 01 17:30:38 crc kubenswrapper[4810]: I1201 17:30:38.309936 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="7597c683-ad23-4d34-aaef-c1841982c7bd" containerName="collect-profiles" Dec 01 17:30:38 crc kubenswrapper[4810]: I1201 17:30:38.310146 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="7597c683-ad23-4d34-aaef-c1841982c7bd" containerName="collect-profiles" Dec 01 17:30:38 crc kubenswrapper[4810]: I1201 17:30:38.312357 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbj4l" Dec 01 17:30:38 crc kubenswrapper[4810]: I1201 17:30:38.318600 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jbj4l"] Dec 01 17:30:38 crc kubenswrapper[4810]: I1201 17:30:38.441940 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c-utilities\") pod \"redhat-operators-jbj4l\" (UID: \"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c\") " pod="openshift-marketplace/redhat-operators-jbj4l" Dec 01 17:30:38 crc kubenswrapper[4810]: I1201 17:30:38.442269 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqtgq\" (UniqueName: \"kubernetes.io/projected/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c-kube-api-access-jqtgq\") pod \"redhat-operators-jbj4l\" (UID: \"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c\") " pod="openshift-marketplace/redhat-operators-jbj4l" Dec 01 17:30:38 crc kubenswrapper[4810]: I1201 17:30:38.442339 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c-catalog-content\") pod \"redhat-operators-jbj4l\" (UID: \"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c\") " pod="openshift-marketplace/redhat-operators-jbj4l" Dec 01 17:30:38 crc kubenswrapper[4810]: I1201 17:30:38.544334 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c-utilities\") pod \"redhat-operators-jbj4l\" (UID: \"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c\") " pod="openshift-marketplace/redhat-operators-jbj4l" Dec 01 17:30:38 crc kubenswrapper[4810]: I1201 17:30:38.544395 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqtgq\" (UniqueName: \"kubernetes.io/projected/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c-kube-api-access-jqtgq\") pod \"redhat-operators-jbj4l\" (UID: \"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c\") " pod="openshift-marketplace/redhat-operators-jbj4l" Dec 01 17:30:38 crc kubenswrapper[4810]: I1201 17:30:38.544485 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c-catalog-content\") pod \"redhat-operators-jbj4l\" (UID: \"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c\") " pod="openshift-marketplace/redhat-operators-jbj4l" Dec 01 17:30:38 crc kubenswrapper[4810]: I1201 17:30:38.544913 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c-catalog-content\") pod \"redhat-operators-jbj4l\" (UID: \"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c\") " pod="openshift-marketplace/redhat-operators-jbj4l" Dec 01 17:30:38 crc kubenswrapper[4810]: I1201 17:30:38.544950 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c-utilities\") pod \"redhat-operators-jbj4l\" (UID: \"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c\") " pod="openshift-marketplace/redhat-operators-jbj4l" Dec 01 17:30:38 crc kubenswrapper[4810]: I1201 17:30:38.572016 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqtgq\" (UniqueName: \"kubernetes.io/projected/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c-kube-api-access-jqtgq\") pod \"redhat-operators-jbj4l\" (UID: \"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c\") " pod="openshift-marketplace/redhat-operators-jbj4l" Dec 01 17:30:38 crc kubenswrapper[4810]: I1201 17:30:38.651865 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbj4l" Dec 01 17:30:39 crc kubenswrapper[4810]: I1201 17:30:39.221610 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jbj4l"] Dec 01 17:30:39 crc kubenswrapper[4810]: I1201 17:30:39.566961 4810 generic.go:334] "Generic (PLEG): container finished" podID="23cddec1-85c4-4d5c-8dfe-eeb4b17f375c" containerID="04629219c6314028f8f9b163577cf75ec16303678fb2d660ec473f69fcccad28" exitCode=0 Dec 01 17:30:39 crc kubenswrapper[4810]: I1201 17:30:39.567508 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbj4l" event={"ID":"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c","Type":"ContainerDied","Data":"04629219c6314028f8f9b163577cf75ec16303678fb2d660ec473f69fcccad28"} Dec 01 17:30:39 crc kubenswrapper[4810]: I1201 17:30:39.567590 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbj4l" event={"ID":"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c","Type":"ContainerStarted","Data":"fbaad6897b55980288fe880e37f79074ad9ec678b2dad413b99c28995eb172a1"} Dec 01 17:30:39 crc kubenswrapper[4810]: I1201 17:30:39.569349 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 17:30:40 crc kubenswrapper[4810]: I1201 17:30:40.118073 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq_2552822d-ff62-49fa-8ba0-321952bda208/util/0.log" Dec 01 17:30:40 crc kubenswrapper[4810]: I1201 17:30:40.403283 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq_2552822d-ff62-49fa-8ba0-321952bda208/pull/0.log" Dec 01 17:30:40 crc kubenswrapper[4810]: I1201 17:30:40.419893 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq_2552822d-ff62-49fa-8ba0-321952bda208/pull/0.log" Dec 01 17:30:40 crc kubenswrapper[4810]: I1201 17:30:40.550206 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq_2552822d-ff62-49fa-8ba0-321952bda208/util/0.log" Dec 01 17:30:40 crc kubenswrapper[4810]: I1201 17:30:40.563992 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq_2552822d-ff62-49fa-8ba0-321952bda208/util/0.log" Dec 01 17:30:40 crc kubenswrapper[4810]: I1201 17:30:40.620575 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq_2552822d-ff62-49fa-8ba0-321952bda208/extract/0.log" Dec 01 17:30:40 crc kubenswrapper[4810]: I1201 17:30:40.681097 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a6hgjq_2552822d-ff62-49fa-8ba0-321952bda208/pull/0.log" Dec 01 17:30:40 crc kubenswrapper[4810]: I1201 17:30:40.771816 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz_b0ed0930-4327-4772-9a1c-2a0ae55f529c/util/0.log" Dec 01 17:30:41 crc kubenswrapper[4810]: I1201 17:30:41.039636 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz_b0ed0930-4327-4772-9a1c-2a0ae55f529c/pull/0.log" Dec 01 17:30:41 crc kubenswrapper[4810]: I1201 17:30:41.250931 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz_b0ed0930-4327-4772-9a1c-2a0ae55f529c/pull/0.log" Dec 01 17:30:41 crc kubenswrapper[4810]: I1201 17:30:41.263013 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz_b0ed0930-4327-4772-9a1c-2a0ae55f529c/util/0.log" Dec 01 17:30:41 crc kubenswrapper[4810]: I1201 17:30:41.462630 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz_b0ed0930-4327-4772-9a1c-2a0ae55f529c/extract/0.log" Dec 01 17:30:41 crc kubenswrapper[4810]: I1201 17:30:41.471096 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz_b0ed0930-4327-4772-9a1c-2a0ae55f529c/util/0.log" Dec 01 17:30:41 crc kubenswrapper[4810]: I1201 17:30:41.513017 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fs4lpz_b0ed0930-4327-4772-9a1c-2a0ae55f529c/pull/0.log" Dec 01 17:30:41 crc kubenswrapper[4810]: I1201 17:30:41.590210 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbj4l" event={"ID":"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c","Type":"ContainerStarted","Data":"dff898b1b52f324955141745d54d13cc9365265f4ee363f45a78ed6a33070d20"} Dec 01 17:30:41 crc kubenswrapper[4810]: I1201 17:30:41.655860 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x_135d31a5-5184-4906-8c27-9f8d645ea84c/util/0.log" Dec 01 17:30:41 crc kubenswrapper[4810]: I1201 17:30:41.914393 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x_135d31a5-5184-4906-8c27-9f8d645ea84c/util/0.log" Dec 01 17:30:41 crc kubenswrapper[4810]: I1201 17:30:41.970363 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x_135d31a5-5184-4906-8c27-9f8d645ea84c/pull/0.log" Dec 01 17:30:41 crc kubenswrapper[4810]: I1201 17:30:41.970607 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x_135d31a5-5184-4906-8c27-9f8d645ea84c/pull/0.log" Dec 01 17:30:42 crc kubenswrapper[4810]: I1201 17:30:42.077689 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x_135d31a5-5184-4906-8c27-9f8d645ea84c/pull/0.log" Dec 01 17:30:42 crc kubenswrapper[4810]: I1201 17:30:42.104503 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x_135d31a5-5184-4906-8c27-9f8d645ea84c/util/0.log" Dec 01 17:30:42 crc kubenswrapper[4810]: I1201 17:30:42.215983 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210spq6x_135d31a5-5184-4906-8c27-9f8d645ea84c/extract/0.log" Dec 01 17:30:42 crc kubenswrapper[4810]: I1201 17:30:42.267407 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv_673fd73e-0701-4cf7-a7dd-273db851c2ba/util/0.log" Dec 01 17:30:42 crc kubenswrapper[4810]: I1201 17:30:42.571701 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv_673fd73e-0701-4cf7-a7dd-273db851c2ba/util/0.log" Dec 01 17:30:42 crc kubenswrapper[4810]: I1201 17:30:42.619400 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv_673fd73e-0701-4cf7-a7dd-273db851c2ba/pull/0.log" Dec 01 17:30:42 crc kubenswrapper[4810]: I1201 17:30:42.619579 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv_673fd73e-0701-4cf7-a7dd-273db851c2ba/pull/0.log" Dec 01 17:30:42 crc kubenswrapper[4810]: I1201 17:30:42.694742 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv_673fd73e-0701-4cf7-a7dd-273db851c2ba/util/0.log" Dec 01 17:30:42 crc kubenswrapper[4810]: I1201 17:30:42.838428 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv_673fd73e-0701-4cf7-a7dd-273db851c2ba/extract/0.log" Dec 01 17:30:42 crc kubenswrapper[4810]: I1201 17:30:42.876645 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mn97w_ba60559e-fda1-4366-b7ee-eabb2b6c2324/extract-utilities/0.log" Dec 01 17:30:42 crc kubenswrapper[4810]: I1201 17:30:42.928672 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f836flpv_673fd73e-0701-4cf7-a7dd-273db851c2ba/pull/0.log" Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.135807 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mn97w_ba60559e-fda1-4366-b7ee-eabb2b6c2324/extract-content/0.log" Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.174793 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mn97w_ba60559e-fda1-4366-b7ee-eabb2b6c2324/extract-content/0.log" Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.176132 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mn97w_ba60559e-fda1-4366-b7ee-eabb2b6c2324/extract-utilities/0.log" Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.287899 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cpblh"] Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.299715 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cpblh" Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.307514 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cpblh"] Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.472798 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l4kd\" (UniqueName: \"kubernetes.io/projected/319f7ef2-d70b-4351-b776-33b68596960f-kube-api-access-7l4kd\") pod \"community-operators-cpblh\" (UID: \"319f7ef2-d70b-4351-b776-33b68596960f\") " pod="openshift-marketplace/community-operators-cpblh" Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.472883 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/319f7ef2-d70b-4351-b776-33b68596960f-catalog-content\") pod \"community-operators-cpblh\" (UID: \"319f7ef2-d70b-4351-b776-33b68596960f\") " pod="openshift-marketplace/community-operators-cpblh" Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.472906 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/319f7ef2-d70b-4351-b776-33b68596960f-utilities\") pod \"community-operators-cpblh\" (UID: \"319f7ef2-d70b-4351-b776-33b68596960f\") " pod="openshift-marketplace/community-operators-cpblh" Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.499809 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mn97w_ba60559e-fda1-4366-b7ee-eabb2b6c2324/extract-content/0.log" Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.526578 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mn97w_ba60559e-fda1-4366-b7ee-eabb2b6c2324/extract-utilities/0.log" Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.575390 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l4kd\" (UniqueName: \"kubernetes.io/projected/319f7ef2-d70b-4351-b776-33b68596960f-kube-api-access-7l4kd\") pod \"community-operators-cpblh\" (UID: \"319f7ef2-d70b-4351-b776-33b68596960f\") " pod="openshift-marketplace/community-operators-cpblh" Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.575771 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/319f7ef2-d70b-4351-b776-33b68596960f-catalog-content\") pod \"community-operators-cpblh\" (UID: \"319f7ef2-d70b-4351-b776-33b68596960f\") " pod="openshift-marketplace/community-operators-cpblh" Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.575794 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/319f7ef2-d70b-4351-b776-33b68596960f-utilities\") pod \"community-operators-cpblh\" (UID: \"319f7ef2-d70b-4351-b776-33b68596960f\") " pod="openshift-marketplace/community-operators-cpblh" Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.576273 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/319f7ef2-d70b-4351-b776-33b68596960f-utilities\") pod \"community-operators-cpblh\" (UID: \"319f7ef2-d70b-4351-b776-33b68596960f\") " pod="openshift-marketplace/community-operators-cpblh" Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.577562 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/319f7ef2-d70b-4351-b776-33b68596960f-catalog-content\") pod \"community-operators-cpblh\" (UID: \"319f7ef2-d70b-4351-b776-33b68596960f\") " pod="openshift-marketplace/community-operators-cpblh" Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.692108 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l4kd\" (UniqueName: \"kubernetes.io/projected/319f7ef2-d70b-4351-b776-33b68596960f-kube-api-access-7l4kd\") pod \"community-operators-cpblh\" (UID: \"319f7ef2-d70b-4351-b776-33b68596960f\") " pod="openshift-marketplace/community-operators-cpblh" Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.699018 4810 generic.go:334] "Generic (PLEG): container finished" podID="23cddec1-85c4-4d5c-8dfe-eeb4b17f375c" containerID="dff898b1b52f324955141745d54d13cc9365265f4ee363f45a78ed6a33070d20" exitCode=0 Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.699055 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbj4l" event={"ID":"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c","Type":"ContainerDied","Data":"dff898b1b52f324955141745d54d13cc9365265f4ee363f45a78ed6a33070d20"} Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.726651 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cpblh" Dec 01 17:30:43 crc kubenswrapper[4810]: I1201 17:30:43.846406 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-th5zw_e62dc78d-0d09-4b02-a206-30b306bb4807/extract-utilities/0.log" Dec 01 17:30:44 crc kubenswrapper[4810]: I1201 17:30:44.128968 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mn97w_ba60559e-fda1-4366-b7ee-eabb2b6c2324/registry-server/0.log" Dec 01 17:30:44 crc kubenswrapper[4810]: I1201 17:30:44.277557 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-th5zw_e62dc78d-0d09-4b02-a206-30b306bb4807/extract-content/0.log" Dec 01 17:30:44 crc kubenswrapper[4810]: I1201 17:30:44.282607 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-th5zw_e62dc78d-0d09-4b02-a206-30b306bb4807/extract-utilities/0.log" Dec 01 17:30:44 crc kubenswrapper[4810]: I1201 17:30:44.287009 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-th5zw_e62dc78d-0d09-4b02-a206-30b306bb4807/extract-content/0.log" Dec 01 17:30:44 crc kubenswrapper[4810]: I1201 17:30:44.471698 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-th5zw_e62dc78d-0d09-4b02-a206-30b306bb4807/extract-content/0.log" Dec 01 17:30:44 crc kubenswrapper[4810]: I1201 17:30:44.473923 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-th5zw_e62dc78d-0d09-4b02-a206-30b306bb4807/extract-utilities/0.log" Dec 01 17:30:44 crc kubenswrapper[4810]: I1201 17:30:44.558338 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-8ss9b_97b03cd3-84c4-4c13-b5cd-65f2ad45cb6b/marketplace-operator/0.log" Dec 01 17:30:44 crc kubenswrapper[4810]: I1201 17:30:44.593695 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cpblh"] Dec 01 17:30:44 crc kubenswrapper[4810]: I1201 17:30:44.704959 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmlkh_8c8f2b09-969f-4adf-8f45-8468d57e76be/extract-utilities/0.log" Dec 01 17:30:44 crc kubenswrapper[4810]: I1201 17:30:44.715905 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpblh" event={"ID":"319f7ef2-d70b-4351-b776-33b68596960f","Type":"ContainerStarted","Data":"176b1234deb9342661387d8600d48be2efb8738678e5a0bdeb57acdbd5534557"} Dec 01 17:30:44 crc kubenswrapper[4810]: I1201 17:30:44.739707 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-th5zw_e62dc78d-0d09-4b02-a206-30b306bb4807/registry-server/0.log" Dec 01 17:30:44 crc kubenswrapper[4810]: I1201 17:30:44.851080 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmlkh_8c8f2b09-969f-4adf-8f45-8468d57e76be/extract-utilities/0.log" Dec 01 17:30:44 crc kubenswrapper[4810]: I1201 17:30:44.860111 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmlkh_8c8f2b09-969f-4adf-8f45-8468d57e76be/extract-content/0.log" Dec 01 17:30:44 crc kubenswrapper[4810]: I1201 17:30:44.910109 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmlkh_8c8f2b09-969f-4adf-8f45-8468d57e76be/extract-content/0.log" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.091613 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2m2rp"] Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.093956 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2m2rp" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.106118 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2m2rp"] Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.135735 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmlkh_8c8f2b09-969f-4adf-8f45-8468d57e76be/extract-utilities/0.log" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.147089 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmlkh_8c8f2b09-969f-4adf-8f45-8468d57e76be/extract-content/0.log" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.179941 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7h8nl_0341c875-f1fa-4a99-b850-202a0f3385db/extract-utilities/0.log" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.218961 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9vrt\" (UniqueName: \"kubernetes.io/projected/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220-kube-api-access-n9vrt\") pod \"redhat-marketplace-2m2rp\" (UID: \"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220\") " pod="openshift-marketplace/redhat-marketplace-2m2rp" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.219058 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220-utilities\") pod \"redhat-marketplace-2m2rp\" (UID: \"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220\") " pod="openshift-marketplace/redhat-marketplace-2m2rp" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.219085 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220-catalog-content\") pod \"redhat-marketplace-2m2rp\" (UID: \"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220\") " pod="openshift-marketplace/redhat-marketplace-2m2rp" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.320780 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9vrt\" (UniqueName: \"kubernetes.io/projected/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220-kube-api-access-n9vrt\") pod \"redhat-marketplace-2m2rp\" (UID: \"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220\") " pod="openshift-marketplace/redhat-marketplace-2m2rp" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.320876 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220-utilities\") pod \"redhat-marketplace-2m2rp\" (UID: \"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220\") " pod="openshift-marketplace/redhat-marketplace-2m2rp" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.320899 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220-catalog-content\") pod \"redhat-marketplace-2m2rp\" (UID: \"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220\") " pod="openshift-marketplace/redhat-marketplace-2m2rp" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.321326 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220-catalog-content\") pod \"redhat-marketplace-2m2rp\" (UID: \"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220\") " pod="openshift-marketplace/redhat-marketplace-2m2rp" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.321524 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220-utilities\") pod \"redhat-marketplace-2m2rp\" (UID: \"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220\") " pod="openshift-marketplace/redhat-marketplace-2m2rp" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.362306 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9vrt\" (UniqueName: \"kubernetes.io/projected/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220-kube-api-access-n9vrt\") pod \"redhat-marketplace-2m2rp\" (UID: \"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220\") " pod="openshift-marketplace/redhat-marketplace-2m2rp" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.411679 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2m2rp" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.521740 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jmlkh_8c8f2b09-969f-4adf-8f45-8468d57e76be/registry-server/0.log" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.534587 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7h8nl_0341c875-f1fa-4a99-b850-202a0f3385db/extract-content/0.log" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.575620 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7h8nl_0341c875-f1fa-4a99-b850-202a0f3385db/extract-utilities/0.log" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.668398 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7h8nl_0341c875-f1fa-4a99-b850-202a0f3385db/extract-content/0.log" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.725543 4810 generic.go:334] "Generic (PLEG): container finished" podID="319f7ef2-d70b-4351-b776-33b68596960f" containerID="3a02ebc51554b3d87b10c6e2929329b43b1c6a61d5e7b29189d9f38178d6e44f" exitCode=0 Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.725617 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpblh" event={"ID":"319f7ef2-d70b-4351-b776-33b68596960f","Type":"ContainerDied","Data":"3a02ebc51554b3d87b10c6e2929329b43b1c6a61d5e7b29189d9f38178d6e44f"} Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.740399 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbj4l" event={"ID":"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c","Type":"ContainerStarted","Data":"0982b226f47cafb10c65d344c61a850a4b853d7b94af8f8f448a24e9f2e15741"} Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.755323 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7h8nl_0341c875-f1fa-4a99-b850-202a0f3385db/extract-utilities/0.log" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.781572 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jbj4l" podStartSLOduration=2.74125337 podStartE2EDuration="7.781552282s" podCreationTimestamp="2025-12-01 17:30:38 +0000 UTC" firstStartedPulling="2025-12-01 17:30:39.568903943 +0000 UTC m=+10605.332413546" lastFinishedPulling="2025-12-01 17:30:44.609202855 +0000 UTC m=+10610.372712458" observedRunningTime="2025-12-01 17:30:45.771868922 +0000 UTC m=+10611.535378525" watchObservedRunningTime="2025-12-01 17:30:45.781552282 +0000 UTC m=+10611.545061895" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.788444 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7h8nl_0341c875-f1fa-4a99-b850-202a0f3385db/extract-content/0.log" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.930519 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbj4l_23cddec1-85c4-4d5c-8dfe-eeb4b17f375c/extract-utilities/0.log" Dec 01 17:30:45 crc kubenswrapper[4810]: I1201 17:30:45.960199 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2m2rp"] Dec 01 17:30:46 crc kubenswrapper[4810]: I1201 17:30:46.180350 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbj4l_23cddec1-85c4-4d5c-8dfe-eeb4b17f375c/extract-content/0.log" Dec 01 17:30:46 crc kubenswrapper[4810]: I1201 17:30:46.241920 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbj4l_23cddec1-85c4-4d5c-8dfe-eeb4b17f375c/extract-utilities/0.log" Dec 01 17:30:46 crc kubenswrapper[4810]: I1201 17:30:46.299012 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbj4l_23cddec1-85c4-4d5c-8dfe-eeb4b17f375c/extract-content/0.log" Dec 01 17:30:46 crc kubenswrapper[4810]: I1201 17:30:46.462967 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7h8nl_0341c875-f1fa-4a99-b850-202a0f3385db/registry-server/0.log" Dec 01 17:30:46 crc kubenswrapper[4810]: I1201 17:30:46.491520 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:30:46 crc kubenswrapper[4810]: E1201 17:30:46.491890 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:30:46 crc kubenswrapper[4810]: I1201 17:30:46.542206 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbj4l_23cddec1-85c4-4d5c-8dfe-eeb4b17f375c/extract-content/0.log" Dec 01 17:30:46 crc kubenswrapper[4810]: I1201 17:30:46.601125 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbj4l_23cddec1-85c4-4d5c-8dfe-eeb4b17f375c/extract-utilities/0.log" Dec 01 17:30:46 crc kubenswrapper[4810]: I1201 17:30:46.607298 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jbj4l_23cddec1-85c4-4d5c-8dfe-eeb4b17f375c/registry-server/0.log" Dec 01 17:30:46 crc kubenswrapper[4810]: I1201 17:30:46.752281 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpblh" event={"ID":"319f7ef2-d70b-4351-b776-33b68596960f","Type":"ContainerStarted","Data":"0a02c0f038c4419644ea7110d0b4d5e6a6a69a97bc3da8ede422949f5976d7e9"} Dec 01 17:30:46 crc kubenswrapper[4810]: I1201 17:30:46.754312 4810 generic.go:334] "Generic (PLEG): container finished" podID="d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220" containerID="42e803fbeffb46eeddee8f4ed84b31d1692be9bfb02d9378b454086755f932af" exitCode=0 Dec 01 17:30:46 crc kubenswrapper[4810]: I1201 17:30:46.754353 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2m2rp" event={"ID":"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220","Type":"ContainerDied","Data":"42e803fbeffb46eeddee8f4ed84b31d1692be9bfb02d9378b454086755f932af"} Dec 01 17:30:46 crc kubenswrapper[4810]: I1201 17:30:46.754377 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2m2rp" event={"ID":"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220","Type":"ContainerStarted","Data":"34dc341561a2b74712ef3ec653abc4096b1cbdd80662e2058b51f8c35f1f6d68"} Dec 01 17:30:47 crc kubenswrapper[4810]: I1201 17:30:47.765026 4810 generic.go:334] "Generic (PLEG): container finished" podID="319f7ef2-d70b-4351-b776-33b68596960f" containerID="0a02c0f038c4419644ea7110d0b4d5e6a6a69a97bc3da8ede422949f5976d7e9" exitCode=0 Dec 01 17:30:47 crc kubenswrapper[4810]: I1201 17:30:47.765130 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpblh" event={"ID":"319f7ef2-d70b-4351-b776-33b68596960f","Type":"ContainerDied","Data":"0a02c0f038c4419644ea7110d0b4d5e6a6a69a97bc3da8ede422949f5976d7e9"} Dec 01 17:30:48 crc kubenswrapper[4810]: I1201 17:30:48.653634 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jbj4l" Dec 01 17:30:48 crc kubenswrapper[4810]: I1201 17:30:48.654252 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jbj4l" Dec 01 17:30:48 crc kubenswrapper[4810]: I1201 17:30:48.779876 4810 generic.go:334] "Generic (PLEG): container finished" podID="d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220" containerID="2d1f85ffa9f6740f958e98c07852ec7eb75a61da06f4f702342b53f8d28ee5dc" exitCode=0 Dec 01 17:30:48 crc kubenswrapper[4810]: I1201 17:30:48.779922 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2m2rp" event={"ID":"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220","Type":"ContainerDied","Data":"2d1f85ffa9f6740f958e98c07852ec7eb75a61da06f4f702342b53f8d28ee5dc"} Dec 01 17:30:49 crc kubenswrapper[4810]: I1201 17:30:49.700919 4810 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jbj4l" podUID="23cddec1-85c4-4d5c-8dfe-eeb4b17f375c" containerName="registry-server" probeResult="failure" output=< Dec 01 17:30:49 crc kubenswrapper[4810]: timeout: failed to connect service ":50051" within 1s Dec 01 17:30:49 crc kubenswrapper[4810]: > Dec 01 17:30:49 crc kubenswrapper[4810]: I1201 17:30:49.793186 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpblh" event={"ID":"319f7ef2-d70b-4351-b776-33b68596960f","Type":"ContainerStarted","Data":"79b736a7cd78f6677ac990e369202fef265f65aa5ca232a6badff3bfcd2de287"} Dec 01 17:30:49 crc kubenswrapper[4810]: I1201 17:30:49.819025 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cpblh" podStartSLOduration=3.923995765 podStartE2EDuration="6.819002274s" podCreationTimestamp="2025-12-01 17:30:43 +0000 UTC" firstStartedPulling="2025-12-01 17:30:45.72940627 +0000 UTC m=+10611.492915873" lastFinishedPulling="2025-12-01 17:30:48.624412779 +0000 UTC m=+10614.387922382" observedRunningTime="2025-12-01 17:30:49.809597232 +0000 UTC m=+10615.573106845" watchObservedRunningTime="2025-12-01 17:30:49.819002274 +0000 UTC m=+10615.582511887" Dec 01 17:30:50 crc kubenswrapper[4810]: I1201 17:30:50.808342 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2m2rp" event={"ID":"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220","Type":"ContainerStarted","Data":"2888ad99862db498d2ff26a312cdb4dfea02d835d8850659c008194cfda9be03"} Dec 01 17:30:50 crc kubenswrapper[4810]: I1201 17:30:50.835873 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2m2rp" podStartSLOduration=3.2664650650000002 podStartE2EDuration="5.835847701s" podCreationTimestamp="2025-12-01 17:30:45 +0000 UTC" firstStartedPulling="2025-12-01 17:30:46.756177013 +0000 UTC m=+10612.519686616" lastFinishedPulling="2025-12-01 17:30:49.325559649 +0000 UTC m=+10615.089069252" observedRunningTime="2025-12-01 17:30:50.827769754 +0000 UTC m=+10616.591279367" watchObservedRunningTime="2025-12-01 17:30:50.835847701 +0000 UTC m=+10616.599357304" Dec 01 17:30:53 crc kubenswrapper[4810]: I1201 17:30:53.727779 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cpblh" Dec 01 17:30:53 crc kubenswrapper[4810]: I1201 17:30:53.728150 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cpblh" Dec 01 17:30:53 crc kubenswrapper[4810]: I1201 17:30:53.814149 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cpblh" Dec 01 17:30:55 crc kubenswrapper[4810]: I1201 17:30:55.412239 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2m2rp" Dec 01 17:30:55 crc kubenswrapper[4810]: I1201 17:30:55.412625 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2m2rp" Dec 01 17:30:55 crc kubenswrapper[4810]: I1201 17:30:55.471868 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2m2rp" Dec 01 17:30:55 crc kubenswrapper[4810]: I1201 17:30:55.916067 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2m2rp" Dec 01 17:30:55 crc kubenswrapper[4810]: I1201 17:30:55.980145 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2m2rp"] Dec 01 17:30:57 crc kubenswrapper[4810]: I1201 17:30:57.873925 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2m2rp" podUID="d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220" containerName="registry-server" containerID="cri-o://2888ad99862db498d2ff26a312cdb4dfea02d835d8850659c008194cfda9be03" gracePeriod=2 Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.542664 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2m2rp" Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.620450 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220-catalog-content\") pod \"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220\" (UID: \"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220\") " Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.620583 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220-utilities\") pod \"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220\" (UID: \"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220\") " Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.620708 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9vrt\" (UniqueName: \"kubernetes.io/projected/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220-kube-api-access-n9vrt\") pod \"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220\" (UID: \"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220\") " Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.621718 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220-utilities" (OuterVolumeSpecName: "utilities") pod "d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220" (UID: "d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.630986 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220-kube-api-access-n9vrt" (OuterVolumeSpecName: "kube-api-access-n9vrt") pod "d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220" (UID: "d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220"). InnerVolumeSpecName "kube-api-access-n9vrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.645005 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220" (UID: "d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.723427 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.723458 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.723481 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9vrt\" (UniqueName: \"kubernetes.io/projected/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220-kube-api-access-n9vrt\") on node \"crc\" DevicePath \"\"" Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.724666 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jbj4l" Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.785645 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jbj4l" Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.887199 4810 generic.go:334] "Generic (PLEG): container finished" podID="d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220" containerID="2888ad99862db498d2ff26a312cdb4dfea02d835d8850659c008194cfda9be03" exitCode=0 Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.887284 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2m2rp" Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.887283 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2m2rp" event={"ID":"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220","Type":"ContainerDied","Data":"2888ad99862db498d2ff26a312cdb4dfea02d835d8850659c008194cfda9be03"} Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.887376 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2m2rp" event={"ID":"d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220","Type":"ContainerDied","Data":"34dc341561a2b74712ef3ec653abc4096b1cbdd80662e2058b51f8c35f1f6d68"} Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.887416 4810 scope.go:117] "RemoveContainer" containerID="2888ad99862db498d2ff26a312cdb4dfea02d835d8850659c008194cfda9be03" Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.955285 4810 scope.go:117] "RemoveContainer" containerID="2d1f85ffa9f6740f958e98c07852ec7eb75a61da06f4f702342b53f8d28ee5dc" Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.955788 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2m2rp"] Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.979613 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2m2rp"] Dec 01 17:30:58 crc kubenswrapper[4810]: I1201 17:30:58.988924 4810 scope.go:117] "RemoveContainer" containerID="42e803fbeffb46eeddee8f4ed84b31d1692be9bfb02d9378b454086755f932af" Dec 01 17:30:59 crc kubenswrapper[4810]: I1201 17:30:59.062399 4810 scope.go:117] "RemoveContainer" containerID="2888ad99862db498d2ff26a312cdb4dfea02d835d8850659c008194cfda9be03" Dec 01 17:30:59 crc kubenswrapper[4810]: E1201 17:30:59.063022 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2888ad99862db498d2ff26a312cdb4dfea02d835d8850659c008194cfda9be03\": container with ID starting with 2888ad99862db498d2ff26a312cdb4dfea02d835d8850659c008194cfda9be03 not found: ID does not exist" containerID="2888ad99862db498d2ff26a312cdb4dfea02d835d8850659c008194cfda9be03" Dec 01 17:30:59 crc kubenswrapper[4810]: I1201 17:30:59.063076 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2888ad99862db498d2ff26a312cdb4dfea02d835d8850659c008194cfda9be03"} err="failed to get container status \"2888ad99862db498d2ff26a312cdb4dfea02d835d8850659c008194cfda9be03\": rpc error: code = NotFound desc = could not find container \"2888ad99862db498d2ff26a312cdb4dfea02d835d8850659c008194cfda9be03\": container with ID starting with 2888ad99862db498d2ff26a312cdb4dfea02d835d8850659c008194cfda9be03 not found: ID does not exist" Dec 01 17:30:59 crc kubenswrapper[4810]: I1201 17:30:59.063107 4810 scope.go:117] "RemoveContainer" containerID="2d1f85ffa9f6740f958e98c07852ec7eb75a61da06f4f702342b53f8d28ee5dc" Dec 01 17:30:59 crc kubenswrapper[4810]: E1201 17:30:59.063356 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d1f85ffa9f6740f958e98c07852ec7eb75a61da06f4f702342b53f8d28ee5dc\": container with ID starting with 2d1f85ffa9f6740f958e98c07852ec7eb75a61da06f4f702342b53f8d28ee5dc not found: ID does not exist" containerID="2d1f85ffa9f6740f958e98c07852ec7eb75a61da06f4f702342b53f8d28ee5dc" Dec 01 17:30:59 crc kubenswrapper[4810]: I1201 17:30:59.063396 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d1f85ffa9f6740f958e98c07852ec7eb75a61da06f4f702342b53f8d28ee5dc"} err="failed to get container status \"2d1f85ffa9f6740f958e98c07852ec7eb75a61da06f4f702342b53f8d28ee5dc\": rpc error: code = NotFound desc = could not find container \"2d1f85ffa9f6740f958e98c07852ec7eb75a61da06f4f702342b53f8d28ee5dc\": container with ID starting with 2d1f85ffa9f6740f958e98c07852ec7eb75a61da06f4f702342b53f8d28ee5dc not found: ID does not exist" Dec 01 17:30:59 crc kubenswrapper[4810]: I1201 17:30:59.063428 4810 scope.go:117] "RemoveContainer" containerID="42e803fbeffb46eeddee8f4ed84b31d1692be9bfb02d9378b454086755f932af" Dec 01 17:30:59 crc kubenswrapper[4810]: E1201 17:30:59.063682 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42e803fbeffb46eeddee8f4ed84b31d1692be9bfb02d9378b454086755f932af\": container with ID starting with 42e803fbeffb46eeddee8f4ed84b31d1692be9bfb02d9378b454086755f932af not found: ID does not exist" containerID="42e803fbeffb46eeddee8f4ed84b31d1692be9bfb02d9378b454086755f932af" Dec 01 17:30:59 crc kubenswrapper[4810]: I1201 17:30:59.063710 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42e803fbeffb46eeddee8f4ed84b31d1692be9bfb02d9378b454086755f932af"} err="failed to get container status \"42e803fbeffb46eeddee8f4ed84b31d1692be9bfb02d9378b454086755f932af\": rpc error: code = NotFound desc = could not find container \"42e803fbeffb46eeddee8f4ed84b31d1692be9bfb02d9378b454086755f932af\": container with ID starting with 42e803fbeffb46eeddee8f4ed84b31d1692be9bfb02d9378b454086755f932af not found: ID does not exist" Dec 01 17:30:59 crc kubenswrapper[4810]: I1201 17:30:59.494112 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jbj4l"] Dec 01 17:30:59 crc kubenswrapper[4810]: I1201 17:30:59.899005 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jbj4l" podUID="23cddec1-85c4-4d5c-8dfe-eeb4b17f375c" containerName="registry-server" containerID="cri-o://0982b226f47cafb10c65d344c61a850a4b853d7b94af8f8f448a24e9f2e15741" gracePeriod=2 Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.017812 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-kx7vb_56425059-45e3-4f39-8e68-92d8ffd61831/prometheus-operator/0.log" Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.242126 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7769c7475b-pdxbm_5c81ea53-c375-492e-8426-88b80c2ccb25/prometheus-operator-admission-webhook/0.log" Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.298760 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7769c7475b-g626h_a3054e82-c7b2-4419-86cf-7089e43c4612/prometheus-operator-admission-webhook/0.log" Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.421409 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbj4l" Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.490684 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:31:00 crc kubenswrapper[4810]: E1201 17:31:00.491156 4810 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rbt4q_openshift-machine-config-operator(6c275317-c741-4d61-a399-d196f37f1745)\"" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.492312 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-7f426_9f0473a9-3b74-4538-909f-59f84c97d416/operator/0.log" Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.503169 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220" path="/var/lib/kubelet/pods/d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220/volumes" Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.566135 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-vcr8r_b78207c2-8eae-48a2-8eaf-aa967d534298/perses-operator/0.log" Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.587845 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c-utilities\") pod \"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c\" (UID: \"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c\") " Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.588134 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqtgq\" (UniqueName: \"kubernetes.io/projected/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c-kube-api-access-jqtgq\") pod \"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c\" (UID: \"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c\") " Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.588395 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c-catalog-content\") pod \"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c\" (UID: \"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c\") " Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.588840 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c-utilities" (OuterVolumeSpecName: "utilities") pod "23cddec1-85c4-4d5c-8dfe-eeb4b17f375c" (UID: "23cddec1-85c4-4d5c-8dfe-eeb4b17f375c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.589297 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.602274 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c-kube-api-access-jqtgq" (OuterVolumeSpecName: "kube-api-access-jqtgq") pod "23cddec1-85c4-4d5c-8dfe-eeb4b17f375c" (UID: "23cddec1-85c4-4d5c-8dfe-eeb4b17f375c"). InnerVolumeSpecName "kube-api-access-jqtgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.691383 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqtgq\" (UniqueName: \"kubernetes.io/projected/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c-kube-api-access-jqtgq\") on node \"crc\" DevicePath \"\"" Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.703651 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23cddec1-85c4-4d5c-8dfe-eeb4b17f375c" (UID: "23cddec1-85c4-4d5c-8dfe-eeb4b17f375c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.793676 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.918125 4810 generic.go:334] "Generic (PLEG): container finished" podID="23cddec1-85c4-4d5c-8dfe-eeb4b17f375c" containerID="0982b226f47cafb10c65d344c61a850a4b853d7b94af8f8f448a24e9f2e15741" exitCode=0 Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.918180 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbj4l" event={"ID":"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c","Type":"ContainerDied","Data":"0982b226f47cafb10c65d344c61a850a4b853d7b94af8f8f448a24e9f2e15741"} Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.918221 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jbj4l" event={"ID":"23cddec1-85c4-4d5c-8dfe-eeb4b17f375c","Type":"ContainerDied","Data":"fbaad6897b55980288fe880e37f79074ad9ec678b2dad413b99c28995eb172a1"} Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.918253 4810 scope.go:117] "RemoveContainer" containerID="0982b226f47cafb10c65d344c61a850a4b853d7b94af8f8f448a24e9f2e15741" Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.919147 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jbj4l" Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.971119 4810 scope.go:117] "RemoveContainer" containerID="dff898b1b52f324955141745d54d13cc9365265f4ee363f45a78ed6a33070d20" Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.980009 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jbj4l"] Dec 01 17:31:00 crc kubenswrapper[4810]: I1201 17:31:00.997765 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jbj4l"] Dec 01 17:31:01 crc kubenswrapper[4810]: I1201 17:31:01.027638 4810 scope.go:117] "RemoveContainer" containerID="04629219c6314028f8f9b163577cf75ec16303678fb2d660ec473f69fcccad28" Dec 01 17:31:01 crc kubenswrapper[4810]: I1201 17:31:01.058713 4810 scope.go:117] "RemoveContainer" containerID="0982b226f47cafb10c65d344c61a850a4b853d7b94af8f8f448a24e9f2e15741" Dec 01 17:31:01 crc kubenswrapper[4810]: E1201 17:31:01.059249 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0982b226f47cafb10c65d344c61a850a4b853d7b94af8f8f448a24e9f2e15741\": container with ID starting with 0982b226f47cafb10c65d344c61a850a4b853d7b94af8f8f448a24e9f2e15741 not found: ID does not exist" containerID="0982b226f47cafb10c65d344c61a850a4b853d7b94af8f8f448a24e9f2e15741" Dec 01 17:31:01 crc kubenswrapper[4810]: I1201 17:31:01.059289 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0982b226f47cafb10c65d344c61a850a4b853d7b94af8f8f448a24e9f2e15741"} err="failed to get container status \"0982b226f47cafb10c65d344c61a850a4b853d7b94af8f8f448a24e9f2e15741\": rpc error: code = NotFound desc = could not find container \"0982b226f47cafb10c65d344c61a850a4b853d7b94af8f8f448a24e9f2e15741\": container with ID starting with 0982b226f47cafb10c65d344c61a850a4b853d7b94af8f8f448a24e9f2e15741 not found: ID does not exist" Dec 01 17:31:01 crc kubenswrapper[4810]: I1201 17:31:01.059315 4810 scope.go:117] "RemoveContainer" containerID="dff898b1b52f324955141745d54d13cc9365265f4ee363f45a78ed6a33070d20" Dec 01 17:31:01 crc kubenswrapper[4810]: E1201 17:31:01.059601 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dff898b1b52f324955141745d54d13cc9365265f4ee363f45a78ed6a33070d20\": container with ID starting with dff898b1b52f324955141745d54d13cc9365265f4ee363f45a78ed6a33070d20 not found: ID does not exist" containerID="dff898b1b52f324955141745d54d13cc9365265f4ee363f45a78ed6a33070d20" Dec 01 17:31:01 crc kubenswrapper[4810]: I1201 17:31:01.059642 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dff898b1b52f324955141745d54d13cc9365265f4ee363f45a78ed6a33070d20"} err="failed to get container status \"dff898b1b52f324955141745d54d13cc9365265f4ee363f45a78ed6a33070d20\": rpc error: code = NotFound desc = could not find container \"dff898b1b52f324955141745d54d13cc9365265f4ee363f45a78ed6a33070d20\": container with ID starting with dff898b1b52f324955141745d54d13cc9365265f4ee363f45a78ed6a33070d20 not found: ID does not exist" Dec 01 17:31:01 crc kubenswrapper[4810]: I1201 17:31:01.059668 4810 scope.go:117] "RemoveContainer" containerID="04629219c6314028f8f9b163577cf75ec16303678fb2d660ec473f69fcccad28" Dec 01 17:31:01 crc kubenswrapper[4810]: E1201 17:31:01.066765 4810 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04629219c6314028f8f9b163577cf75ec16303678fb2d660ec473f69fcccad28\": container with ID starting with 04629219c6314028f8f9b163577cf75ec16303678fb2d660ec473f69fcccad28 not found: ID does not exist" containerID="04629219c6314028f8f9b163577cf75ec16303678fb2d660ec473f69fcccad28" Dec 01 17:31:01 crc kubenswrapper[4810]: I1201 17:31:01.066821 4810 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04629219c6314028f8f9b163577cf75ec16303678fb2d660ec473f69fcccad28"} err="failed to get container status \"04629219c6314028f8f9b163577cf75ec16303678fb2d660ec473f69fcccad28\": rpc error: code = NotFound desc = could not find container \"04629219c6314028f8f9b163577cf75ec16303678fb2d660ec473f69fcccad28\": container with ID starting with 04629219c6314028f8f9b163577cf75ec16303678fb2d660ec473f69fcccad28 not found: ID does not exist" Dec 01 17:31:02 crc kubenswrapper[4810]: I1201 17:31:02.504734 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23cddec1-85c4-4d5c-8dfe-eeb4b17f375c" path="/var/lib/kubelet/pods/23cddec1-85c4-4d5c-8dfe-eeb4b17f375c/volumes" Dec 01 17:31:03 crc kubenswrapper[4810]: I1201 17:31:03.782161 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cpblh" Dec 01 17:31:04 crc kubenswrapper[4810]: I1201 17:31:04.482421 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cpblh"] Dec 01 17:31:04 crc kubenswrapper[4810]: I1201 17:31:04.482673 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cpblh" podUID="319f7ef2-d70b-4351-b776-33b68596960f" containerName="registry-server" containerID="cri-o://79b736a7cd78f6677ac990e369202fef265f65aa5ca232a6badff3bfcd2de287" gracePeriod=2 Dec 01 17:31:04 crc kubenswrapper[4810]: I1201 17:31:04.979230 4810 generic.go:334] "Generic (PLEG): container finished" podID="319f7ef2-d70b-4351-b776-33b68596960f" containerID="79b736a7cd78f6677ac990e369202fef265f65aa5ca232a6badff3bfcd2de287" exitCode=0 Dec 01 17:31:04 crc kubenswrapper[4810]: I1201 17:31:04.979569 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpblh" event={"ID":"319f7ef2-d70b-4351-b776-33b68596960f","Type":"ContainerDied","Data":"79b736a7cd78f6677ac990e369202fef265f65aa5ca232a6badff3bfcd2de287"} Dec 01 17:31:05 crc kubenswrapper[4810]: I1201 17:31:05.105571 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cpblh" Dec 01 17:31:05 crc kubenswrapper[4810]: I1201 17:31:05.184345 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/319f7ef2-d70b-4351-b776-33b68596960f-catalog-content\") pod \"319f7ef2-d70b-4351-b776-33b68596960f\" (UID: \"319f7ef2-d70b-4351-b776-33b68596960f\") " Dec 01 17:31:05 crc kubenswrapper[4810]: I1201 17:31:05.184407 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7l4kd\" (UniqueName: \"kubernetes.io/projected/319f7ef2-d70b-4351-b776-33b68596960f-kube-api-access-7l4kd\") pod \"319f7ef2-d70b-4351-b776-33b68596960f\" (UID: \"319f7ef2-d70b-4351-b776-33b68596960f\") " Dec 01 17:31:05 crc kubenswrapper[4810]: I1201 17:31:05.184543 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/319f7ef2-d70b-4351-b776-33b68596960f-utilities\") pod \"319f7ef2-d70b-4351-b776-33b68596960f\" (UID: \"319f7ef2-d70b-4351-b776-33b68596960f\") " Dec 01 17:31:05 crc kubenswrapper[4810]: I1201 17:31:05.185931 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/319f7ef2-d70b-4351-b776-33b68596960f-utilities" (OuterVolumeSpecName: "utilities") pod "319f7ef2-d70b-4351-b776-33b68596960f" (UID: "319f7ef2-d70b-4351-b776-33b68596960f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:31:05 crc kubenswrapper[4810]: I1201 17:31:05.210775 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/319f7ef2-d70b-4351-b776-33b68596960f-kube-api-access-7l4kd" (OuterVolumeSpecName: "kube-api-access-7l4kd") pod "319f7ef2-d70b-4351-b776-33b68596960f" (UID: "319f7ef2-d70b-4351-b776-33b68596960f"). InnerVolumeSpecName "kube-api-access-7l4kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:31:05 crc kubenswrapper[4810]: I1201 17:31:05.231278 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/319f7ef2-d70b-4351-b776-33b68596960f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "319f7ef2-d70b-4351-b776-33b68596960f" (UID: "319f7ef2-d70b-4351-b776-33b68596960f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:31:05 crc kubenswrapper[4810]: I1201 17:31:05.287328 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7l4kd\" (UniqueName: \"kubernetes.io/projected/319f7ef2-d70b-4351-b776-33b68596960f-kube-api-access-7l4kd\") on node \"crc\" DevicePath \"\"" Dec 01 17:31:05 crc kubenswrapper[4810]: I1201 17:31:05.287358 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/319f7ef2-d70b-4351-b776-33b68596960f-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:31:05 crc kubenswrapper[4810]: I1201 17:31:05.287367 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/319f7ef2-d70b-4351-b776-33b68596960f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:31:05 crc kubenswrapper[4810]: I1201 17:31:05.992424 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpblh" event={"ID":"319f7ef2-d70b-4351-b776-33b68596960f","Type":"ContainerDied","Data":"176b1234deb9342661387d8600d48be2efb8738678e5a0bdeb57acdbd5534557"} Dec 01 17:31:05 crc kubenswrapper[4810]: I1201 17:31:05.992500 4810 scope.go:117] "RemoveContainer" containerID="79b736a7cd78f6677ac990e369202fef265f65aa5ca232a6badff3bfcd2de287" Dec 01 17:31:05 crc kubenswrapper[4810]: I1201 17:31:05.992525 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cpblh" Dec 01 17:31:06 crc kubenswrapper[4810]: I1201 17:31:06.024492 4810 scope.go:117] "RemoveContainer" containerID="0a02c0f038c4419644ea7110d0b4d5e6a6a69a97bc3da8ede422949f5976d7e9" Dec 01 17:31:06 crc kubenswrapper[4810]: I1201 17:31:06.077401 4810 scope.go:117] "RemoveContainer" containerID="3a02ebc51554b3d87b10c6e2929329b43b1c6a61d5e7b29189d9f38178d6e44f" Dec 01 17:31:06 crc kubenswrapper[4810]: I1201 17:31:06.079425 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cpblh"] Dec 01 17:31:06 crc kubenswrapper[4810]: I1201 17:31:06.097270 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cpblh"] Dec 01 17:31:06 crc kubenswrapper[4810]: I1201 17:31:06.506453 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="319f7ef2-d70b-4351-b776-33b68596960f" path="/var/lib/kubelet/pods/319f7ef2-d70b-4351-b776-33b68596960f/volumes" Dec 01 17:31:11 crc kubenswrapper[4810]: I1201 17:31:11.492137 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:31:12 crc kubenswrapper[4810]: I1201 17:31:12.081363 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"f9babb667b0dcfca4a95ffaddf7776eb16633358a7568c734c573f654ae0a286"} Dec 01 17:31:26 crc kubenswrapper[4810]: E1201 17:31:26.686493 4810 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.182:48810->38.102.83.182:41089: write tcp 38.102.83.182:48810->38.102.83.182:41089: write: broken pipe Dec 01 17:31:29 crc kubenswrapper[4810]: E1201 17:31:29.579096 4810 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.182:48926->38.102.83.182:41089: write tcp 38.102.83.182:48926->38.102.83.182:41089: write: broken pipe Dec 01 17:33:02 crc kubenswrapper[4810]: I1201 17:33:02.483087 4810 generic.go:334] "Generic (PLEG): container finished" podID="58292b71-6004-474e-aacf-6b8c7349ab13" containerID="dbdccde095f3fc50cffdc9f2bf7cf984033ebde7ce9452363bf8db0845f2c38e" exitCode=0 Dec 01 17:33:02 crc kubenswrapper[4810]: I1201 17:33:02.483137 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xdmnf/must-gather-tcrzd" event={"ID":"58292b71-6004-474e-aacf-6b8c7349ab13","Type":"ContainerDied","Data":"dbdccde095f3fc50cffdc9f2bf7cf984033ebde7ce9452363bf8db0845f2c38e"} Dec 01 17:33:02 crc kubenswrapper[4810]: I1201 17:33:02.484521 4810 scope.go:117] "RemoveContainer" containerID="dbdccde095f3fc50cffdc9f2bf7cf984033ebde7ce9452363bf8db0845f2c38e" Dec 01 17:33:02 crc kubenswrapper[4810]: I1201 17:33:02.891723 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xdmnf_must-gather-tcrzd_58292b71-6004-474e-aacf-6b8c7349ab13/gather/0.log" Dec 01 17:33:12 crc kubenswrapper[4810]: I1201 17:33:12.111336 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xdmnf/must-gather-tcrzd"] Dec 01 17:33:12 crc kubenswrapper[4810]: I1201 17:33:12.112072 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-xdmnf/must-gather-tcrzd" podUID="58292b71-6004-474e-aacf-6b8c7349ab13" containerName="copy" containerID="cri-o://d1e8f6aa8ffa77af42e3972468d9704bc6945998d53e8e8b15871ea23457fad2" gracePeriod=2 Dec 01 17:33:12 crc kubenswrapper[4810]: I1201 17:33:12.120268 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xdmnf/must-gather-tcrzd"] Dec 01 17:33:12 crc kubenswrapper[4810]: I1201 17:33:12.611941 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xdmnf_must-gather-tcrzd_58292b71-6004-474e-aacf-6b8c7349ab13/copy/0.log" Dec 01 17:33:12 crc kubenswrapper[4810]: I1201 17:33:12.612698 4810 generic.go:334] "Generic (PLEG): container finished" podID="58292b71-6004-474e-aacf-6b8c7349ab13" containerID="d1e8f6aa8ffa77af42e3972468d9704bc6945998d53e8e8b15871ea23457fad2" exitCode=143 Dec 01 17:33:12 crc kubenswrapper[4810]: I1201 17:33:12.612747 4810 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d56fb36b1e641e3ce8484ae761267b34b71dca0509a32b473cb848a9bf65f26" Dec 01 17:33:12 crc kubenswrapper[4810]: I1201 17:33:12.670969 4810 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xdmnf_must-gather-tcrzd_58292b71-6004-474e-aacf-6b8c7349ab13/copy/0.log" Dec 01 17:33:12 crc kubenswrapper[4810]: I1201 17:33:12.671724 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xdmnf/must-gather-tcrzd" Dec 01 17:33:12 crc kubenswrapper[4810]: I1201 17:33:12.754726 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/58292b71-6004-474e-aacf-6b8c7349ab13-must-gather-output\") pod \"58292b71-6004-474e-aacf-6b8c7349ab13\" (UID: \"58292b71-6004-474e-aacf-6b8c7349ab13\") " Dec 01 17:33:12 crc kubenswrapper[4810]: I1201 17:33:12.754827 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87jvt\" (UniqueName: \"kubernetes.io/projected/58292b71-6004-474e-aacf-6b8c7349ab13-kube-api-access-87jvt\") pod \"58292b71-6004-474e-aacf-6b8c7349ab13\" (UID: \"58292b71-6004-474e-aacf-6b8c7349ab13\") " Dec 01 17:33:12 crc kubenswrapper[4810]: I1201 17:33:12.778783 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58292b71-6004-474e-aacf-6b8c7349ab13-kube-api-access-87jvt" (OuterVolumeSpecName: "kube-api-access-87jvt") pod "58292b71-6004-474e-aacf-6b8c7349ab13" (UID: "58292b71-6004-474e-aacf-6b8c7349ab13"). InnerVolumeSpecName "kube-api-access-87jvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:33:12 crc kubenswrapper[4810]: I1201 17:33:12.857616 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87jvt\" (UniqueName: \"kubernetes.io/projected/58292b71-6004-474e-aacf-6b8c7349ab13-kube-api-access-87jvt\") on node \"crc\" DevicePath \"\"" Dec 01 17:33:12 crc kubenswrapper[4810]: I1201 17:33:12.980249 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58292b71-6004-474e-aacf-6b8c7349ab13-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "58292b71-6004-474e-aacf-6b8c7349ab13" (UID: "58292b71-6004-474e-aacf-6b8c7349ab13"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:33:13 crc kubenswrapper[4810]: I1201 17:33:13.063008 4810 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/58292b71-6004-474e-aacf-6b8c7349ab13-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 01 17:33:13 crc kubenswrapper[4810]: I1201 17:33:13.621313 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xdmnf/must-gather-tcrzd" Dec 01 17:33:14 crc kubenswrapper[4810]: I1201 17:33:14.505989 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58292b71-6004-474e-aacf-6b8c7349ab13" path="/var/lib/kubelet/pods/58292b71-6004-474e-aacf-6b8c7349ab13/volumes" Dec 01 17:33:19 crc kubenswrapper[4810]: I1201 17:33:19.326326 4810 scope.go:117] "RemoveContainer" containerID="d1e8f6aa8ffa77af42e3972468d9704bc6945998d53e8e8b15871ea23457fad2" Dec 01 17:33:19 crc kubenswrapper[4810]: I1201 17:33:19.355306 4810 scope.go:117] "RemoveContainer" containerID="dbdccde095f3fc50cffdc9f2bf7cf984033ebde7ce9452363bf8db0845f2c38e" Dec 01 17:33:32 crc kubenswrapper[4810]: I1201 17:33:32.972831 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:33:32 crc kubenswrapper[4810]: I1201 17:33:32.973498 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:34:02 crc kubenswrapper[4810]: I1201 17:34:02.972004 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:34:02 crc kubenswrapper[4810]: I1201 17:34:02.972683 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:34:32 crc kubenswrapper[4810]: I1201 17:34:32.972449 4810 patch_prober.go:28] interesting pod/machine-config-daemon-rbt4q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 17:34:32 crc kubenswrapper[4810]: I1201 17:34:32.973985 4810 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 17:34:32 crc kubenswrapper[4810]: I1201 17:34:32.974088 4810 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" Dec 01 17:34:32 crc kubenswrapper[4810]: I1201 17:34:32.974881 4810 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f9babb667b0dcfca4a95ffaddf7776eb16633358a7568c734c573f654ae0a286"} pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 17:34:32 crc kubenswrapper[4810]: I1201 17:34:32.975136 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" podUID="6c275317-c741-4d61-a399-d196f37f1745" containerName="machine-config-daemon" containerID="cri-o://f9babb667b0dcfca4a95ffaddf7776eb16633358a7568c734c573f654ae0a286" gracePeriod=600 Dec 01 17:34:33 crc kubenswrapper[4810]: I1201 17:34:33.552738 4810 generic.go:334] "Generic (PLEG): container finished" podID="6c275317-c741-4d61-a399-d196f37f1745" containerID="f9babb667b0dcfca4a95ffaddf7776eb16633358a7568c734c573f654ae0a286" exitCode=0 Dec 01 17:34:33 crc kubenswrapper[4810]: I1201 17:34:33.552837 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerDied","Data":"f9babb667b0dcfca4a95ffaddf7776eb16633358a7568c734c573f654ae0a286"} Dec 01 17:34:33 crc kubenswrapper[4810]: I1201 17:34:33.554007 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rbt4q" event={"ID":"6c275317-c741-4d61-a399-d196f37f1745","Type":"ContainerStarted","Data":"ef87b622b9ca11fa6f0d45d42d6a4a11bba873825da0b38cea5d3a321ff96cb0"} Dec 01 17:34:33 crc kubenswrapper[4810]: I1201 17:34:33.554066 4810 scope.go:117] "RemoveContainer" containerID="beaf420f649880a26e7b3593a73cb57f5d9e48259971e060ba74d7db19c614bd" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.746824 4810 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fnb7l"] Dec 01 17:35:59 crc kubenswrapper[4810]: E1201 17:35:59.747825 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="319f7ef2-d70b-4351-b776-33b68596960f" containerName="registry-server" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.747840 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="319f7ef2-d70b-4351-b776-33b68596960f" containerName="registry-server" Dec 01 17:35:59 crc kubenswrapper[4810]: E1201 17:35:59.747855 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23cddec1-85c4-4d5c-8dfe-eeb4b17f375c" containerName="extract-content" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.747864 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="23cddec1-85c4-4d5c-8dfe-eeb4b17f375c" containerName="extract-content" Dec 01 17:35:59 crc kubenswrapper[4810]: E1201 17:35:59.747890 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58292b71-6004-474e-aacf-6b8c7349ab13" containerName="gather" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.747899 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="58292b71-6004-474e-aacf-6b8c7349ab13" containerName="gather" Dec 01 17:35:59 crc kubenswrapper[4810]: E1201 17:35:59.747924 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220" containerName="extract-utilities" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.747932 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220" containerName="extract-utilities" Dec 01 17:35:59 crc kubenswrapper[4810]: E1201 17:35:59.747948 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58292b71-6004-474e-aacf-6b8c7349ab13" containerName="copy" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.747956 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="58292b71-6004-474e-aacf-6b8c7349ab13" containerName="copy" Dec 01 17:35:59 crc kubenswrapper[4810]: E1201 17:35:59.747972 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="319f7ef2-d70b-4351-b776-33b68596960f" containerName="extract-utilities" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.747980 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="319f7ef2-d70b-4351-b776-33b68596960f" containerName="extract-utilities" Dec 01 17:35:59 crc kubenswrapper[4810]: E1201 17:35:59.747999 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="319f7ef2-d70b-4351-b776-33b68596960f" containerName="extract-content" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.748007 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="319f7ef2-d70b-4351-b776-33b68596960f" containerName="extract-content" Dec 01 17:35:59 crc kubenswrapper[4810]: E1201 17:35:59.748019 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23cddec1-85c4-4d5c-8dfe-eeb4b17f375c" containerName="registry-server" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.748027 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="23cddec1-85c4-4d5c-8dfe-eeb4b17f375c" containerName="registry-server" Dec 01 17:35:59 crc kubenswrapper[4810]: E1201 17:35:59.748050 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220" containerName="extract-content" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.748057 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220" containerName="extract-content" Dec 01 17:35:59 crc kubenswrapper[4810]: E1201 17:35:59.748071 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23cddec1-85c4-4d5c-8dfe-eeb4b17f375c" containerName="extract-utilities" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.748079 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="23cddec1-85c4-4d5c-8dfe-eeb4b17f375c" containerName="extract-utilities" Dec 01 17:35:59 crc kubenswrapper[4810]: E1201 17:35:59.748097 4810 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220" containerName="registry-server" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.748105 4810 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220" containerName="registry-server" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.748387 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="319f7ef2-d70b-4351-b776-33b68596960f" containerName="registry-server" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.748401 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="58292b71-6004-474e-aacf-6b8c7349ab13" containerName="gather" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.748417 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="58292b71-6004-474e-aacf-6b8c7349ab13" containerName="copy" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.748439 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="23cddec1-85c4-4d5c-8dfe-eeb4b17f375c" containerName="registry-server" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.748456 4810 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5bb84bd-7fbd-4864-be5f-6cb9d3e1d220" containerName="registry-server" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.750575 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fnb7l"] Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.750669 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fnb7l" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.808252 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlrf6\" (UniqueName: \"kubernetes.io/projected/98a424a6-03ae-4b17-a926-8395babb1ae4-kube-api-access-zlrf6\") pod \"certified-operators-fnb7l\" (UID: \"98a424a6-03ae-4b17-a926-8395babb1ae4\") " pod="openshift-marketplace/certified-operators-fnb7l" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.808462 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a424a6-03ae-4b17-a926-8395babb1ae4-utilities\") pod \"certified-operators-fnb7l\" (UID: \"98a424a6-03ae-4b17-a926-8395babb1ae4\") " pod="openshift-marketplace/certified-operators-fnb7l" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.808670 4810 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a424a6-03ae-4b17-a926-8395babb1ae4-catalog-content\") pod \"certified-operators-fnb7l\" (UID: \"98a424a6-03ae-4b17-a926-8395babb1ae4\") " pod="openshift-marketplace/certified-operators-fnb7l" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.910659 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlrf6\" (UniqueName: \"kubernetes.io/projected/98a424a6-03ae-4b17-a926-8395babb1ae4-kube-api-access-zlrf6\") pod \"certified-operators-fnb7l\" (UID: \"98a424a6-03ae-4b17-a926-8395babb1ae4\") " pod="openshift-marketplace/certified-operators-fnb7l" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.910772 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a424a6-03ae-4b17-a926-8395babb1ae4-utilities\") pod \"certified-operators-fnb7l\" (UID: \"98a424a6-03ae-4b17-a926-8395babb1ae4\") " pod="openshift-marketplace/certified-operators-fnb7l" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.910861 4810 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a424a6-03ae-4b17-a926-8395babb1ae4-catalog-content\") pod \"certified-operators-fnb7l\" (UID: \"98a424a6-03ae-4b17-a926-8395babb1ae4\") " pod="openshift-marketplace/certified-operators-fnb7l" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.911382 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a424a6-03ae-4b17-a926-8395babb1ae4-utilities\") pod \"certified-operators-fnb7l\" (UID: \"98a424a6-03ae-4b17-a926-8395babb1ae4\") " pod="openshift-marketplace/certified-operators-fnb7l" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.911414 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a424a6-03ae-4b17-a926-8395babb1ae4-catalog-content\") pod \"certified-operators-fnb7l\" (UID: \"98a424a6-03ae-4b17-a926-8395babb1ae4\") " pod="openshift-marketplace/certified-operators-fnb7l" Dec 01 17:35:59 crc kubenswrapper[4810]: I1201 17:35:59.932137 4810 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlrf6\" (UniqueName: \"kubernetes.io/projected/98a424a6-03ae-4b17-a926-8395babb1ae4-kube-api-access-zlrf6\") pod \"certified-operators-fnb7l\" (UID: \"98a424a6-03ae-4b17-a926-8395babb1ae4\") " pod="openshift-marketplace/certified-operators-fnb7l" Dec 01 17:36:00 crc kubenswrapper[4810]: I1201 17:36:00.103178 4810 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fnb7l" Dec 01 17:36:00 crc kubenswrapper[4810]: I1201 17:36:00.693095 4810 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fnb7l"] Dec 01 17:36:00 crc kubenswrapper[4810]: W1201 17:36:00.697761 4810 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98a424a6_03ae_4b17_a926_8395babb1ae4.slice/crio-2449c20c0613de14b7ae6c740489fb0b4e2ec74ecdd0a1428f5e6c0bf0710ef3 WatchSource:0}: Error finding container 2449c20c0613de14b7ae6c740489fb0b4e2ec74ecdd0a1428f5e6c0bf0710ef3: Status 404 returned error can't find the container with id 2449c20c0613de14b7ae6c740489fb0b4e2ec74ecdd0a1428f5e6c0bf0710ef3 Dec 01 17:36:01 crc kubenswrapper[4810]: I1201 17:36:01.569768 4810 generic.go:334] "Generic (PLEG): container finished" podID="98a424a6-03ae-4b17-a926-8395babb1ae4" containerID="3fd3393a113802d66f84eb4ac1bd69af7de573bcf5e7437ff017a02f7c1272ef" exitCode=0 Dec 01 17:36:01 crc kubenswrapper[4810]: I1201 17:36:01.569869 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnb7l" event={"ID":"98a424a6-03ae-4b17-a926-8395babb1ae4","Type":"ContainerDied","Data":"3fd3393a113802d66f84eb4ac1bd69af7de573bcf5e7437ff017a02f7c1272ef"} Dec 01 17:36:01 crc kubenswrapper[4810]: I1201 17:36:01.570138 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnb7l" event={"ID":"98a424a6-03ae-4b17-a926-8395babb1ae4","Type":"ContainerStarted","Data":"2449c20c0613de14b7ae6c740489fb0b4e2ec74ecdd0a1428f5e6c0bf0710ef3"} Dec 01 17:36:01 crc kubenswrapper[4810]: I1201 17:36:01.573794 4810 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 17:36:02 crc kubenswrapper[4810]: I1201 17:36:02.587969 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnb7l" event={"ID":"98a424a6-03ae-4b17-a926-8395babb1ae4","Type":"ContainerStarted","Data":"a37527baf704d5111ca298569c2b473cff053ec11a658bbeba0213b854257c1b"} Dec 01 17:36:03 crc kubenswrapper[4810]: I1201 17:36:03.611867 4810 generic.go:334] "Generic (PLEG): container finished" podID="98a424a6-03ae-4b17-a926-8395babb1ae4" containerID="a37527baf704d5111ca298569c2b473cff053ec11a658bbeba0213b854257c1b" exitCode=0 Dec 01 17:36:03 crc kubenswrapper[4810]: I1201 17:36:03.611916 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnb7l" event={"ID":"98a424a6-03ae-4b17-a926-8395babb1ae4","Type":"ContainerDied","Data":"a37527baf704d5111ca298569c2b473cff053ec11a658bbeba0213b854257c1b"} Dec 01 17:36:04 crc kubenswrapper[4810]: I1201 17:36:04.627079 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnb7l" event={"ID":"98a424a6-03ae-4b17-a926-8395babb1ae4","Type":"ContainerStarted","Data":"722ef52a8685d352c255e087ce0b7a8a86f28ffb1ed1ecad284412f476ac0d48"} Dec 01 17:36:04 crc kubenswrapper[4810]: I1201 17:36:04.658123 4810 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fnb7l" podStartSLOduration=3.121254465 podStartE2EDuration="5.658093356s" podCreationTimestamp="2025-12-01 17:35:59 +0000 UTC" firstStartedPulling="2025-12-01 17:36:01.573146981 +0000 UTC m=+10927.336656624" lastFinishedPulling="2025-12-01 17:36:04.109985912 +0000 UTC m=+10929.873495515" observedRunningTime="2025-12-01 17:36:04.653072492 +0000 UTC m=+10930.416582135" watchObservedRunningTime="2025-12-01 17:36:04.658093356 +0000 UTC m=+10930.421602969" Dec 01 17:36:10 crc kubenswrapper[4810]: I1201 17:36:10.104145 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fnb7l" Dec 01 17:36:10 crc kubenswrapper[4810]: I1201 17:36:10.104745 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fnb7l" Dec 01 17:36:10 crc kubenswrapper[4810]: I1201 17:36:10.164238 4810 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fnb7l" Dec 01 17:36:10 crc kubenswrapper[4810]: I1201 17:36:10.749185 4810 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fnb7l" Dec 01 17:36:10 crc kubenswrapper[4810]: I1201 17:36:10.812803 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fnb7l"] Dec 01 17:36:12 crc kubenswrapper[4810]: I1201 17:36:12.724879 4810 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fnb7l" podUID="98a424a6-03ae-4b17-a926-8395babb1ae4" containerName="registry-server" containerID="cri-o://722ef52a8685d352c255e087ce0b7a8a86f28ffb1ed1ecad284412f476ac0d48" gracePeriod=2 Dec 01 17:36:13 crc kubenswrapper[4810]: I1201 17:36:13.737604 4810 generic.go:334] "Generic (PLEG): container finished" podID="98a424a6-03ae-4b17-a926-8395babb1ae4" containerID="722ef52a8685d352c255e087ce0b7a8a86f28ffb1ed1ecad284412f476ac0d48" exitCode=0 Dec 01 17:36:13 crc kubenswrapper[4810]: I1201 17:36:13.737715 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnb7l" event={"ID":"98a424a6-03ae-4b17-a926-8395babb1ae4","Type":"ContainerDied","Data":"722ef52a8685d352c255e087ce0b7a8a86f28ffb1ed1ecad284412f476ac0d48"} Dec 01 17:36:13 crc kubenswrapper[4810]: I1201 17:36:13.899179 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fnb7l" Dec 01 17:36:13 crc kubenswrapper[4810]: I1201 17:36:13.930418 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a424a6-03ae-4b17-a926-8395babb1ae4-utilities\") pod \"98a424a6-03ae-4b17-a926-8395babb1ae4\" (UID: \"98a424a6-03ae-4b17-a926-8395babb1ae4\") " Dec 01 17:36:13 crc kubenswrapper[4810]: I1201 17:36:13.930789 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a424a6-03ae-4b17-a926-8395babb1ae4-catalog-content\") pod \"98a424a6-03ae-4b17-a926-8395babb1ae4\" (UID: \"98a424a6-03ae-4b17-a926-8395babb1ae4\") " Dec 01 17:36:13 crc kubenswrapper[4810]: I1201 17:36:13.931918 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98a424a6-03ae-4b17-a926-8395babb1ae4-utilities" (OuterVolumeSpecName: "utilities") pod "98a424a6-03ae-4b17-a926-8395babb1ae4" (UID: "98a424a6-03ae-4b17-a926-8395babb1ae4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:36:13 crc kubenswrapper[4810]: I1201 17:36:13.995734 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98a424a6-03ae-4b17-a926-8395babb1ae4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98a424a6-03ae-4b17-a926-8395babb1ae4" (UID: "98a424a6-03ae-4b17-a926-8395babb1ae4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 17:36:14 crc kubenswrapper[4810]: I1201 17:36:14.032043 4810 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlrf6\" (UniqueName: \"kubernetes.io/projected/98a424a6-03ae-4b17-a926-8395babb1ae4-kube-api-access-zlrf6\") pod \"98a424a6-03ae-4b17-a926-8395babb1ae4\" (UID: \"98a424a6-03ae-4b17-a926-8395babb1ae4\") " Dec 01 17:36:14 crc kubenswrapper[4810]: I1201 17:36:14.032553 4810 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a424a6-03ae-4b17-a926-8395babb1ae4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:14 crc kubenswrapper[4810]: I1201 17:36:14.032571 4810 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a424a6-03ae-4b17-a926-8395babb1ae4-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:14 crc kubenswrapper[4810]: I1201 17:36:14.037259 4810 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98a424a6-03ae-4b17-a926-8395babb1ae4-kube-api-access-zlrf6" (OuterVolumeSpecName: "kube-api-access-zlrf6") pod "98a424a6-03ae-4b17-a926-8395babb1ae4" (UID: "98a424a6-03ae-4b17-a926-8395babb1ae4"). InnerVolumeSpecName "kube-api-access-zlrf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 17:36:14 crc kubenswrapper[4810]: I1201 17:36:14.134297 4810 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlrf6\" (UniqueName: \"kubernetes.io/projected/98a424a6-03ae-4b17-a926-8395babb1ae4-kube-api-access-zlrf6\") on node \"crc\" DevicePath \"\"" Dec 01 17:36:14 crc kubenswrapper[4810]: I1201 17:36:14.751154 4810 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fnb7l" event={"ID":"98a424a6-03ae-4b17-a926-8395babb1ae4","Type":"ContainerDied","Data":"2449c20c0613de14b7ae6c740489fb0b4e2ec74ecdd0a1428f5e6c0bf0710ef3"} Dec 01 17:36:14 crc kubenswrapper[4810]: I1201 17:36:14.751298 4810 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fnb7l" Dec 01 17:36:14 crc kubenswrapper[4810]: I1201 17:36:14.751534 4810 scope.go:117] "RemoveContainer" containerID="722ef52a8685d352c255e087ce0b7a8a86f28ffb1ed1ecad284412f476ac0d48" Dec 01 17:36:14 crc kubenswrapper[4810]: I1201 17:36:14.784605 4810 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fnb7l"] Dec 01 17:36:14 crc kubenswrapper[4810]: I1201 17:36:14.792358 4810 scope.go:117] "RemoveContainer" containerID="a37527baf704d5111ca298569c2b473cff053ec11a658bbeba0213b854257c1b" Dec 01 17:36:14 crc kubenswrapper[4810]: I1201 17:36:14.802843 4810 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fnb7l"] Dec 01 17:36:14 crc kubenswrapper[4810]: I1201 17:36:14.818710 4810 scope.go:117] "RemoveContainer" containerID="3fd3393a113802d66f84eb4ac1bd69af7de573bcf5e7437ff017a02f7c1272ef" Dec 01 17:36:16 crc kubenswrapper[4810]: I1201 17:36:16.511235 4810 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98a424a6-03ae-4b17-a926-8395babb1ae4" path="/var/lib/kubelet/pods/98a424a6-03ae-4b17-a926-8395babb1ae4/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515113351027024443 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015113351030017352 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015113323171016502 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015113323172015453 5ustar corecore